Hello Orthanc users,
I would like your advice on the problem I’m currently facing.
I am importing with a batch a large quantity of studies inside an orthanc instance. But those studies are sometimes duplicated, anonymized twice and I do not have control over this data of bad quality.
In the end, I end up with multiple studies in orthanc having the same studyInstanceUid, multiple series with the same studyInstanceUid and multiple instances with the same sopInstanceUid.
My goal is to centralize all my old data to keep it easily accessible from Orthanc. But having such bad data is problematic: Orthanc issues multiple warning saying my data “breaks the DICOM model of the real world” and nevertheless it would be a bad idea to keep duplicated objects when they are supposed to be unique in the real world (and all your scripts/programs logic is based on the uniqueness).
Has anyone ever set up strategies or script to handle this ?
I can have two strategies:
- either I clean my data before injecting it in Orthanc
- or I clean it after it being stored in Orthanc
I was leaning toward the second solution by creating a script I would run everyday that would delete all but one study for studies whom studyInstanceUID appears multiple times.
I think cleaning my data before sending it to Orthanc would be quite complicated.
What do you think ? Do you have tools or advices ?
Regards,
Francois