WO2017198518A1 - Dispositif de traitement de données d'image - Google Patents
Dispositif de traitement de données d'image Download PDFInfo
- Publication number
- WO2017198518A1 WO2017198518A1 PCT/EP2017/061246 EP2017061246W WO2017198518A1 WO 2017198518 A1 WO2017198518 A1 WO 2017198518A1 EP 2017061246 W EP2017061246 W EP 2017061246W WO 2017198518 A1 WO2017198518 A1 WO 2017198518A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image data
- mapped
- image
- partial
- features
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 77
- 238000012937 correction Methods 0.000 claims abstract description 20
- 230000009466 transformation Effects 0.000 claims abstract description 19
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 14
- 238000000034 method Methods 0.000 claims description 44
- 238000013507 mapping Methods 0.000 claims description 31
- 230000011218 segmentation Effects 0.000 claims description 26
- 238000002600 positron emission tomography Methods 0.000 claims description 25
- 238000003384 imaging method Methods 0.000 claims description 12
- 238000005259 measurement Methods 0.000 claims description 11
- 238000000691 measurement method Methods 0.000 claims description 9
- 238000004590 computer program Methods 0.000 claims description 8
- 230000001131 transforming effect Effects 0.000 claims description 7
- 238000002591 computed tomography Methods 0.000 claims description 6
- 238000012307 MRI technique Methods 0.000 claims description 3
- 210000004556 brain Anatomy 0.000 description 19
- 210000003710 cerebral cortex Anatomy 0.000 description 8
- 230000008901 benefit Effects 0.000 description 5
- 230000000694 effects Effects 0.000 description 5
- 238000007619 statistical method Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 210000004885 white matter Anatomy 0.000 description 4
- 230000002159 abnormal effect Effects 0.000 description 3
- 230000001419 dependent effect Effects 0.000 description 3
- 238000002595 magnetic resonance imaging Methods 0.000 description 3
- 230000002503 metabolic effect Effects 0.000 description 3
- 239000000700 radioactive tracer Substances 0.000 description 3
- 238000001356 surgical procedure Methods 0.000 description 3
- 206010072731 White matter lesion Diseases 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 230000017531 blood circulation Effects 0.000 description 2
- 238000002610 neuroimaging Methods 0.000 description 2
- 238000009206 nuclear medicine Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- ZCXUVYAZINUVJD-AHXZWLDOSA-N 2-deoxy-2-((18)F)fluoro-alpha-D-glucose Chemical compound OC[C@H]1O[C@H](O)[C@H]([18F])[C@@H](O)[C@@H]1O ZCXUVYAZINUVJD-AHXZWLDOSA-N 0.000 description 1
- 230000002547 anomalous effect Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 210000004204 blood vessel Anatomy 0.000 description 1
- 210000004027 cell Anatomy 0.000 description 1
- 210000001175 cerebrospinal fluid Anatomy 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000007428 craniotomy Methods 0.000 description 1
- 238000012864 cross contamination Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 238000009792 diffusion process Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 210000004884 grey matter Anatomy 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 230000004060 metabolic process Effects 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 230000003562 morphometric effect Effects 0.000 description 1
- 238000013425 morphometry Methods 0.000 description 1
- 230000002450 orbitofrontal effect Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000002560 therapeutic procedure Methods 0.000 description 1
- 238000002604 ultrasonography Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
- G06T7/0014—Biomedical image inspection using an image reference approach
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
Definitions
- the present invention relates to an image data processing device, a method for operating an image data processing device and to a computer program.
- WO 2014/139024 Al describes planning, navigation and simulation systems and methods for minimally invasive therapy in which the planning method and system uses patient specific pre-operative images.
- the planning system allows for multiple paths to be developed from the pre-operative images, and scores the paths depending on desired surgical outcome of the surgery and the navigation systems allow for minimally invasive port based surgical procedures, as well as craniotomies in the particular case of brain surgery.
- WO 2009/134820 A2 describes a method and apparatus for magnetic source magnetic resonance imaging.
- the method includes collecting energy signals from an object, providing additional information of characteristics of the object, and generating the image of the object from the energy signals and from the additional information such that the image includes a representation of a quantitative estimation of the characteristics.
- the additional information may comprise predetermined characteristics of the object, a magnitude image generated from the object, or magnetic signals collected from different relative orientations between the object and the imaging system.
- the image is generated by an inversion operation based on the collected signals and the additional information. The inversion operation minimizes a cost function obtained by combining the data extracted from the collected signals and the additional information of the object. Additionally, the image is used to detect a number of diagnostic features including microbleeds, contract agents and the like.
- WO 2009/065079 A2 describes devices, systems and techniques that relate to detecting volumetric changes in imaged structures over time, including devices, systems and techniques that enable precise registration of structures (e.g., brain areas) after large or subtle deformations, allowing precise registration at small spatial scales with a low boundary contrast.
- structures e.g., brain areas
- first and second mutually registered sets of image data representing measurement data acquired by different measurement techniques from an identical external object of an object type, the respective sets of image data containing at least one respective image feature with respective internal contrast features, wherein the first set of image data represents a first image having a higher spatial resolution than a second image represented by the second set of image data;
- segmentation information assign, based on at least one predetermined segmentation criterion, segmentation information to the image data of the first set of image data, the segmentation information representing a segmentation of the first image into at least two image segments;
- the processing unit is configured to:
- sample image data pertaining to a plurality of sample mapped image features from a statistical population of further external objects of the object type, the sample mapped image features having the reference shape;
- the lower spatial resolution of the image represented by the image data can be due to partial- volume effects (PVE) in the form of, for example, cross contamination between different image regions caused by a point spread function of a measurement system performing the measurement technique used to acquire the second set of image data.
- PVE partial- volume effects
- the present invention allows turning this into an advantage in that it can be used to actually limit scan times in obtaining image data, since low-resolution images obtained with such limited scan time can be enhanced in this regard by post-processing in accordance with the present invention.
- the image feature is segmented or parcellated into at least two image segments. This may be done by identifying certain known reference features which are common to most of the objects of the object type and allows for a significant spatial correspondence of the image features of both sets of image data.
- the image date processing device is configured to perform a partial- volume correction algorithm to combine the higher spatial resolution of the first set of image data with the measurement data represented by the second set of image data.
- a partial- volume correction algorithm to combine the higher spatial resolution of the first set of image data with the measurement data represented by the second set of image data.
- image enhancement techniques which primarily rely on recovering resolution directly from the second set of image data
- image-domain correction techniques which rely on information from the first set of image data to determine the appropriate correction
- projection based correction projection based correction.
- the processing unit receives the partial- volume corrected set of image data representing an image with a partial- volume corrected image feature from the transformation unit and determines mapped image data according to a predetermined mapping rule.
- This mapping rule transforms the partial- volume corrected image feature into a mapped image feature having a predetermined reference shape.
- the processing unit is thus configured to transform the shape of the partial- volume corrected image feature to a predetermined reference shape that includes mapped image features.
- the predetermined mapping rule is suitably designed to map the internal contrast features (i.e., the regions within the image feature that have different color or brightness values ranges), into the mapped image feature. This is preferably performed in a way that a position of a given mapped internal contrast feature within the mapped image feature can be tracked back to the position of the corresponding internal contrast feature of the partial- volume corrected image feature. Since the mapping rule transforms the shape of the image feature into the reference shape, it also changes a shape of the internal contrast features. The result of this mapping is mapped image data representing a mapped image feature that has mapped internal contrast features that have positions within the mapped image feature related via the mapping rule to positions the corresponding internal contrast features of the partial- volume corrected contrast features.
- the processing unit also receives sample image data pertaining to a plurality of sample mapped image features that belong to a statistical population of further external objects of the object type. Since the sample mapped image features also have the same reference shape, the processing unit is advantageously configured to identify significant mapped internal contrast features of the mapped image feature that differ in a statistically significant manner from corresponding sample mapped internal contrast features of the statistical population.
- the sample image data thus forms a statistical population to which the transformed image data can be compared in search of predetermined statistical deviations. This identification of significant mapped internal contrast features can be performed due to the fact that both the image data and the sample image data have been equivalently transformed and both sets of image data share the same reference shape, regardless of the original shape of the image feature, which is likely to be different and thus not matching.
- the image data processing device is suitably configured to provide a partial- volume corrected set of image data representing an image with a higher spatial resolution as the image represented by the second set of image data, to map the partial- volume corrected contrast feature onto a mapped image feature that has a reference shape and to compare the mapped image feature with sample mapped image features of objects of the same object type even in cases where the image features of the individual objects of the same object type do not have the same shape.
- the processing unit is also configured to provide output data depending on comparison result.
- the transformation unit and the processing unit share a common housing, whereas in other embodiments they can be separate units connected to each other.
- the transfer of the partial- volume corrected set of image data from the transformation unit to the processing unit is performed via a patch cable or a wireless communication system.
- the transformation unit and the processing unit are not connected to each other and the transfer of the partial- volume corrected set of image data from the transformation unit to the processing unit is performed via a data storage device that is configured to be connected to the transformation unit and to receive and store the partial- volume corrected set of image data, and to be connected and provide the stored partial- volume corrected set of image data to the processing unit.
- the first and second mutually registered sets of image data are acquired simultaneously or sequentially within a short time window. This allows reducing a potential mismatch between the two sets of image data caused by e.g.
- morphological variations of the object under measurement such as, but not limited to, a change of size or of orientation of the object.
- the processing unit is also configured to identify at least one landmark feature of the folded image feature using at least a section of the folded image feature and a set of pre-stored landmark template features.
- the landmark features included in the landmark template features represent certain features that are thought to be present in partial- volume corrected image feature regardless of its actual shape.
- the processing unit provides the mapped image data using the partial- volume corrected set of image data and the identified landmark features according to a predetermined mapping rule.
- the mapping rule transforms the partial- volume corrected image feature into an inflated image feature having the reference shape and further maps the internal contrast features onto the inflated image feature.
- PET positron emission tomography
- PET scans tend to appear blurred due to a number of reasons such as the limited resolution of the PET-image acquisition equipment, or in cases where the shape of the measured object is convoluted thus having a folded shape which cannot be fully resolved in the second set of image data due to the limited resolution of the measurement equipment.
- PET further provides information regarding tracer uptake and metabolic activity in an image feature in the form of internal contrast features.
- a partial- volume correction is performed using an alternative set of image data obtained by magnetic resonance imaging techniques or by computer tomography scan techniques. These alternative sets of image data represent an image having a higher spatial resolution than the image represented by the PET set of image data.
- a partial- volume corrected set of image data is obtained, which has a corrected spatial resolution higher than that of the PET set of image data.
- PET image data is an advantageous, but non- restrictive example and that the image data processing device of the present invention is applicable in a similar manner for image data obtained using other imaging modalities.
- another application case involves a statistical comparison of fractional anisotropy values obtained by a low-resolution diffusion magnetic resonance (MR) imaging modaility for white matter areas in the brain with reference data as imaged using a high-resolution anatomical MR modality, in order to identify areas with white matter lesions by detection of abnormal fractional anisotropy.
- the image data processing device is advantageously configured to provide a statistical comparison of fractional anisotropy values for white matter areas in the brain.
- high-resolution anatomical MR data is compared to reference data to identify areas with white matter lesions based on abnormal fractional anisotropy.
- Application of the present invention is not restricted to brain imaging.
- Another exemplary application case from the field cardiology is a statistical comparison of blood flow values from Dopper ultrasound data forming a low-resolution imaging modality in blood vessels to normative data which have been previously segmented in high resolution, for instance by MR imaging, in order to identify areas of abnormal blood flow.
- the present application case forms an example of an embodiment of the present invention that makes use of image data obtained by at least one non-tomographic image modality.
- the image data processing device is advantageously configured to process the first and the second mutually registered sets of image data wherein the first set of image data is magnetic resonance image data acquired by magnetic resonance imaging technique or computed tomography image data acquired by computed tomography scan technique, and wherein the second set of image data is positron emission tomography image data acquired by positron emission tomography imaging technique.
- embodiments are suitably configured to perform a partial- volume correction algorithms using magnetic resonance image data or computed tomography image data as a first set of image data and positron emission tomography image data as a second set of image data, to map the partial- volume corrected image feature onto a mapped image feature having a reference shape and to statistically compare the mapped image features to a plurality of sample mapped image features belonging to a statistical population of external objects of the object type.
- first and second sets of image data, the partial- volume corrected set of image data, the mapped image data sample image data, and the sample transformed image data comprise a respective plurality of voxels in a three dimensional coordinate grid, and wherein the processing unit is configured to identify the significant mapped internal contrast features voxel- wise.
- These sets of image data are divided in voxels that have a respective position in a three dimensional coordinate grid.
- Each voxel has a respective voxel value indicative of, for example, an amount of brightness or an amount of color component intensity.
- the voxels values of the voxels forming the mapped image data are then compared to those respective voxels of the sample image data to identify the significant mapped internal contrast features. In some embodiments only the voxels forming the mapped image feature are used for the identification of the significant mapped internal contrast features.
- the processing unit is further configured to provide inverse-mapped image data of the identified significant mapped internal contrast feature by applying to the mapped image data of the identified significant mapped internal contrast features a predetermined second mapping rule.
- This second mapping rule is inverse with respect to the mapping rule applied to the partial- volume corrected image data, so that the partial- volume corrected image feature can be reestablished from the mapped image data.
- the image data processing device of these embodiments further comprise an output user interface, which receives the inverse-mapped image data and is configured to provide a graphical representation of the inverse mapped image data for graphical output together with the partial- volume corrected image feature.
- the output user interface receives the mapped image data and is configured to provide a graphical representation of the mapped image data for graphical output together with the reference shape.
- a method for operating an image data processing device comprises:
- first and second mutually registered sets of image data representing measurement data acquired by different measurement techniques from an identical external object of an object type, the respective sets of image data containing at least one respective image feature with respective internal contrast features, wherein the first set of image data represents a first image having a higher spatial resolution than a second image represented by the second set of image data; an object having a folded shape and forming an folded image feature in view of adjacent surroundings external to the object, the folded image feature further containing internal contrast features;
- segmentation information assigning, based on at least one predetermined segmentation criterion, segmentation information to the image data of the first set of image data, the segmentation information representing a segmentation of the first image into at least two image segments identifying at least one landmark feature of the folded image feature using at least a section of the folded image feature and a set of pre-stored landmark template features;
- mapping rule transforming the partial- volume corrected image feature into a mapped image feature having a predetermined reference shape
- sample image data pertaining to a plurality of sample mapped image features from a statistical population of further external objects of the object type, the sample mapped image features having the reference shape;
- sample image data representing sample transformed objects that form a statistical population of sample transformed objects having the unfolded reference shape, each sample transformed object having respective sample mapped internal contrast features
- the method of the second aspect of the present invention shares the advantages of the image data processing device of the first aspect.
- performing a statistical comparison further comprises identifying significant mapped internal contrast features that differ from corresponding sample mapped internal contrast features of the statistical population in a statistically significant manner.
- output data is thus subsequently provided which is depending on to the identified significant mapped internal contrast features.
- the method further comprises, before determining the mapped image data, identifying and selecting image points of the partial- volume corrected image feature using a predetermined image data selection criterion, and determining the mapped image data using only the selected image data.
- inventions of the method of the second aspect further comprise providing inverse-mapped image data of the identified significant mapped internal contrast feature by applying to the mapped image data of the identified significant mapped internal contrast features a predetermined second mapping rule, which is inverse with respect to the mapping rule applied to the partial- volume corrected image data, and providing a graphical representation of the inverse mapped image data for graphical output together with the partial- volume corrected image feature
- a computer program is presented.
- the computer program is configured to control an image data processing device, the computer program comprising executable code for executing the method of the second aspect of the present invention or one of its embodiments when executed by a processor of a computer.
- the computer program of the third aspect shares the advantages of the method of the second aspect of the invention.
- Fig. 1 shows an exemplary block diagram of an embodiment of an image data processing device.
- Fig. 3a shows an exemplary reference shape with mapped internal contrast features.
- Fig. 3b shows the exemplary reference shape with other mapped internal contrast features.
- Fig. 5 shows another exemplary reference shape with mapped internal contrast features.
- Fig. 6 shows a flow diagram of an exemplary embodiment of a method for operating an image data processing device.
- Fig. 1 shows a schematic block diagram of an exemplary image data processing device 100.
- the description of this embodiment will be with reference to Figs, laic and Figs. 3a-3b which show respective images represented by different sets of image data as it will be explained in the following.
- the image data processing device comprises a transformation unit 102 which is configured to receive first 104 and second 106 sets of image data representing measurements acquired by different measurement techniques from an identical external object of an object type.
- Fig 2. a shows an image 200. a of an object represented by a first set of image data whereas Fig. 2b shows another image 200. b of the same object represented by a second set of image data. Both images contain a respective image feature 202. a and 202.b.
- Image 200.b also presents internal contrast features, symbolized in Fig. 2b by different filling patterns.
- Three exemplary internal contrast features are labeled 204.b, 206.b and 208.b in Fig. 2b.
- the first and the second sets of image data are acquired by different measurement techniques.
- the first and the second sets of image data are, in a particular embodiment, acquired by magnetic resonance imaging and by positron emission tomography imagining techniques respectively.
- the image represented in Fig. 2a has a higher spatial resolution that the image shown in Fig. 2b.
- a convoluted shape e.g. a shape having folds and grooves
- the object is represented in a more accurate manner by the image in Fig. 2a than by the image in Fig. 2b.
- the internal contrast features 204.b, 206.b and 208.b are distinguishable from each other and from adjacent surroundings based on an intensity range of a given color component (e.g. in a RGB color code), or on a brightness value range in a gray scale code.
- the image data comprises a plurality of pixels or voxels in a coordinate pixel or voxel grid, where each pixel or voxel has at least a respective pixel value or voxel value that is indicative of the intensity of a given color component or indicative of a brightness component in a gray scale.
- the transformation unit 102 performs a partial- volume correction algorithm using both sets of image data and provides a partial- volume corrected set of image data representing an image 200.
- c having a partial- volume corrected image feature 202. c as shown in Fig. 2c.
- Image 200. c has a higher spatial resolution than image 200.b, and shows internal contrast features (e.g. 204. c 206. c and 208. c) that correspond to the internal contrast features 204.b, 206.b and 208.b of the image feature 202. a.
- the image data processing device 100 further comprises a processing unit 108 that receives the partial- volume corrected set of image data 200. c and determines mapped image data 300. a as shown in Fig. 3a.
- the mapped image data pertains to the external object according to a predetermined mapping rule that transforms the partial- volume corrected image feature 202. c into a mapped image feature 302. a having a predetermined reference shape 304.
- the mapped image feature 302. a further contains mapped internal contrast features (e.g. 306, 308 and 310) that correspond to the internal contrast features 204. c, 206. c and 208. c of the partial- volume corrected image feature 202. c.
- the processing unit 108 further receives sample image data 110 pertaining to a plurality of sample mapped image features from a statistical population of further external objects of the object type.
- Fig. 3b shows exemplary mapped image data 300.b which represents a sample mapped image feature 302.b.
- the sample mapped image feature also has the reference shape 304.
- the fact that the mapped image feature 302. a and the sample image features 302.b share the same shape allows for a comparison of the internal contrast features in a statistical manner.
- the processing unit thus identifies, using the received mapped image data and the received sample image data, significant mapped internal contrast features 312 of the mapped image feature 302. a that differ in a statistically significant manner from corresponding sample mapped internal contrast features of the statistical population in a statistically significant manner. In some embodiments this statistical analysis is performed voxel-wise, by comparing each voxel value of the mapped image feature with the voxel values of the sample mapped image features of the sample image data 110.
- the sample image data represents a plurality of sample mapped image features that have the predetermined reference shape. They represent respective objects from a statistical population of objects of the object type.
- the image mapped image features and the sample mapped image features represent partial- volume corrected images of the brain cortex of a plurality of individuals, obtained by performing a partial- volume correction algorithm on a set of PET image data and a set of magnetic resonance image data.
- Both the mapped image data and the sample image data comprise image features in accordance with a predetermined reference shape, although the mapped internal contrast feature of the mapped image feature and the sample mapped internal contrast features obtained from the sample image data may differ.
- the processing unit 108 is also configured to identify significant mapped internal contrast features that differ from corresponding sample mapped internal contrast features of the statistical population in a statistically significant manner.
- the processing unit thus compares the mapped image feature with the sample mapped image feature in search of statistically relevant deviations from the sample image data.
- Some embodiments use the mean value and standard deviation of the sample image data and identify mapped internal contrast features of the mapped image data that lie outside a predetermined contrast range centered at the mean value and having a range width given by twice the standard deviation.
- Other embodiments use other statistical approaches to identify the significant mapped internal contrast features that differ from the corresponding sample mapped internal contrast features in a statistically relevant manner.
- the processing unit provides output data depending on the identified significant mapped internal contrast features.
- the output data comprises in some embodiments relative position data pertaining to the identified mapped internal contrast feature within the mapped image feature having the reference shape. In other embodiments, the output data comprises the relative approximate position data pertaining to the identified mapped internal contrast feature within the partial- volume corrected image feature. In some embodiments the output data further comprises intensity data pertaining to the statistical analysis, such us for example, intensity data pertaining to the intensity difference between the mapped internal contrast feature and the sample image data.
- the determination of mapped image data by the processing unit is performed by identifying landmark features of the partial- volume corrected image feature by using a set of pre-stored landmark template features.
- Fig. 4 shows an example of a partial- volume corrected set of image data 400.
- This set of image data represents an image feature 402 in view of adjacent surroundings 404.
- the image feature is distinguishable from the surroundings based on a contrast difference (e.g. different values in a gray or RGB color scale).
- the image feature 402 further contains internal contrast features. These are labeled in Fig. 4 as 406. a-c. These internal contrast features are also distinguishable based on a contrast difference (e.g.
- this contrast difference is based on the brightness values in a gray scale).
- the landmark features 408. a and 408. b included in the set of landmark template features represent certain features that are considered to be present in the partial- volume corrected image feature regardless of the actual shape (e.g. determined by folds and grooves) that the represented object presents.
- the transformation unit is configured to identify these landmarks features among the plurality of folds and grooves of the partial- volume corrected image feature based on the set of landmark template features and to determine mapped image data according to a predetermined mapping rule that is based on the identified landmark features 408. a and 408.b.
- These identified landmark features serve as a position reference for the internal contrast features 406. a-c.
- the image feature is divided into regions defined by the landmarks 408. a and 408.b as shown by the dotted lines in Fig. 4.
- Fig. 5 shows mapped image data 500 that is the result of a
- Mapped positions 506. a and 506.b of the identified landmark features 408. a and 408. b on the mapped image feature also define reference regions of the mapped image feature.
- the internal contrast features 406. a-c are mapped onto the mapped image feature that has the predetermined reference shape.
- the mapped internal contrast features are labeled as 504. a-c and correspond to the internal contrast features 406.
- the mapped image feature 502 is in accordance with a predetermined reference shape means that, regardless of the respective folded structure of the image data, i.e., the individual folds and grooves of the partial- volume corrected image feature, the mapped image feature will have the predetermined reference shape and a statistical analysis of the internal contrast features performed by comparing these to a sample set of image features, all having the same shape, is therefore possible.
- An exemplary embodiment of the image data processing device is advantageously configured to compare image data representing partial- volume corrected (PVC) positron emission tomography (PET) scans of an object that has a folded shape such as a brain or a brain cortex, with a sample image data also representing PVC-PET scans of other objects of the same object type.
- PET brain scans that are not partial- volume corrected appear blurred due to the limited resolution of the PET detector cells. This results in spill- in of gray matter PET activity into areas of white matter and vice versa.
- PVC partial- volume correction
- MRI magnetic resonance imaging
- the processing input unit of this embodiment receives partial- volume corrected PET brain scans as a partial- volume corrected set of image data.
- This set of image data shows the cerebral cortex region, i.e. the region of interest, as a region of pixels having a distinguishable pixel values in comparison with the surrounding regions (white matter, cerebrospinal fluid, etc.).
- the processing unit identifies landmark positions in the cerebral cortex according to predetermined landmark template feature.
- the landmarks are thus features in the cerebral cortex which can be found in almost every brain independently of the individual gyri and sulci. According to some models, the landmarks define structural regions of the brain.
- the Midboggle-101 dataset serves as a normative dataset to establish morphometric variation in a healthy population for comparison against clinical populations and uses the Desikan-Killiany-Tourville (DKT) protocol to improve the consistency and accuracy of labeling human cortical areas.
- DKT Desikan-Killiany-Tourville
- Some of these structural regions are labeled according to some models as superior frontal, rostral middle frontal, triangularis, lateral orbito frontal, transverse temporal, etc.
- These structural regions are defined by brain features or landmarks (mostly sulci) which are present in most brains.
- the processing unit determines mapped image data representing a mapped image feature onto which the internal contrast features are mapped.
- the mapped image data can be then directly compared in a statistical manner with sample image data.
- the sample image data comprises mapped representations of sample brains which are considered to share a predetermined characteristic such as, for example, being regarded as representation of healthy brains not showing any medical condition,. Nevertheless, the sample image data shows certain differences in the form of different sample mapped internal contrast features. These differences are due to the fact that the sample image data represents a plurality of sample transformed objects that are not identical to each other. The sample image data can be thus regarded as registered set of reference scans.
- the processing unit is configured to identify mapped internal contrast features that differ from corresponding sample mapped internal contrast features of the statistical population in a statistically significant manner.
- Fig. 6 shows show a flow diagram of an exemplary method 600 for operating an image data processing device.
- the method 600 comprises receiving in a step 602 first and second mutually registered sets of image data representing measurement data from an identical external object of an object type. These two sets of image data are acquired by different measurement techniques.
- the respective sets of image data contain at least one respective image feature with respective internal contrast features.
- the first set of image data represents a first image having a higher spatial resolution than a second image represented by the second set of image data.
- the method also receives, in a step 612 sample image data pertaining to a plurality of sample mapped image features from a statistical population of further external objects of the object type.
- the sample mapped image features have also the reference shape.
- the method identifies, using the received mapped image data and the received sample image data, significant mapped internal contrast features of the mapped image feature that differ in a statistically significant manner from corresponding sample mapped internal contrast features of the statistical population, and finally, in a step 616, the method provides output data depending on the identified significant mapped internal contrast features.
- a simple embodiment of a data processing device thus comprises a
- transformation unit configured to receive two distinct sets of image data containing respective image features and pertaining to an identical external object, and to perform a partial volume correction algorithm using both sets of image data to obtain a partial volume corrected set of image data containing a partial- volume corrected image feature and. It further comprises a processing unit configured to determine mapped image data based on the partial- volume corrected set of image data, to receive sample mapped image data, to identify significant mapped internal contrast features of the mapped image feature that differ in a statistically significant matter form corresponding mapped internal contrast features of the statistical population and to provide output information that depends on the identified significant mapped internal contrast features.
- processing unit can be provided as a stand-alone device.
- processing unit which is configured to:
- mapping rule transforming a partial-volume corrected image feature comprised by the partial- volume corrected set of image data into a mapped image feature having a predetermined reference shape
Landscapes
- Engineering & Computer Science (AREA)
- Quality & Reliability (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
- Image Processing (AREA)
Abstract
La présente invention concerne un dispositif de traitement de données d'image (100) consistant en une unité de transformation (102) configurée pour recevoir deux ensembles distincts de données d'image (104, 106) comportant des caractéristiques d'image respectives et ayant trait à un objet externe identique et pour effectuer un algorithme de correction de volume partiel à l'aide des deux ensembles de données d'image en vue d'obtenir un ensemble corrigé de volume partiel de données d'image comportant une caractéristique d'image corrigée de volume partiel. Elle consiste en outre en une unité de traitement (108) configurée pour déterminer des données d'image mappées en fonction de l'ensemble corrigé de volume partiel des données d'image, en vue de recevoir des données d'image échantillons mappées (110), d'identifier des caractéristiques de contraste interne mappées significatives de la caractéristique d'image mappée qui diffèrent dans une forme de substance significative statistiquement correspondant à des caractéristiques de contraste interne mappées de la population statistique et de fournir des informations de sortie qui dépendent des caractéristiques de contraste interne mappées significatives identifiées.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201662336772P | 2016-05-16 | 2016-05-16 | |
US62/336772 | 2016-05-16 | ||
EP16173332.4 | 2016-06-07 | ||
EP16173332 | 2016-06-07 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017198518A1 true WO2017198518A1 (fr) | 2017-11-23 |
Family
ID=56117542
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2017/061246 WO2017198518A1 (fr) | 2016-05-16 | 2017-05-11 | Dispositif de traitement de données d'image |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2017198518A1 (fr) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021154671A1 (fr) * | 2020-01-29 | 2021-08-05 | University Of Cincinnati | Procédé d'amélioration de contraste destiné aux objets à effet de volume partiel intrinsèque |
CN118654590A (zh) * | 2024-06-04 | 2024-09-17 | 广州鲸展科技有限公司 | 加油枪体变形程度的云计算鉴定系统 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2009065079A2 (fr) * | 2007-11-14 | 2009-05-22 | The Regents Of The University Of California | Alignement longitudinal de l'anatomie en imagerie par résonance magnétique |
US20090164132A1 (en) * | 2007-12-13 | 2009-06-25 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Methods and systems for comparing media content |
WO2009134820A2 (fr) * | 2008-04-28 | 2009-11-05 | Cornell University | Outil pour une quantification précise dans une imagerie par résonance magnétique moléculaire |
US20110199084A1 (en) * | 2010-02-16 | 2011-08-18 | Board Of Regents Of The University Of Texas System | Method and system for diffusion tensor imaging |
WO2014139024A1 (fr) * | 2013-03-15 | 2014-09-18 | Synaptive Medical (Barbados) Inc. | Systèmes de planification, de navigation et de simulation, et procédés de thérapie mini-invasive |
-
2017
- 2017-05-11 WO PCT/EP2017/061246 patent/WO2017198518A1/fr active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2009065079A2 (fr) * | 2007-11-14 | 2009-05-22 | The Regents Of The University Of California | Alignement longitudinal de l'anatomie en imagerie par résonance magnétique |
US20090164132A1 (en) * | 2007-12-13 | 2009-06-25 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Methods and systems for comparing media content |
WO2009134820A2 (fr) * | 2008-04-28 | 2009-11-05 | Cornell University | Outil pour une quantification précise dans une imagerie par résonance magnétique moléculaire |
US20110199084A1 (en) * | 2010-02-16 | 2011-08-18 | Board Of Regents Of The University Of Texas System | Method and system for diffusion tensor imaging |
WO2014139024A1 (fr) * | 2013-03-15 | 2014-09-18 | Synaptive Medical (Barbados) Inc. | Systèmes de planification, de navigation et de simulation, et procédés de thérapie mini-invasive |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021154671A1 (fr) * | 2020-01-29 | 2021-08-05 | University Of Cincinnati | Procédé d'amélioration de contraste destiné aux objets à effet de volume partiel intrinsèque |
CN118654590A (zh) * | 2024-06-04 | 2024-09-17 | 广州鲸展科技有限公司 | 加油枪体变形程度的云计算鉴定系统 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7118606B2 (ja) | 医用画像処理装置及び医用画像処理プログラム | |
Lohmann et al. | LIPSIA—a new software system for the evaluation of functional magnetic resonance images of the human brain | |
US8837791B2 (en) | Feature location method and system | |
CN106133790B (zh) | 用于在组织种类分离的帮助下基于磁共振图像生成一幅或多幅计算机断层摄影图像的方法和设备 | |
US7995864B2 (en) | Method and system for performing image registration | |
US20150131882A1 (en) | Medical image data processing apparatus and method | |
US9361686B2 (en) | Method and apparatus for the assessment of medical images | |
US20130044927A1 (en) | Image processing method and system | |
US20090034812A1 (en) | Superimposing brain atlas images and brain images with delineation of infarct and penumbra for stroke diagnosis | |
Khalifé et al. | Subject-specific bone attenuation correction for brain PET/MR: can ZTE-MRI substitute CT scan accurately? | |
Bardera et al. | Semi-automated method for brain hematoma and edema quantification using computed tomography | |
US20100128841A1 (en) | Smoothing of Dynamic Data Sets | |
RU2752690C2 (ru) | Обнаружение изменений на медицинских изображениях | |
Calmon et al. | Automatic measurement of changes in brain volume on consecutive 3D MR images by segmentation propagation | |
CN112950648B (zh) | 确定磁共振图像中的正中矢状平面的方法和设备 | |
US20180064409A1 (en) | Simultaneously displaying medical images | |
De Silva et al. | Registration of MRI to intraoperative radiographs for target localization in spinal interventions | |
Sun et al. | Automated template-based PET region of interest analyses in the aging brain | |
Alam et al. | Evaluation of medical image registration techniques based on nature and domain of the transformation | |
Liu et al. | Rigid model‐based 3D segmentation of the bones of joints in MR and CT images for motion analysis | |
WO2017198518A1 (fr) | Dispositif de traitement de données d'image | |
Nowinski et al. | Rapid and automatic calculation of the midsagittal plane in magnetic resonance diffusion and perfusion images | |
Collins et al. | Use of registration for cohort studies | |
Liu et al. | Employing symmetry features for automatic misalignment correction in neuroimages | |
EP1969563A2 (fr) | Procede d'inspection croisee dans le temps pour diagnostic medical |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17723073 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17723073 Country of ref document: EP Kind code of ref document: A1 |