WO2017051943A1 - 영상 생성 방법 및 장치, 및 영상 분석 방법 - Google Patents
영상 생성 방법 및 장치, 및 영상 분석 방법 Download PDFInfo
- Publication number
- WO2017051943A1 WO2017051943A1 PCT/KR2015/010085 KR2015010085W WO2017051943A1 WO 2017051943 A1 WO2017051943 A1 WO 2017051943A1 KR 2015010085 W KR2015010085 W KR 2015010085W WO 2017051943 A1 WO2017051943 A1 WO 2017051943A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- noise
- window
- learning
- reference image
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 57
- 238000003703 image analysis method Methods 0.000 title abstract description 6
- 238000012549 training Methods 0.000 claims description 74
- 238000013528 artificial neural network Methods 0.000 claims description 62
- 238000010586 diagram Methods 0.000 description 16
- 210000000988 bone and bone Anatomy 0.000 description 11
- 210000004369 blood Anatomy 0.000 description 7
- 239000008280 blood Substances 0.000 description 7
- 210000000056 organ Anatomy 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 239000000284 extract Substances 0.000 description 5
- 206010028980 Neoplasm Diseases 0.000 description 4
- 238000002591 computed tomography Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 3
- 238000004422 calculation algorithm Methods 0.000 description 2
- 201000010099 disease Diseases 0.000 description 2
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 2
- 230000003902 lesion Effects 0.000 description 2
- 238000002595 magnetic resonance imaging Methods 0.000 description 2
- 230000003278 mimic effect Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 210000002569 neuron Anatomy 0.000 description 2
- 238000002600 positron emission tomography Methods 0.000 description 2
- 210000001519 tissue Anatomy 0.000 description 2
- 210000001015 abdomen Anatomy 0.000 description 1
- 238000010521 absorption reaction Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000001149 cognitive effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000013501 data transformation Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 210000004072 lung Anatomy 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012567 pattern recognition method Methods 0.000 description 1
- 210000004872 soft tissue Anatomy 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24133—Distances to prototypes
- G06F18/24137—Distances to cluster centroïds
- G06F18/2414—Smoothing the distance, e.g. radial basis function networks [RBFN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/77—Retouching; Inpainting; Scratch removal
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
Definitions
- the following description relates to an image generating method and apparatus, and an image analyzing method, and more specifically, to a method and apparatus for generating a training image used for neural network learning, and an input image using a neural network learned through the generated training image. It is about how to analyze.
- Such an artificial neural network includes a relatively large number of layers. In order to learn a large-structured artificial neural network including many layers, a large amount of training data is required, and it is not necessary to overfit the specific training data.
- An image generating method includes receiving a reference image; And adding a noise to at least one parameter of a window width and a window level for pixel values of the reference image to generate a learning image from the reference image.
- the generating of the learning image may include adding the noise added parameter and the noise when there are remaining parameters in which the noise is not added among the window width and the window level. Based on the remaining parameters, a learning image may be generated from the reference image.
- the window width and the window level may include preset values for an object to be analyzed by a neural network trained based on the training image.
- the window width may indicate a range of pixel values to be included in the learning image among pixel values included in the reference image.
- the window level may represent a center of a range of pixel values included in the training image.
- the reference image may be a medical image photographing an object to be analyzed by a neural network trained based on the training image.
- the generating of the training image may include generating a window width and a window level of the window width and the window level such that the window width and the window level deviate from a preset value for the object of the neural network trained based on the training image. It is possible to modify the value for at least one parameter.
- the image generating method may further include adding noise to pixel values of the training image.
- the noise added to the pixel value of the training image may be generated based on at least one of a characteristic of a device photographing the reference image and an object included in the reference image.
- an image analysis method includes: receiving an input image; And analyzing the input image based on a neural network, wherein the neural network is trained based on a learning image extracted from a reference image, and the learning image includes a window width and a window for pixel values of the reference image. Noise is added to at least one parameter of the level to generate the reference image.
- An image generating apparatus may include a memory in which an image generating method is stored; And a processor executing the image generating method, wherein the processor generates a learning image from the reference image by adding noise to at least one parameter of a window width and a window level for pixel values of the reference image.
- a learning image to which natural noise is applied may be obtained, improving a learning effect of a neural network to be learned, and being robust to various changes. You can expect a robust neural network.
- noise is added to at least one parameter of a window width and a window level used when extracting a training image from a reference image, thereby effectively modifying the training image used for neural network learning and increasing the amount of the training image. Can be increased.
- FIG. 1 is a diagram illustrating a method of generating an image, according to an exemplary embodiment.
- FIG. 2 is a diagram for describing a window width and a window level, according to an exemplary embodiment.
- FIG. 3 is a diagram illustrating an example in which noise is added to a window width according to an embodiment.
- FIG. 4 is a diagram illustrating an example in which noise is added to a window level according to an embodiment.
- FIG. 5 is a diagram illustrating an example in which noise is added to a window width and a window level according to an embodiment.
- FIG. 6 is a diagram illustrating a method of generating an image, according to another exemplary embodiment.
- FIG. 7 is a diagram illustrating an image generating apparatus, according to an exemplary embodiment.
- FIG. 8 is a diagram illustrating an image analysis method, according to an exemplary embodiment.
- FIG. 1 is a diagram illustrating a method of generating an image, according to an exemplary embodiment.
- the image generating method may be performed by a processor included in the image generating apparatus.
- the image generating apparatus may be widely used in an area for generating training data (eg, a training image) for training a neural network for analyzing (eg, recognizing, classifying, detecting, etc.) an input image.
- Neural networks are cognitive models implemented in software or hardware that mimic the computational power of biological systems using a large number of artificial neurons connected by connecting lines.
- the neural network may include a plurality of layers.
- the neural network may include an input layer, a hidden layer, and an output layer.
- the input layer may receive an input for performing learning (eg, training data) and transmit the input to the hidden layer, and the output layer may generate an output of the neural network based on signals received from nodes of the hidden layer.
- the hidden layer is located between the input layer and the output layer, and can change the training data transmitted through the input layer to a predictable value.
- the neural network may include a plurality of hidden layers.
- a neural network including a plurality of hidden layers is called a deep neural network, and learning a deep neural network is called deep learning.
- the training image generated by the image generating apparatus may be input to a neural network to be trained.
- the image generating apparatus may variously modify data input to the neural network by applying random noise to the training image.
- the neural networks can be trained to have noise-resistant characteristics without overfitting the specific training images.
- a process of generating a learning image by the image generating apparatus using random noise will be described later.
- the image generating apparatus receives a reference image.
- the image generating apparatus may receive a reference image from an external device through a built-in sensor or a network.
- the reference image is a medical image of an object (eg, bones, organs, blood, etc.) to be analyzed by the neural network, and may be composed of, for example, pixels having a value of 12 bits. While the reference image includes a pixel value of 12 bits, a general display device may express an 8 bit pixel value, and thus the reference image is difficult to be displayed on the display device. Therefore, in order to visualize the medical image on the display device, it is necessary to convert the 12-bit reference image into an 8-bit image (or an image of 8 bits or less).
- the image generating apparatus may convert the reference image into an image that can be visualized by limiting the range of pixel values of the reference image to be displayed on the display device and determining the center of the range of pixel values to be expressed.
- the range of pixel values to be represented refers to a window width, and the center of the range of pixel values to be represented may mean a window level.
- the image generating apparatus generates a learning image from the reference image by adding noise to at least one parameter of a window width and a window level for pixel values of the reference image.
- Noise is added to at least one parameter of a window width and a window level for pixel values of the reference image.
- the window width and the window level may refer to parameters used when the image generating apparatus generates the training image from the reference image.
- the image generating apparatus may add noise to at least one parameter of the window width and the window level.
- the image generating apparatus may add noise to both the window width and the window level, or may add noise to only one of the window width and the window level.
- a more specific embodiment of adding noise to at least one parameter of the window width and the window level will be described later with reference to FIGS. 2 to 5.
- the image generating apparatus may generate a learning image from the reference image based on the parameter to which the noise is added.
- the parameter to which the noise is added may mean a window width and a window level.
- the image generating apparatus may generate a training image from the reference image based on the parameter to which the noise is added and the remaining parameter to which the noise is not added. have.
- the image generating apparatus may generate a learning image from the reference image based on the parameters and the remaining parameters.
- the parameter may mean that noise of the window width and the window level is added
- the remaining parameters may mean that noise of the window width and the window level is not added.
- FIG. 2 is a diagram for describing a window width and a window level, according to an exemplary embodiment.
- the reference image is a medical image of an object to be analyzed by the neural network.
- the reference image is photographed through various techniques such as magnetic resonance imaging (MRI), computed tomography (CT), X-ray, and positron emission tomography (PET). It may include a video.
- MRI magnetic resonance imaging
- CT computed tomography
- PET positron emission tomography
- the reference image corresponds to a gray-scale image and may have a pixel value of 12 bits.
- One pixel constituting the reference image may have a value of about 4000 steps, which is outside the range (eg, 8 bits) that the pixels of the general image can represent.
- the reference image may include a hounsfield unit (HU) value.
- HU may indicate the extent to which the X-ray is absorbed by the body according to the difference in density of the tissue through which the X-ray penetrates the body.
- the HU can be obtained by setting the water to 0 HU, the bone to 1000 HU, the lowest absorbing air to -1000 HU, and calculating the relative linear attenuation coefficient based on the relative X-ray absorption of each tissue.
- HU may also be referred to as CT Number.
- A may represent -1000 HU, which is the lowest HU value that the reference image may have, and B may represent +3000 HU, which is the maximum HU value that the reference image may have.
- the human eye cannot recognize all 12-bit pixel values included in the reference image. Therefore, the reference image needs to be converted into an 8-bit image that can be recognized by the human eye. To this end, it is possible to limit the range of the HU to be expressed in the reference image, and determine the center of the range of the HU to be expressed, wherein the range of the HU to be expressed represents the window width 210 and is intended to be expressed. The center of the range of HU is the window level 220.
- the window width 210 and the window level 220 may be predetermined according to an object to be analyzed by the neural network. For example, when the object to be analyzed by the neural network is the soft tissue of the abdomen, the window width 210 may be determined as 350 to 400 HU, and the window level 220 may be determined as 50 HU. In addition, when the object to be analyzed by the neural network is a lung, the window width 210 may be determined as 1500 to 1600 HU, and the window level 220 may be determined as ⁇ 700 HU. In this case, the specific window width 210 and the window level 220 may be set to the HU value input by the user or to the HU value determined by receiving N points for the object to be analyzed from the user. .
- the image generating apparatus may add noise to at least one parameter of the window width 210 and the window level 220, and generate a learning image from the reference image using the parameter to which the noise is added. have.
- the image generating apparatus may generate various learning images for learning the neural network, and by being learned through various learning images, the neural network may have characteristics that are robust to noise without overfitting the specific learning image.
- FIG. 3 is a diagram illustrating an example in which noise is added to a window width according to an embodiment.
- noise-added window widths 310-1, 310-2, and 310-3 have various ranges, while the window level 320 without added noise may have a single value.
- the first window width 310-1 shown in FIG. 3 may have a narrower range than the second window width 310-2 and the third window width 310-3.
- the training image extracted through the first window width 310-1 and the window level 320 may be expressed more than the training image extracted using the second window width 310-2 or the third window width 310-3.
- the range of possible pixel values may be narrow.
- the training image extracted through the third window width 310-3 and the window level 320 is the training image extracted using the first window width 310-1 or the second window width 310-2.
- the range of pixel values that can be more represented can be wide.
- the training extracted through the second window width 310-2 is performed.
- the image may more clearly represent bone than the learning image extracted by using the first window width 310-1 or the third window width 310-3.
- the training image extracted through the first window width 310-1 may include only a part of the bone, not the whole portion.
- the training image extracted through the third window width 310-3 may include other parts of the body including bones. It may also be included.
- the image generating apparatus may generate a training image to which natural noise is applied by extracting the training image through various window widths 310-1, 310-2, and 310-3 to which noise is added.
- FIG. 4 is a diagram illustrating an example in which noise is added to a window level according to an embodiment.
- noise-added window levels 420-1, 420-2, and 420-3 have various values, while the window width 410 without the noise can have the same size range.
- the first window level 420-1 illustrated in FIG. 4 may include a value larger than the second window level 420-2 and smaller than the third window level 420-3.
- the second window level 420-2 may have a smaller value than the first window level 420-1
- the third window level 420-3 may have a larger value than the first window level 420-1. have.
- some common HU ranges are included in the learning image extracted from the reference image using the first window level 420-1 and the learning image extracted from the reference image using the second window level 420-2. As such, there may be a part expressed in common in the extracted learning images.
- the training image extracted using the third window level 420-3 has a HU range in common with the training image extracted using the first window level 420-1 or the second window level 420-2. Since there is no, a part commonly expressed in the extracted learning images may not exist.
- the image generating apparatus may generate a training image to which natural noise is applied by extracting the training image using various window levels 420-1, 420-2, and 420-3 to which noise is added.
- FIG. 5 is a diagram illustrating an example in which noise is added to a window width and a window level according to an embodiment.
- noise-added window widths 510-1, 510-2, and 510-3 have various ranges, and the noise-added window levels 520-1, 520-2, and 520-3 can have various values. have.
- the window widths 510-1, 510-2, and 510-3 in which the noise of FIG. 5 is added may include the second window width 510-2, the first window width 510-1, and the third window width 510- 3) the window levels 520-1, 520-2, and 520-3 increase in order, and the second window width 510-2, the first window width 510-1, and the third window width It may have a value increasing in the order (510-3).
- the training image extracted through the first window width 510-1 and the first window level 520-1, the second window width 510-2, and the second window level 520-2 are determined. Commonly expressed parts may not exist between the extracted learning images.
- the training image is extracted through the first window width 510-1 and the first window level 520-1, and is extracted through the third window width 510-3 and the third window level 520-3. There may be a part expressed in common among the learning images.
- the image generating apparatus extracts a training image through various window widths 510-1, 510-2, and 510-3 and window levels 520-1, 520-2, and 520-3 to which noise is added, thereby providing natural noise. Can generate a learning image applied.
- An embodiment of adding noise to at least one parameter of the window width and the window level described with reference to FIGS. 3 to 5 may be variously changed according to a design.
- FIG. 6 is a diagram illustrating a method of generating an image, according to another exemplary embodiment.
- the image generating method according to an embodiment may be performed by a processor included in the image generating apparatus.
- the image generating apparatus receives a reference image.
- the reference image is a medical image of an object (eg, bone, organ, blood, etc.) to be analyzed by the neural network, and may be composed of, for example, pixels having a value of 12 bits.
- the image generating apparatus generates a learning image from the reference image by adding noise to at least one parameter of a window width and a window level for pixel values of the reference image.
- the window width and the window level may refer to parameters used by the image generating apparatus to generate a learning image from the reference image.
- the image generating apparatus generates a training image from the reference image using the parameter to which the noise is added. For example, when noise is added to both the window width and the window level, the image generating apparatus may extract the training image from the reference image based on the parameter to which the noise is added.
- the parameter to which the noise is added may mean a window width and a window level.
- the image generating apparatus may extract the training image from the reference image based on the parameter to which the noise is added and the remaining parameter to which the noise is not added. have.
- the image generating apparatus may generate a learning image from the reference image based on the parameters and the remaining parameters.
- the image generating apparatus may add noise to pixel values of the training image.
- the training image generated in operation 620 may be an image generated from a reference image using a parameter added with noise, and may be an image without noise added to the pixel value itself.
- the image generating apparatus may add random noise to the pixel value of the training image generated in operation 620.
- the image generating apparatus may generate a noise pattern based on the characteristics of the device photographing the reference image, and add the generated noise pattern to the pixel value of the training image. For example, the image generating apparatus may identify the device through the information of the device photographing the reference image, and generate a noise pattern based on the identified device.
- the device that photographed the reference image refers to a medical device for photographing an object using various techniques such as MRI, CT, X-ray, PET, etc.
- the characteristics of the device may include information about the manufacturer of the device. Can be.
- the image generating apparatus may generate a noise pattern based on an object included in the reference image, and add the generated noise pattern to the pixel value of the training image. For example, the image generating apparatus may generate a noise pattern based on whether the object included in the reference image is bone, organ, blood, or tumor. In addition, the image generating apparatus may generate a noise pattern based on the shape of bone, organ, blood, or tumor.
- the image generating apparatus may train the neural network based on the training image.
- the learning image is an image extracted from the reference image using a parameter to which noise is added, and may include noise in pixel values.
- FIG. 7 is a diagram illustrating an image generating apparatus, according to an exemplary embodiment.
- the image generating apparatus 700 includes a memory 710 and a processor 720.
- the image generating apparatus 700 may be widely used in an area for generating training data (eg, a learning image) for learning a neural network for analyzing (eg, recognizing, classifying, or detecting) an input image.
- the image generating apparatus 700 may be mounted on various computing devices and / or systems, such as a smart phone, a tablet computer, a laptop computer, a desktop computer, a television, a wearable device, a security system, and a smart home system.
- the memory 710 stores an image generating method.
- the image generating method stored in the memory 710 relates to a method of generating a training image for learning a neural network, and may be executed by the processor 720.
- the memory 710 may store a training image generated by the processor 720 or a neural network learned based on the generated training image.
- the processor 720 executes an image generating method.
- the processor 720 adds noise to at least one parameter of a window width and a window level for pixel values of the reference image.
- the window width and the window level may refer to parameters used when the processor 720 generates the learning image from the reference image.
- the processor 720 generates a training image from the reference image by using the parameter to which noise is added. For example, when noise is added to both the window width and the window level, the processor 720 may extract the training image from the reference image based on the parameter to which the noise is added. In this case, the parameter to which the noise is added may mean a window width and a window level.
- the processor 720 may extract the training image from the reference image based on the parameter to which the noise is added and the remaining parameter to which the noise is not added. Can be. In other words, when there is a remaining parameter without adding noise among the window width and the window level, the processor 720 may generate a learning image from the reference image based on the parameter and the remaining parameter.
- the processor 720 stores the training image extracted from the reference image in the memory 710 or stores the learned neural network based on the extracted training image in the memory 710. ) Can be stored.
- the processor 720 may add noise to the pixel value of the training image based on at least one of a characteristic of a device photographing the reference image and an object included in the reference image. have.
- the processor 720 may generate a noise pattern based on the characteristics of the device photographing the reference image, and add the generated noise pattern to the pixel value of the training image. For example, the processor 720 may identify the device through the information of the device capturing the reference image, and generate a noise pattern based on the identified device.
- the processor 720 may generate a noise pattern based on an object included in the reference image, and add the generated noise pattern to the pixel value of the training image. For example, the processor 720 may generate a noise pattern based on whether an object included in the reference image is a bone, an organ, blood, or a tumor. In addition, the processor 720 may generate a noise pattern based on the shape of bone, organ, blood, or tumor.
- the processor 720 may store the generated learning image in the memory 710.
- the processor 720 may train the neural network based on the training image.
- the learning image is an image extracted from the reference image using a parameter to which noise is added, and may include noise in pixel values.
- the processor 720 may store the learned neural network in the memory 710. For example, the processor 720 may store the parameters for the learned neural network in the memory 710.
- FIG. 8 is a diagram illustrating an image analysis method, according to an exemplary embodiment.
- the image analysis method may be performed by a processor included in the image analysis apparatus.
- the image analyzing apparatus receives an input image.
- the input image may be a medical image including an object (eg, bone, organ, blood, etc.) to be analyzed.
- the image analyzing apparatus may receive an input image from an external device through a built-in sensor or a network.
- the image analyzing apparatus analyzes the input image based on the neural network.
- the neural network is a previously learned neural network and may be trained based on a learning image extracted from a reference image.
- the learning image may be generated from the reference image by adding noise to at least one parameter of a window width and a window level for pixel values of the reference image.
- the image analyzing apparatus may classify the input image using a neural network. For example, the image analyzing apparatus may classify an input image including an object into a specific disease or check the progress of the disease based on a neural network. Alternatively, the image analyzing apparatus may detect a lesion included in the input image using a neural network. In this case, the neural network may be learned based on various medical images including a lesion.
- the embodiments described above may be implemented as hardware components, software components, and / or combinations of hardware components and software components.
- the devices, methods, and components described in the embodiments may include, for example, processors, controllers, arithmetic logic units (ALUs), digital signal processors, microcomputers, field programmable gates (FPGAs). It may be implemented using one or more general purpose or special purpose computers, such as an array, a programmable logic unit (PLU), a microprocessor, or any other device capable of executing and responding to instructions.
- ALU arithmetic logic unit
- FPGAs field programmable gates
- PLU programmable logic unit
- one processing device may be described as being used, but one of ordinary skill in the art will appreciate that the processing device includes a plurality of processing elements and / or a plurality of types of processing elements. It can be seen that it may include.
- the processing device may include a plurality of processors or one processor and one controller.
- other processing configurations are possible, such as parallel
- the software may include a computer program, code, instructions, or a combination of one or more of the above, and configure the processing device to operate as desired, or process it independently or collectively. You can command the device.
- the software may be distributed over networked computer systems so that they may be stored or executed in a distributed manner. Software and data may be stored on one or more computer readable recording media.
- the method according to the embodiment may be embodied in the form of program instructions that can be executed by various computer means and recorded in a computer readable medium.
- the computer readable medium may include program instructions, data files, data structures, etc. alone or in combination.
- Examples of computer-readable recording media include magnetic media such as hard disks, floppy disks, and magnetic tape, optical media such as CD-ROMs, DVDs, and magnetic disks, such as floppy disks.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Databases & Information Systems (AREA)
- Computing Systems (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Quality & Reliability (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims (19)
- 기준 영상을 수신하는 단계; 및상기 기준 영상의 픽셀 값들에 대한 윈도우 폭 및 윈도우 레벨 중 적어도 하나의 파라미터에 노이즈를 추가하여 상기 기준 영상으로부터 학습 영상을 생성하는 단계를 포함하는 영상 생성 방법.
- 제1항에 있어서,상기 학습 영상을 생성하는 단계는,상기 윈도우 폭 및 윈도우 레벨 중 상기 노이즈가 추가되지 않은 나머지 파라미터가 존재하는 경우, 상기 노이즈가 추가된 파라미터와 상기 노이즈가 추가되지 않은 나머지 파라미터에 기초하여 상기 기준 영상으로부터 학습 영상을 생성하는, 영상 생성 방법.
- 제1항에 있어서,상기 윈도우 폭 및 상기 윈도우 레벨은,상기 학습 영상에 기초하여 학습되는 신경망이 분석하고자 하는 오브젝트에 대해 미리 설정된 값을 포함하는, 영상 생성 방법.
- 제1항에 있어서,상기 윈도우 폭은,상기 기준 영상에 포함된 픽셀 값들 중에서 상기 학습 영상에 포함시키고자 하는 픽셀 값들의 범위를 나타내는, 영상 생성 방법.
- 제1항에 있어서,상기 윈도우 레벨은,상기 학습 영상에 포함되는 픽셀 값들의 범위에 대한 중심을 나타내는, 영상 생성 방법.
- 제1항에 있어서,상기 기준 영상은,상기 학습 영상에 기초하여 학습되는 신경망이 분석하고자 하는 오브젝트를 촬영한 의료 영상인, 영상 생성 방법.
- 제1항에 있어서,상기 학습 영상을 생성하는 단계는,윈도우 폭 및 윈도우 레벨이 상기 학습 영상에 기초하여 학습되는 신경망의 오브젝트에 대해 미리 설정된 값에서 벗어나도록 상기 윈도우 폭 및 윈도우 레벨 중 적어도 하나의 파라미터에 대한 값을 변형시키는, 영상 생성 방법.
- 제1항에 있어서,상기 학습 영상의 픽셀 값에 노이즈를 추가하는 단계를 더 포함하는, 영상 생성 방법.
- 제8항에 있어서,상기 학습 영상의 픽셀 값에 추가된 노이즈는,상기 기준 영상을 촬영한 기기의 특성 및 상기 기준 영상에 포함된 오브젝트 중 적어도 하나에 기반하여 생성되는, 영상 생성 방법.
- 입력 영상을 수신하는 단계; 및신경망에 기초하여 상기 입력 영상을 분석하는 단계를 포함하고,상기 신경망은, 기준 영상으로부터 추출된 학습 영상에 기초하여 학습되고,상기 학습 영상은,상기 기준 영상의 픽셀 값들에 대한 윈도우 폭 및 윈도우 레벨 중 적어도 하나의 파라미터에 노이즈를 추가하여 상기 기준 영상으로부터 생성되는, 영상 분석 방법.
- 영상 생성 방법이 저장된 메모리; 및상기 영상 생성 방법을 실행하는 프로세서를 포함하고,상기 프로세서는,기준 영상의 픽셀 값들에 대한 윈도우 폭 및 윈도우 레벨 중 적어도 하나의 파라미터에 노이즈를 추가하여 상기 기준 영상으로부터 학습 영상을 생성하는 영상 생성 장치.
- 제11항에 있어서,상기 프로세서는,상기 윈도우 폭 및 윈도우 레벨 중 상기 노이즈가 추가되지 않은 나머지 파라미터가 존재하는 경우, 상기 노이즈가 추가된 파라미터와 상기 노이즈가 추가되지 않은 나머지 파라미터에 기초하여 상기 기준 영상으로부터 학습 영상을 생성하는, 영상 생성 장치.
- 제11항에 있어서,상기 윈도우 폭 및 상기 윈도우 레벨은,상기 학습 영상에 기초하여 학습되는 신경망이 분석하고자 하는 오브젝트에 대해 미리 설정된 값을 포함하는, 영상 생성 장치.
- 제11항에 있어서,상기 윈도우 폭은,상기 기준 영상에 포함된 픽셀 값들 중에서 상기 학습 영상에 포함시키고자 하는 픽셀 값들의 범위를 나타내는, 영상 생성 장치.
- 제11항에 있어서,상기 윈도우 레벨은,상기 학습 영상에 포함되는 픽셀 값들의 범위에 대한 중심을 나타내는, 영상 생성 장치.
- 제11항에 있어서,상기 기준 영상은,상기 학습 영상에 기초하여 학습되는 신경망이 분석하고자 하는 오브젝트를 촬영한 의료 영상인, 영상 생성 장치.
- 제11항에 있어서,상기 프로세서는,윈도우 폭 및 윈도우 레벨이 상기 학습 영상에 기초하여 학습되는 신경망의 오브젝트에 대해 미리 설정된 값에서 벗어나도록 상기 윈도우 폭 및 윈도우 레벨 중 적어도 하나의 파라미터에 대한 값을 변형시키는, 영상 생성 장치.
- 제11항에 있어서,상기 프로세서는,상기 학습 영상의 픽셀 값에 노이즈를 추가하는, 영상 생성 장치.
- 제18항에 있어서,상기 학습 영상의 픽셀 값에 추가된 노이즈는,상기 기준 영상을 촬영한 기기의 특성 및 상기 기준 영상에 포함된 오브젝트 중 적어도 하나에 기반하여 생성되는, 영상 생성 장치.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/787,728 US20170256038A1 (en) | 2015-09-24 | 2015-09-24 | Image Generating Method and Apparatus, and Image Analyzing Method |
KR1020177036475A KR101880035B1 (ko) | 2015-09-24 | 2015-09-24 | 영상 생성 방법 및 장치, 및 영상 분석 방법 |
PCT/KR2015/010085 WO2017051943A1 (ko) | 2015-09-24 | 2015-09-24 | 영상 생성 방법 및 장치, 및 영상 분석 방법 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/KR2015/010085 WO2017051943A1 (ko) | 2015-09-24 | 2015-09-24 | 영상 생성 방법 및 장치, 및 영상 분석 방법 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017051943A1 true WO2017051943A1 (ko) | 2017-03-30 |
Family
ID=58386849
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2015/010085 WO2017051943A1 (ko) | 2015-09-24 | 2015-09-24 | 영상 생성 방법 및 장치, 및 영상 분석 방법 |
Country Status (3)
Country | Link |
---|---|
US (1) | US20170256038A1 (ko) |
KR (1) | KR101880035B1 (ko) |
WO (1) | WO2017051943A1 (ko) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108537794A (zh) * | 2018-04-19 | 2018-09-14 | 上海联影医疗科技有限公司 | 医学图像数据处理方法、装置和计算机可读存储介质 |
CN110462689A (zh) * | 2017-04-05 | 2019-11-15 | 通用电气公司 | 基于深度学习的断层摄影重建 |
US20220261670A1 (en) * | 2017-12-18 | 2022-08-18 | Aptiv Technologies Limited | Methods of Processing and Generating Image Data in a Connectionist Network |
Families Citing this family (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11517197B2 (en) | 2017-10-06 | 2022-12-06 | Canon Medical Systems Corporation | Apparatus and method for medical image reconstruction using deep learning for computed tomography (CT) image noise and artifacts reduction |
US10803984B2 (en) | 2017-10-06 | 2020-10-13 | Canon Medical Systems Corporation | Medical image processing apparatus and medical image processing system |
CN109782989B (zh) | 2017-11-14 | 2021-08-03 | 腾讯科技(深圳)有限公司 | 安全验证方法及相关装置 |
US10891762B2 (en) * | 2017-11-20 | 2021-01-12 | ClariPI Inc. | Apparatus and method for medical image denoising based on deep learning |
KR102033743B1 (ko) * | 2017-11-20 | 2019-11-08 | 주식회사 클라리파이 | 딥러닝 기반 ct 이미지 잡음 저감 장치 및 방법 |
US10825132B2 (en) * | 2018-02-20 | 2020-11-03 | Element Ai Inc. | Training method for convolutional neural networks for use in artistic style transfers for video |
US11341277B2 (en) * | 2018-04-20 | 2022-05-24 | Nec Corporation | Method and system for securing machine learning models |
US12248877B2 (en) * | 2018-05-23 | 2025-03-11 | Movidius Ltd. | Hybrid neural network pruning |
KR102211541B1 (ko) | 2019-01-16 | 2021-02-02 | 연세대학교 산학협력단 | 학습 영상 유효성 검증 장치 및 방법 |
DE102020133963A1 (de) * | 2020-03-10 | 2021-09-16 | Samsung Electronics Co., Ltd. | Systeme und Verfahren zur Abbildrauschunterdrückung unter Verwendung von tiefen gefalteten Netzwerken |
KR102208690B1 (ko) * | 2020-07-23 | 2021-01-28 | 주식회사 어반베이스 | 데이터 증강 기반 스타일 분석 모델 학습 장치 및 방법 |
KR102208685B1 (ko) * | 2020-07-23 | 2021-01-28 | 주식회사 어반베이스 | 데이터 증강 기반 공간 분석 모델 학습 장치 및 방법 |
KR102208688B1 (ko) * | 2020-07-23 | 2021-01-28 | 주식회사 어반베이스 | 데이터 증강 기반 사물 분석 모델 학습 장치 및 방법 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7848555B2 (en) * | 2005-01-04 | 2010-12-07 | Carestream Health, Inc. | Computer aided detection of microcalcification clusters |
WO2013104938A2 (en) * | 2012-01-11 | 2013-07-18 | 77 Elektronika Muszeripari Kft. | Neural network and a method for teaching thereof |
KR20140145801A (ko) * | 2013-06-14 | 2014-12-24 | 전북대학교산학협력단 | 신경망을 이용한 영상의 화질 개선 시스템 및 방법 |
KR20150088157A (ko) * | 2014-01-23 | 2015-07-31 | 삼성전자주식회사 | 행동 인식을 위한 특징 벡터 생성 방법, 히스토그램 생성 방법, 및 분류기 학습 방법 |
US20150238148A1 (en) * | 2013-10-17 | 2015-08-27 | Siemens Aktiengesellschaft | Method and system for anatomical object detection using marginal space deep neural networks |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7130776B2 (en) * | 2002-03-25 | 2006-10-31 | Lockheed Martin Corporation | Method and computer program product for producing a pattern recognition training set |
-
2015
- 2015-09-24 KR KR1020177036475A patent/KR101880035B1/ko active Active
- 2015-09-24 US US14/787,728 patent/US20170256038A1/en not_active Abandoned
- 2015-09-24 WO PCT/KR2015/010085 patent/WO2017051943A1/ko active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7848555B2 (en) * | 2005-01-04 | 2010-12-07 | Carestream Health, Inc. | Computer aided detection of microcalcification clusters |
WO2013104938A2 (en) * | 2012-01-11 | 2013-07-18 | 77 Elektronika Muszeripari Kft. | Neural network and a method for teaching thereof |
KR20140145801A (ko) * | 2013-06-14 | 2014-12-24 | 전북대학교산학협력단 | 신경망을 이용한 영상의 화질 개선 시스템 및 방법 |
US20150238148A1 (en) * | 2013-10-17 | 2015-08-27 | Siemens Aktiengesellschaft | Method and system for anatomical object detection using marginal space deep neural networks |
KR20150088157A (ko) * | 2014-01-23 | 2015-07-31 | 삼성전자주식회사 | 행동 인식을 위한 특징 벡터 생성 방법, 히스토그램 생성 방법, 및 분류기 학습 방법 |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110462689A (zh) * | 2017-04-05 | 2019-11-15 | 通用电气公司 | 基于深度学习的断层摄影重建 |
KR20190133728A (ko) * | 2017-04-05 | 2019-12-03 | 제너럴 일렉트릭 캄파니 | 딥러닝에 기초한 단층촬영 재구성 |
KR102257637B1 (ko) * | 2017-04-05 | 2021-05-31 | 제너럴 일렉트릭 캄파니 | 딥러닝에 기초한 단층촬영 재구성 |
CN110462689B (zh) * | 2017-04-05 | 2023-06-02 | 通用电气公司 | 基于深度学习的断层摄影重建 |
US20220261670A1 (en) * | 2017-12-18 | 2022-08-18 | Aptiv Technologies Limited | Methods of Processing and Generating Image Data in a Connectionist Network |
US12266191B2 (en) * | 2017-12-18 | 2025-04-01 | Aptiv Technologies AG | Methods of processing and generating image data in a connectionist network |
CN108537794A (zh) * | 2018-04-19 | 2018-09-14 | 上海联影医疗科技有限公司 | 医学图像数据处理方法、装置和计算机可读存储介质 |
CN108537794B (zh) * | 2018-04-19 | 2021-09-21 | 上海联影医疗科技股份有限公司 | 医学图像数据处理方法、装置和计算机可读存储介质 |
Also Published As
Publication number | Publication date |
---|---|
US20170256038A1 (en) | 2017-09-07 |
KR20180004824A (ko) | 2018-01-12 |
KR101880035B1 (ko) | 2018-07-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2017051943A1 (ko) | 영상 생성 방법 및 장치, 및 영상 분석 방법 | |
Harrison et al. | Progressive and multi-path holistically nested neural networks for pathological lung segmentation from CT images | |
US11151721B2 (en) | System and method for automatic detection, localization, and semantic segmentation of anatomical objects | |
BR112020025391A2 (pt) | Sistemas e métodos para processar vídeo em tempo real de um dispositivo de imagem médica e detectar objetos no vídeo | |
WO2021071288A1 (ko) | 골절 진단모델의 학습 방법 및 장치 | |
CN106777953A (zh) | 医学影像数据的分析方法及系统 | |
Poap et al. | Chest radiographs segmentation by the use of nature-inspired algorithm for lung disease detection | |
WO2019172498A1 (ko) | 종양의 악성도와 악성도 추론의 근거를 제공하는 컴퓨터 보조 진단 시스템 및 그 방법 | |
WO2023229415A1 (ko) | Ar 영상 제공 방법 및 ar 영상 제공 장치 | |
WO2019143021A1 (ko) | 영상의 열람을 지원하는 방법 및 이를 이용한 장치 | |
WO2022145841A1 (ko) | 병변 판독 방법 및 이를 위한 장치 | |
WO2021034138A1 (ko) | 치매 평가 방법 및 이를 이용한 장치 | |
WO2017221412A1 (ja) | 画像処理装置、学習装置、画像処理方法、識別基準の作成方法、学習方法およびプログラム | |
Chang et al. | Medical image segmentation using a contextual-constraint-based Hopfield neural cube | |
WO2021054700A1 (ko) | 치아 병변 정보 제공 방법 및 이를 이용한 장치 | |
WO2021002669A1 (ko) | 병변 통합 학습 모델을 구축하는 장치와 방법, 및 상기 병변 통합 학습 모델을 사용하여 병변을 진단하는 장치와 방법 | |
WO2020139011A1 (ko) | 뇌병변 정보 제공 장치 및 방법 | |
WO2023027248A1 (ko) | 데이터 생성 방법 및 이를 이용한 학습방법 및 장치 | |
Alkhalisy et al. | Students Behavior Detection Based on Improved YOLOv5 Algorithm Combining with CBAM Attention Mechanism. | |
Olsen | Motion tracking of infants in risk of cerebral palsy | |
WO2019168280A1 (ko) | 뉴럴 네트워크를 이용하여 캡슐 내시경 영상으로부터 병변 판독 방법 및 장치 | |
Dwiyanto et al. | Dealing with COVID-19 using deep learning for computer vision | |
WO2024117555A1 (ko) | 이미지 데이터에 대한 분류 모델을 학습시키는 방법 및 이를 위한 장치 | |
WO2023033392A1 (ko) | 기계학습 방법 및 장치 | |
WO2024106799A1 (ko) | 늑골 골절에 대한 정보 제공 방법 및 이를 이용한 디바이스 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 14787728 Country of ref document: US |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15904754 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20177036475 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 06/08/2018) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15904754 Country of ref document: EP Kind code of ref document: A1 |