[go: up one dir, main page]

CN111882561A - Cancer cell identification and diagnosis system - Google Patents

Cancer cell identification and diagnosis system Download PDF

Info

Publication number
CN111882561A
CN111882561A CN202010561274.3A CN202010561274A CN111882561A CN 111882561 A CN111882561 A CN 111882561A CN 202010561274 A CN202010561274 A CN 202010561274A CN 111882561 A CN111882561 A CN 111882561A
Authority
CN
China
Prior art keywords
image
cell
nuclear
segmentation
value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010561274.3A
Other languages
Chinese (zh)
Inventor
车俐
韩梦玲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guilin University of Electronic Technology
Original Assignee
Guilin University of Electronic Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guilin University of Electronic Technology filed Critical Guilin University of Electronic Technology
Priority to CN202010561274.3A priority Critical patent/CN111882561A/en
Publication of CN111882561A publication Critical patent/CN111882561A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/187Segmentation; Edge detection involving region growing; involving region merging; involving connected component labelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/62Analysis of geometric attributes of area, perimeter, diameter or volume
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10056Microscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30024Cell structures in vitro; Tissue sections in vitro

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Geometry (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)
  • Investigating Or Analysing Biological Materials (AREA)

Abstract

The invention discloses a cancer cell identification and diagnosis system, which comprises: a cell image preprocessing module: an improved mathematical morphology on-off filtering algorithm is adopted, and a proper structural element is selected, so that background impurities and small normal discrete cells of a cell image are removed; cell image segmentation module: an improved region growing algorithm is adopted for extracting a cell nucleus region, a watershed algorithm based on price marker control is adopted for extracting the outline of an unlaminated cell body, and a segmentation algorithm based on a snake model is adopted for extracting the outline of the overlapped cell; cancer cell image feature extraction module: and calculating the nuclear-to-cytoplasmic ratio of each segmented region, whether the cell nucleus is uniform in size, abnormal nuclear staining and whether the nuclear distance is uniform. Cell image classification and identification module: cancer cells are identified by neural network identification techniques. The system can efficiently and accurately carry out quantitative analysis and detection identification on the cell image.

Description

一种癌细胞识别诊断系统A cancer cell identification and diagnosis system

技术领域technical field

本发明涉及细胞图像识别技术领域,具体为一种癌细胞识别诊断系统。The invention relates to the technical field of cell image recognition, in particular to a cancer cell recognition and diagnosis system.

背景技术Background technique

癌症是危害人类身体健康的常见疾病,癌症的早期诊断是治疗的关键。采用细胞学计算机辅助诊断技术,可以有效地降低医生的工作强度以及提高诊断的准确性。由于细胞图像背景复杂、细胞形态多样及重叠等问题,使得计算机检测识别难度较大。本发明对典型医学图像分割算法做出了新的技术改良,针对不重叠细胞图像及重叠细胞图像分别给出了改进算法。Cancer is a common disease that endangers human health, and early diagnosis of cancer is the key to treatment. The use of cytology computer-aided diagnosis technology can effectively reduce the work intensity of doctors and improve the accuracy of diagnosis. Due to the complex background of cell images, diverse cell shapes and overlapping problems, it is difficult to detect and identify by computer. The present invention makes new technical improvements to the typical medical image segmentation algorithm, and provides improved algorithms for non-overlapping cell images and overlapping cell images respectively.

发明内容SUMMARY OF THE INVENTION

本发明的目的是针对现有技术中存在的不足,而提供一种癌细胞识别诊断系统。这种诊断系统,可实现目标区域更精准地提取,对细胞图像进行更高效、准确地定量分析和检测识别。The purpose of the present invention is to provide a system for identifying and diagnosing cancer cells in view of the deficiencies in the prior art. This diagnostic system can achieve more accurate extraction of target areas, and perform more efficient and accurate quantitative analysis, detection and identification of cell images.

实现本发明目的的技术方案是:The technical scheme that realizes the object of the present invention is:

一种癌细胞识别诊断系统,包括,A cancer cell identification diagnostic system comprising,

细胞图像预处理模块:首先将细胞图片转化为数字图像并对其进行灰度化处理。为了更加准确的提取图像中有用细胞特征,需要将图像中小,离散正常细胞以及加性噪声一起去除。针对这一情况采用改进开闭滤波算法进行图像去噪,选取结构元素B为正常淋巴细胞的两倍。针对小于结构元素B的噪声,运用开运算消除背景中胡椒状噪声,然后利用闭运算消除砂眼噪声。针对大于结构元素B的噪声,首先将滤波的图像二值化,并对原图像进行叠加提取。图像二值化是将原图像转化为二值图像,“0”代表目标图像,“1”代表背景,用该二值图像对原图像的内容进行选择,二值图像中值为“0”的像素位置均保留原灰度值,值为“1”的像素位置灰度值为“255”。Cell image preprocessing module: First, convert the cell image into a digital image and perform grayscale processing on it. In order to more accurately extract useful cell features in the image, it is necessary to remove small, discrete normal cells and additive noise in the image together. In view of this situation, an improved open-close filter algorithm is used to denoise the image, and the structural element B is selected to be twice that of normal lymphocytes. For the noise smaller than the structuring element B, the open operation is used to eliminate the pepper noise in the background, and then the closed operation is used to eliminate the trachoma noise. For the noise larger than the structural element B, the filtered image is first binarized, and the original image is superimposed and extracted. Image binarization is to convert the original image into a binary image, "0" represents the target image, "1" represents the background, and the binary image is used to select the content of the original image. The value in the binary image is "0". The pixel position retains the original gray value, and the gray value of the pixel position with a value of "1" is "255".

细胞图像分割模块:Cell Image Segmentation Module:

第一步:提取细胞核区域采取自动种子点的区域生长算法,将阈值分割和区域生长算法联合。首先设定阈值T=115,将图像分割成大于阈值T的对象点和小于阈值T的背景点两部分。原图像f(x,y)转化为输出图像g(x,y),当f(x,y)>T,g(x,y)=1;当f(x,y)≤T,g(x,y)=0。此法实现了“粗轮廓”提取细胞核的大部分区域。设置种子的生长规则为相邻像素点灰度值与种子区域平均灰度值的差的绝对值小于阈值0.15.以粗轮廓的质心作为种子的生长点,每个种子按生长规则迭代生长,差值小于阈值的加入种子,直到找不到满足规则的点就结束生长。此法实现了“细轮廓”提取细胞核。最后采用Sobel算子做边缘检测,Sobel算子的模板及对应公式如下:The first step: extracting the nucleus region adopts the region growing algorithm of automatic seed points, and combines the threshold segmentation and the region growing algorithm. First, the threshold value T=115 is set, and the image is divided into two parts: the object points larger than the threshold value T and the background points smaller than the threshold value T. The original image f(x,y) is converted into the output image g(x,y), when f(x,y)>T, g(x,y)=1; when f(x,y)≤T, g( x, y)=0. This method achieves "coarse outline" extraction of most regions of the nucleus. Set the growth rule of seeds as the absolute value of the difference between the gray value of adjacent pixels and the average gray value of the seed area is less than the threshold value of 0.15. The centroid of the thick outline is used as the growth point of the seed, and each seed grows iteratively according to the growth rule. Seeds with a value less than the threshold are added until no point that satisfies the rule is found, and the growth ends. This method achieves "fine contour" extraction of cell nuclei. Finally, the Sobel operator is used for edge detection. The template and corresponding formula of the Sobel operator are as follows:

Figure BDA0002546163020000021
Figure BDA0002546163020000021

最后输出图像。The final output image.

第二步:提取单个细胞体区域采取标记符控制的分水岭算法。为了方便确定内外标记,结合细胞显微镜图像背景量,目标暗的特点,先对细胞灰度图像做反相处理,得到目标物体较亮、背景区域较暗的图像。将图像进行二值化处理,将二值图像进行距离变化,使得分界线能保住目标物体而又不会过于接近物体边缘。最后采用分水岭变换,得到的分界线作为外部标记。将反处理的图像做基于重建的开操作和闭操作,去除目标内部细节并对局部极大值区域作为内部标记。结合内外标记,将反处理后的图像计算梯度幅值,在改进后的梯度图像上进行分水岭变换,在细胞灰度图像上显示细胞体分割轮廓。The second step: extracting a single cell body region adopts a marker-controlled watershed algorithm. In order to facilitate the determination of internal and external markers, combined with the background volume of the cell microscope image and the characteristics of the dark target, the grayscale image of the cell was first inverted to obtain an image with a brighter target object and a darker background area. The image is binarized, and the distance of the binary image is changed, so that the dividing line can keep the target object without being too close to the edge of the object. Finally, a watershed transformation is used, and the obtained dividing line is used as an external marker. Reconstruction-based opening and closing operations are performed on the inversely processed image to remove the internal details of the target and use the local maximum area as an internal marker. Combined with the inner and outer markers, the gradient magnitude is calculated from the inversely processed image, and the watershed transformation is performed on the improved gradient image, and the cell body segmentation contour is displayed on the cell grayscale image.

第三步:提取叠层的细胞图像中的细胞轮廓采用了基于snake模型的分割算法。针对snake模型对初始位置敏感的特点,采用细胞的稀疏轮廓点模型,利用环形动态轮廓搜索算法,自动定位出主体细胞的轮廓点,之后利用snake模型方法进行分割,经过曲线对细胞体边缘的拟合,最终获得细胞体轮廓。其分割精度约为94%,过分割率约为2.5%,欠分割率约3.5%。Step 3: Extract the cell contours in the stacked cell images using a segmentation algorithm based on the snake model. In view of the characteristic that the snake model is sensitive to the initial position, the sparse contour point model of the cell is used, and the circular dynamic contour search algorithm is used to automatically locate the contour points of the main cell. Finally, the outline of the cell body is obtained. Its segmentation accuracy is about 94%, the over-segmentation rate is about 2.5%, and the under-segmentation rate is about 3.5%.

癌细胞图像特征提取模块:计算分割后每个区域的核质比,细胞核大小是否均匀,核染色异常以及核间距是否均匀。Cancer cell image feature extraction module: Calculate the nuclear-to-cytoplasmic ratio of each region after segmentation, whether the size of the nucleus is uniform, whether the nuclear staining is abnormal, and whether the nuclear spacing is uniform.

核质比:对于不叠层的细胞图像,通过segment子程序将图像分为一个分块一个单细胞,调用ncratio子程序计算出每个细胞的核质比。当该细胞的核质比远超于正常细胞核质比0.5:1时,判定疑似癌细胞。对于叠层的细胞图像,调用segment子程序,将图像分区,计算整个区域的核质比,当其比值远超于正常细胞核质比0.5:1时,可判定疑似癌细胞。Nucleocytoplasmic ratio: For non-overlapping cell images, the segment subroutine divides the image into a block and a single cell, and calls the ncratio subroutine to calculate the nucleocytoplasmic ratio of each cell. When the nucleocytoplasmic ratio of the cell is much higher than the normal cell nucleocytoplasmic ratio of 0.5:1, it is determined to be a suspected cancer cell. For the stacked cell images, the segment subroutine is called, the image is divided into sections, and the nuclear-cytoplasmic ratio of the entire area is calculated. When the ratio is much higher than the normal cell nuclear-cytoplasmic ratio of 0.5:1, a suspected cancer cell can be determined.

细胞核大小:将分割后得到的图像通过segment分块处理,将得到的单个细胞核大小。若细胞核大小分布比较分散,则符合癌细胞特征。Nucleus size: The image obtained after segmentation is processed into segments to obtain the size of a single nucleus. If the nuclear size distribution is relatively scattered, it is consistent with the characteristics of cancer cells.

核染色体异常:通过计算核深染面积与细胞核面积的比值,若占比大于31.91%,可以判定疑似癌细胞。Nuclear chromosome abnormality: By calculating the ratio of nuclear hyperchromatic area to nuclear area, if the proportion is greater than 31.91%, it can be determined as a suspected cancer cell.

成团细胞相邻核间距:通过调用distant计算成团细胞相邻核间距,若成团细胞细胞核分布不均匀,可以判定疑似癌细胞。The distance between adjacent nuclei of agglomerated cells: Calculate the distance between adjacent nuclei of agglomerated cells by calling distance. If the nuclei of agglomerated cells are not uniformly distributed, a suspected cancer cell can be determined.

细胞图像分类识别模块:采用BP神经网络识别技术识别癌细胞。首先初始化网络的权值和阈值。网络权值的初始值即为网络从误差曲面的哪一点开始训练,该点影响网络的训练时间。开始向前计算,求出所有神经元的所有输出。计算输出层误差梯度。然后向后计算各隐层误差梯度,计算并保存各权值修正量,修正权值。判断是否达到训练目标。如达到,训练结束,达不到则转入向前计算。Cell image classification and recognition module: using BP neural network recognition technology to identify cancer cells. First initialize the weights and thresholds of the network. The initial value of the network weight is the point at which the network starts to train on the error surface, which affects the training time of the network. Start computing forward, finding all outputs of all neurons. Calculate the output layer error gradient. Then calculate the error gradient of each hidden layer backward, calculate and save the correction amount of each weight, and correct the weight. Determine whether the training goal is achieved. If it is reached, the training is over. If it is not reached, it will be transferred to the forward calculation.

附图说明Description of drawings

图1为实施例的结构示意图。FIG. 1 is a schematic structural diagram of an embodiment.

图2为实施例中提取细胞核区域的流程示意图。FIG. 2 is a schematic flow chart of extracting the nucleus region in the embodiment.

图3为实施例中BP神经网络识别算法的流程示意图FIG. 3 is a schematic flowchart of the BP neural network identification algorithm in the embodiment

具体实施方式Detailed ways

下面结合附图及具体实施例对本发明作进一步的详细描述,但不是对本发明的限定The present invention will be described in further detail below in conjunction with the accompanying drawings and specific embodiments, but it is not intended to limit the present invention.

实施例:Example:

参照图1,一种癌细胞识别诊断系统,包括,Referring to Figure 1, a cancer cell identification and diagnosis system includes,

细胞图像预处理模块:Cell image preprocessing module:

首先将细胞图片转化为数字图像并对其进行灰度化处理。为了更加准确的提取图像中有用细胞特征,需要将图像中小,离散正常细胞以及加性噪声一起去除。针对这一情况采用改进开闭滤波算法进行图像去噪,选取结构元素B为正常淋巴细胞的两倍。针对小于结构元素B的噪声,运用开运算消除背景中胡椒状噪声,然后利用闭运算消除砂眼噪声。针对大于结构元素B的噪声,首先将滤波的图像二值化,并对原图像进行叠加提取。图像二值化是将原图像转化为二值图像,“0”代表目标图像,“1”代表背景,用该二值图像对原图像的内容进行选择,二值图像中值为“0”的像素位置均保留原灰度值,值为“1”的像素位置灰度值为“255”。如图2所示,细胞图像分割模块:The cell pictures are first converted into digital images and grayscaled. In order to more accurately extract useful cell features in the image, it is necessary to remove small, discrete normal cells and additive noise in the image together. In view of this situation, an improved open-close filter algorithm is used to denoise the image, and the structural element B is selected to be twice that of normal lymphocytes. For the noise smaller than the structuring element B, the open operation is used to eliminate the pepper noise in the background, and then the closed operation is used to eliminate the trachoma noise. For the noise larger than the structural element B, the filtered image is first binarized, and the original image is superimposed and extracted. Image binarization is to convert the original image into a binary image, "0" represents the target image, "1" represents the background, and the binary image is used to select the content of the original image. The value in the binary image is "0". The pixel position retains the original gray value, and the gray value of the pixel position with a value of "1" is "255". As shown in Figure 2, the cell image segmentation module:

第一步:提取细胞核区域采取自动种子点的区域生长算法,将阈值分割和区域生长算法联合。首先设定阈值T=115,将图像分割成大于阈值T的对象点和小于阈值T的背景点两部分。原图像f(x,y)转化为输出图像g(x,y),当f(x,y)>T,g(x,y)=1;当f(x,y)≤T,g(x,y)=0。此法实现了“粗轮廓”提取细胞核的大部分区域。设置种子的生长规则为相邻像素点灰度值与种子区域平均灰度值的差的绝对值小于阈值0.15.以粗轮廓的质心作为种子的生长点,每个种子按生长规则迭代生长,差值小于阈值的加入种子,直到找不到满足规则的点就结束生长。此法实现了“细轮廓”提取细胞核。最后采用Sobel算子做边缘检测,Sobel算子的模板及对应公式如下:The first step: extracting the nucleus region adopts the region growing algorithm of automatic seed points, and combines the threshold segmentation and the region growing algorithm. First, the threshold value T=115 is set, and the image is divided into two parts: the object points larger than the threshold value T and the background points smaller than the threshold value T. The original image f(x,y) is converted into the output image g(x,y), when f(x,y)>T, g(x,y)=1; when f(x,y)≤T, g( x, y)=0. This method achieves "coarse outline" extraction of most regions of the nucleus. Set the growth rule of seeds as the absolute value of the difference between the gray value of adjacent pixels and the average gray value of the seed area is less than the threshold value of 0.15. The centroid of the thick outline is used as the growth point of the seed, and each seed grows iteratively according to the growth rule. Seeds with a value less than the threshold are added until no point that satisfies the rule is found, and the growth ends. This method achieves "fine contour" extraction of cell nuclei. Finally, the Sobel operator is used for edge detection. The template and corresponding formula of the Sobel operator are as follows:

Figure BDA0002546163020000041
Figure BDA0002546163020000041

最后输出图像。The final output image.

第二步:提取单个细胞体区域采取标记符控制的分水岭算法。为了方便确定内外标记,结合细胞显微镜图像背景量,目标暗的特点,先对细胞灰度图像做反相处理,得到目标物体较亮、背景区域较暗的图像。将图像进行二值化处理,将二值图像进行距离变化,使得分界线能保住目标物体而又不会过于接近物体边缘。最后采用分水岭变换,得到的分界线作为外部标记。将反处理的图像做基于重建的开操作和闭操作,去除目标内部细节并对局部极大值区域作为内部标记。结合内外标记,将反处理后的图像计算梯度幅值,在改进后的梯度图像上进行分水岭变换,在细胞灰度图像上显示细胞体分割轮廓。The second step: extracting a single cell body region adopts a marker-controlled watershed algorithm. In order to facilitate the determination of internal and external markers, combined with the background volume of the cell microscope image and the characteristics of the dark target, the grayscale image of the cell was first inverted to obtain an image with a brighter target object and a darker background area. The image is binarized, and the distance of the binary image is changed, so that the dividing line can keep the target object without being too close to the edge of the object. Finally, a watershed transformation is used, and the obtained dividing line is used as an external marker. Reconstruction-based opening and closing operations are performed on the inversely processed image to remove the internal details of the target and use the local maximum area as an internal marker. Combined with the inner and outer markers, the gradient magnitude is calculated from the inversely processed image, and the watershed transformation is performed on the improved gradient image, and the cell body segmentation contour is displayed on the cell grayscale image.

第三步:提取叠层的细胞图像中的细胞轮廓采用了基于snake模型的分割算法。针对snake模型对初始位置敏感的特点,采用细胞的稀疏轮廓点模型,利用环形动态轮廓搜索算法,自动定位出主体细胞的轮廓点,之后利用snake模型方法进行分割,经过曲线对细胞体边缘的拟合,最终获得细胞体轮廓。其分割精度约为94%,过分割率约为2.5%,欠分割率约3.5%。Step 3: Extract the cell contours in the stacked cell images using a segmentation algorithm based on the snake model. In view of the characteristic that the snake model is sensitive to the initial position, the sparse contour point model of the cell is used, and the circular dynamic contour search algorithm is used to automatically locate the contour points of the main cell. Finally, the outline of the cell body is obtained. Its segmentation accuracy is about 94%, the over-segmentation rate is about 2.5%, and the under-segmentation rate is about 3.5%.

癌细胞图像特征提取模块:计算分割后每个区域的核质比,细胞核大小是否均匀,核染色异常以及核间距是否均匀。Cancer cell image feature extraction module: Calculate the nuclear-to-cytoplasmic ratio of each region after segmentation, whether the size of the nucleus is uniform, whether the nuclear staining is abnormal, and whether the nuclear spacing is uniform.

核质比:对于不叠层的细胞图像,通过segment子程序将图像分为一个分块一个单细胞,调用ncratio子程序计算出每个细胞的核质比。当该细胞的核质比远超于正常细胞核质比0.5:1时,判定疑似癌细胞。对于叠层的细胞图像,调用segment子程序,将图像分区,计算整个区域的核质比,当其比值远超于正常细胞核质比0.5:1时,可判定疑似癌细胞。Nucleocytoplasmic ratio: For non-overlapping cell images, the segment subroutine divides the image into a block and a single cell, and calls the ncratio subroutine to calculate the nucleocytoplasmic ratio of each cell. When the nucleocytoplasmic ratio of the cell is much higher than the normal cell nucleocytoplasmic ratio of 0.5:1, it is determined to be a suspected cancer cell. For the stacked cell images, the segment subroutine is called, the image is divided into sections, and the nuclear-cytoplasmic ratio of the entire area is calculated. When the ratio is much higher than the normal cell nuclear-cytoplasmic ratio of 0.5:1, a suspected cancer cell can be determined.

细胞核大小:将分割后得到的图像通过segment分块处理,将得到的单个细胞核大小。若细胞核大小分布比较分散,则符合癌细胞特征。Nucleus size: The image obtained after segmentation is processed into segments to obtain the size of a single nucleus. If the nuclear size distribution is relatively scattered, it is consistent with the characteristics of cancer cells.

核染色体异常:通过计算核深染面积与细胞核面积的比值,若占比大于31.91%,可以判定疑似癌细胞。Nuclear chromosome abnormality: By calculating the ratio of nuclear hyperchromatic area to nuclear area, if the proportion is greater than 31.91%, it can be determined as a suspected cancer cell.

成团细胞相邻核间距:通过调用distant计算成团细胞相邻核间距,若成团细胞细胞核分布不均匀,可以判定疑似癌细胞。The distance between adjacent nuclei of agglomerated cells: Calculate the distance between adjacent nuclei of agglomerated cells by calling distance. If the nuclei of agglomerated cells are not uniformly distributed, a suspected cancer cell can be determined.

如图3所示,细胞图像分类识别模块:采用BP神经网络识别技术识别癌细胞。首先初始化网络的权值和阈值。网络权值的初始值即为网络从误差曲面的哪一点开始训练,该点影响网络的训练时间。开始向前计算,求出所有神经元的所有输出。计算输出层误差梯度。然后向后计算各隐层误差梯度,计算并保存各权值修正量,修正权值。判断是否达到训练目标。如达到,训练结束,达不到则转入向前计算。As shown in Figure 3, cell image classification and recognition module: using BP neural network recognition technology to identify cancer cells. First initialize the weights and thresholds of the network. The initial value of the network weight is the point at which the network starts to train on the error surface, which affects the training time of the network. Start computing forward, finding all outputs of all neurons. Calculate the output layer error gradient. Then calculate the error gradient of each hidden layer backward, calculate and save the correction amount of each weight, and correct the weight. Determine whether the training goal is achieved. If it is reached, the training is over. If it is not reached, it will be transferred to the forward calculation.

Claims (1)

1. A cancer cell identification and diagnosis system, comprising,
a cell image preprocessing module: firstly, converting a cell picture into a digital image and carrying out gray processing on the digital image. In order to extract useful cell features in an image more accurately, small and discrete normal cells in the image need to be removed together with additive noise. Aiming at the situation, an improved on-off filtering algorithm is adopted to carry out image denoising, and the structural element B is selected to be twice of that of a normal lymphocyte. And for the noise smaller than the structural element B, the opening operation is used for eliminating the pepper-shaped noise in the background, and then the closing operation is used for eliminating the sand hole noise. And for the noise larger than the structural element B, firstly, binarizing the filtered image, and superposing and extracting the original image. The image binarization is to convert an original image into a binary image, wherein '0' represents a target image, and '1' represents a background, the content of the original image is selected by using the binary image, the original gray value is reserved at the pixel position with the value of '0' in the binary image, and the gray value at the pixel position with the value of '1' is '255'.
Cell image segmentation module:
the first step is as follows: and extracting a cell nucleus region, and combining threshold segmentation and a region growing algorithm by adopting a region growing algorithm of an automatic seed point. First, a threshold value T is set to 115, and the image is divided into two parts, i.e., an object point larger than the threshold value T and a background point smaller than the threshold value T. Converting an original image f (x, y) into an output image g (x, y), wherein when f (x, y) > T, g (x, y) is 1; when f (x, y) is less than or equal to T, g (x, y) is 0. The method realizes the extraction of most areas of cell nucleuses by 'coarse contour'. Setting the growth rule of the seeds to be that the absolute value of the difference between the gray value of the adjacent pixel point and the average gray value of the seed area is smaller than the threshold value 0.15, taking the mass center of the coarse contour as the growth point of the seeds, iteratively growing each seed according to the growth rule, adding the seeds of which the difference value is smaller than the threshold value, and ending the growth until the point meeting the rule cannot be found. The method realizes the extraction of cell nucleus with 'fine contour'. And finally, using a Sobel operator for edge detection, wherein the template and the corresponding formula of the Sobel operator are as follows:
Figure FDA0002546163010000011
and finally, outputting the image.
The second step is that: and extracting a single cell body region and adopting a marker-controlled watershed algorithm. In order to conveniently determine the internal and external marks, the characteristics of background quantity and dark target of the cell microscope image are combined, the cell gray level image is subjected to phase inversion processing, and an image with a bright target object and a dark background area is obtained. And (4) carrying out binarization processing on the image, and carrying out distance change on the binary image so that the boundary line can keep the target object and cannot be too close to the edge of the object. And finally, adopting watershed transformation to obtain a boundary as an external mark. And performing opening operation and closing operation based on reconstruction on the reversely processed image, removing internal details of the target and taking the local maximum value area as an internal mark. And combining the internal and external marks, calculating a gradient amplitude value of the image after the reverse processing, performing watershed transformation on the improved gradient image, and displaying a cell body segmentation contour on the cell gray level image.
The third step: cell contours in the cell images of the stack are extracted using a snake model based segmentation algorithm. Aiming at the characteristic that a snake model is sensitive to an initial position, a sparse contour point model of cells is adopted, a circular dynamic contour searching algorithm is utilized, contour points of a main body cell are automatically positioned, then a snake model method is utilized for segmentation, and a cell body contour is finally obtained through fitting of a curve to the edge of a cell body. The segmentation precision is about 94%, the over-segmentation rate is about 2.5%, and the under-segmentation rate is about 3.5%.
Cancer cell image feature extraction module: and calculating the nuclear-to-cytoplasmic ratio of each segmented region, whether the cell nucleus is uniform in size, abnormal nuclear staining and whether the nuclear distance is uniform.
Nuclear-to-cytoplasmic ratio: for the image of the cells without overlapping layers, the image is divided into blocks and single cells by a segment subroutine, and the ncratio subroutine is called to calculate the nuclear-to-cytoplasmic ratio of each cell. When the nuclear-to-cytoplasmic ratio of the cell is far higher than that of the normal cell by 0.5:1, the cell is judged to be a suspected cancer cell. For the laminated cell image, a segment subroutine is called to divide the image into regions, the nuclear-to-cytoplasmic ratio of the whole region is calculated, and when the ratio is far higher than the normal nuclear-to-cytoplasmic ratio by 0.5:1, the suspected cancer cell can be judged.
Size of cell nucleus: and (4) processing the segmented image by segment partitioning to obtain the size of a single cell nucleus. If the size distribution of the cell nucleus is dispersed, the characteristics of the cancer cells are met.
Nuclear chromosomal abnormalities: and (3) judging the suspected cancer cells by calculating the ratio of the deep staining area of the nuclei to the area of the nuclei if the ratio is more than 31.91 percent.
Adjacent nuclear spacing of clumped cells: and (4) calculating the distance between adjacent nuclei of the clustered cells by calling distant, and if the nuclei of the clustered cells are not uniformly distributed, judging the suspected cancer cells.
Cell image classification and identification module: and (3) identifying the cancer cells by adopting a BP neural network identification technology. The weights and thresholds of the network are first initialized. The initial value of the network weight is the point from which the network starts training on the error surface, and the point influences the training time of the network. The forward calculation is started and all outputs of all neurons are found. An output layer error gradient is calculated. And then calculating the error gradient of each hidden layer backwards, calculating and storing the correction quantity of each weight and correcting the weight. And judging whether the training target is reached. If the result is reached, the training is finished, and if the result is not reached, the calculation is carried out forward.
CN202010561274.3A 2020-06-18 2020-06-18 Cancer cell identification and diagnosis system Pending CN111882561A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010561274.3A CN111882561A (en) 2020-06-18 2020-06-18 Cancer cell identification and diagnosis system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010561274.3A CN111882561A (en) 2020-06-18 2020-06-18 Cancer cell identification and diagnosis system

Publications (1)

Publication Number Publication Date
CN111882561A true CN111882561A (en) 2020-11-03

Family

ID=73157683

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010561274.3A Pending CN111882561A (en) 2020-06-18 2020-06-18 Cancer cell identification and diagnosis system

Country Status (1)

Country Link
CN (1) CN111882561A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112801939A (en) * 2020-12-31 2021-05-14 杭州迪英加科技有限公司 Method for improving index accuracy of pathological image KI67
CN113222969A (en) * 2021-05-28 2021-08-06 大连海事大学 Imaging flow cytometry adhesion counting and activity detection method and device based on bidirectional background difference method
CN114004851A (en) * 2021-11-26 2022-02-01 广州市艾贝泰生物科技有限公司 Cell image segmentation method and device and cell counting method
CN115100473A (en) * 2022-06-29 2022-09-23 武汉兰丁智能医学股份有限公司 Lung cell image classification method based on parallel neural network
CN115100647A (en) * 2022-07-29 2022-09-23 烟台至公生物医药科技有限公司 Method and device for identifying cervical cancer cells
CN116934742A (en) * 2023-09-13 2023-10-24 中山大学肿瘤防治中心(中山大学附属肿瘤医院、中山大学肿瘤研究所) A lymphatic structure image recognition method and system
CN117288659A (en) * 2023-11-24 2023-12-26 深圳市前海高新国际医疗管理有限公司 Morphological analysis system of tumor stem cells and normal cells based on AI
CN117392154A (en) * 2023-12-07 2024-01-12 华侨大学 A phase contrast microscopy cell image segmentation method and system
CN117808813A (en) * 2024-03-01 2024-04-02 乐恩(北京)医药技术有限公司 Respiratory system lung nodule tumor cell image detection method and system and storage medium
CN119147104A (en) * 2024-11-20 2024-12-17 杭州抱抱堂传媒有限公司 Fault self-diagnosis early warning method, equipment and system for intelligent electric box

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112801939B (en) * 2020-12-31 2022-07-22 杭州迪英加科技有限公司 Method for improving index accuracy of pathological image KI67
CN112801939A (en) * 2020-12-31 2021-05-14 杭州迪英加科技有限公司 Method for improving index accuracy of pathological image KI67
CN113222969A (en) * 2021-05-28 2021-08-06 大连海事大学 Imaging flow cytometry adhesion counting and activity detection method and device based on bidirectional background difference method
CN113222969B (en) * 2021-05-28 2024-05-14 大连海事大学 Imaging flow type cell adhesion counting and activity detecting method and device based on bidirectional background difference method
CN114004851A (en) * 2021-11-26 2022-02-01 广州市艾贝泰生物科技有限公司 Cell image segmentation method and device and cell counting method
CN114004851B (en) * 2021-11-26 2022-11-29 广州市艾贝泰生物科技有限公司 Cell image segmentation method and device and cell counting method
CN115100473A (en) * 2022-06-29 2022-09-23 武汉兰丁智能医学股份有限公司 Lung cell image classification method based on parallel neural network
CN115100647A (en) * 2022-07-29 2022-09-23 烟台至公生物医药科技有限公司 Method and device for identifying cervical cancer cells
CN116934742B (en) * 2023-09-13 2024-02-27 中山大学肿瘤防治中心(中山大学附属肿瘤医院、中山大学肿瘤研究所) Lymphatic structure image recognition method and system
CN116934742A (en) * 2023-09-13 2023-10-24 中山大学肿瘤防治中心(中山大学附属肿瘤医院、中山大学肿瘤研究所) A lymphatic structure image recognition method and system
CN117288659A (en) * 2023-11-24 2023-12-26 深圳市前海高新国际医疗管理有限公司 Morphological analysis system of tumor stem cells and normal cells based on AI
CN117288659B (en) * 2023-11-24 2024-02-13 深圳市前海高新国际医疗管理有限公司 Morphological analysis system of tumor stem cells and normal cells based on AI
CN117392154B (en) * 2023-12-07 2024-04-12 华侨大学 Phase contrast microscopic cell image segmentation method and system
CN117392154A (en) * 2023-12-07 2024-01-12 华侨大学 A phase contrast microscopy cell image segmentation method and system
CN117808813A (en) * 2024-03-01 2024-04-02 乐恩(北京)医药技术有限公司 Respiratory system lung nodule tumor cell image detection method and system and storage medium
CN119147104A (en) * 2024-11-20 2024-12-17 杭州抱抱堂传媒有限公司 Fault self-diagnosis early warning method, equipment and system for intelligent electric box

Similar Documents

Publication Publication Date Title
CN111882561A (en) Cancer cell identification and diagnosis system
CN107194937B (en) Traditional Chinese medicine tongue picture image segmentation method in open environment
CN106651846B (en) Segmentation method of retinal blood vessel images
WO2016091016A1 (en) Nucleus marker watershed transformation-based method for splitting adhered white blood cells
CN107845098A (en) Liver cancer image full-automatic partition method based on random forest and fuzzy clustering
CN102298700A (en) Method for recognizing and positioning cells in bone marrow pathology image
CN110120056A (en) Blood leucocyte dividing method based on self-adapting histogram threshold value and contour detecting
CN103778627A (en) Sea oil spill detection method based on SAR image
Lai et al. Effective segmentation for dental X-ray images using texture-based fuzzy inference system
CN116188786B (en) Image segmentation system for hepatic duct and biliary tract calculus
Purnama et al. Follicle detection on the usg images to support determination of polycystic ovary syndrome
Mittal et al. Identification & enhancement of different skin lesion images by segmentation techniques
Bhagya et al. Analysis of image segmentation algorithms for the effective detection of leukemic cells
Soni et al. CT scan based brain tumor recognition and extraction using Prewitt and morphological dilation
Visalaxi et al. Lesion extraction of endometriotic images using open computer vision
Anandgaonkar et al. Brain tumor detection and identification from T1 post contrast MR images using cluster based segmentation
CN113850792B (en) A cell classification and counting method and system based on computer vision
CN110930346B (en) Automatic detection method and storage device for eyeground image microangioma
Barhoumi et al. Pigment network detection in dermatoscopic images for melanoma diagnosis
Ram et al. Symmetry-based detection of nuclei in microscopy images
Li et al. Automatic detection of leukocytes for cytometry with color decomposition
Tiwari et al. Brain tumor segmentation using CNN
CN111062909A (en) Method and equipment for judging benign and malignant breast tumor
Essaf et al. An improved lung parenchyma segmentation using the maximum inter-class variance method (OTSU)
Khan et al. Segmentation of single and overlapping leaves by extracting appropriate contours

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20201103