[go: up one dir, main page]

CN113781488A - Method, device and medium for segmentation of tongue image - Google Patents

Method, device and medium for segmentation of tongue image Download PDF

Info

Publication number
CN113781488A
CN113781488A CN202110881511.9A CN202110881511A CN113781488A CN 113781488 A CN113781488 A CN 113781488A CN 202110881511 A CN202110881511 A CN 202110881511A CN 113781488 A CN113781488 A CN 113781488A
Authority
CN
China
Prior art keywords
tongue
picture
tongue image
network
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110881511.9A
Other languages
Chinese (zh)
Other versions
CN113781488B (en
Inventor
谭峻东
杨光华
李少杰
路煜
王珂
李月溶
王耀南
成于伽
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hengqin Jingzhun Intelligent Medical Technology Co Ltd
Original Assignee
Hengqin Jingzhun Intelligent Medical Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hengqin Jingzhun Intelligent Medical Technology Co Ltd filed Critical Hengqin Jingzhun Intelligent Medical Technology Co Ltd
Priority to CN202110881511.9A priority Critical patent/CN113781488B/en
Publication of CN113781488A publication Critical patent/CN113781488A/en
Application granted granted Critical
Publication of CN113781488B publication Critical patent/CN113781488B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a technical scheme of a segmentation method, a segmentation device and a segmentation medium of a tongue picture image, which comprises the following steps: acquiring a tongue picture, and performing standard color processing on the tongue picture; acquiring tongue information of the tongue picture, performing frame selection on the tongue information, performing color space conversion processing on the tongue picture, and further performing tongue picture detection to obtain a rough tongue picture position of the tongue picture; determining a corresponding threshold range according to a Y-Cb-Cr color space conversion algorithm, intersecting the threshold range with a frame selection range, and performing tongue picture cutting processing; performing multi-detail reduction to generate a tongue picture with higher definition; training and testing are carried out, and a complete segmentation tongue picture is output. The invention has the beneficial effects that: improved confidence of tongue picture results; the calculation pressure of the deep convolutional neural network is reduced, the calculation speed is increased, the accuracy is also improved, the hardware cost is reduced, and the applicability of the model is enhanced.

Description

Tongue picture image segmentation method, apparatus and medium
Technical Field
The invention relates to the field of computers and medical treatment, in particular to a tongue picture image segmentation method, electronic equipment and a medium.
Background
For research in the medical field, it has become an important research direction to focus on the comprehensive physical condition of people rather than only on a single disease. The method is consistent with the idea that the human body is regarded as a whole in traditional Chinese medicine and the human body is in a healthy state by adjusting the circulation of the whole system. For a single individual, along with the continuous improvement of living standard, people pay more and more attention to the health, and pay more attention to the noninvasive and painless detection of diseases. The theory of traditional Chinese medicine holds that the body is an organic whole, and the tongue is connected with the viscera of the five viscera and the lung through the meridians and collaterals. The traditional Chinese medicine considers that the tongue inspection can understand the deficiency and excess of the lung, the disease nature cold and heat, the location of disease evil and the abundance and insufficiency of qi and blood of a patient, and plays an important role in disease condition evaluation and prescription development medication. Tongue diagnosis the pathological condition of a patient is diagnosed and analyzed by observing tongue picture. Therefore, the most obvious advantage of tongue diagnosis in traditional Chinese medicine is painlessness and no trauma. These conjunction points provide opportunities for the future development of tongue diagnosis in traditional Chinese medicine.
In recent years, image processing techniques have been rapidly developed, and more researchers are dedicated to quantification and standardization of tongue diagnosis. The tongue diagnosis is objectively performed in a plurality of aspects of research from tongue picture collection, tongue body segmentation, tongue coating separation, tongue picture characteristic extraction and attribute identification, and related products such as a tongue picture diagnosis system, a tongue diagnosis instrument and the like are published, and even are practically applied in clinic.
Although related devices such as tongue picture diagnosis systems and tongue diagnosis instruments are available at present, the tongue picture collection is generally required to be completed in a closed space with stable illumination, and specific collection devices are required, although a high-quality tongue picture can be obtained by means of constant illumination and professional devices, and the difficulty in subsequent tongue picture processing is reduced, on one hand, most of the collection or diagnosis devices are expensive and not easy to carry, and only some hospitals and some families have related equipment, so that the collection or diagnosis devices have certain limitations, and thus the collection or diagnosis devices are not widely applied. On the other hand, in consideration of popularization of smart phones, mobile phone camera shooting technology is greatly improved. Therefore, the tongue picture collection and intelligent analysis by using mobile devices in natural environment are gradually becoming new development directions. The segmentation and identification of tongue images taken by different devices under natural light conditions also become one of the main research works of computer tongue diagnosis, which is also the main content of research. In order to more accurately identify and analyze the tongue picture, two preprocessing operations of color correction and tongue body segmentation are carried out on the tongue picture.
At present, the tongue picture processing needs to train according to the tongue picture colors under different environmental illumination to obtain an illumination condition classifier, then train a color correction matrix under different illumination, finally pass the corrected picture through a tongue body segmentation network, and finally output the segmented tongue picture.
On one hand, hardware cost is increased by training a plurality of classifiers, and a large amount of time is consumed for accurate information marking, so that a better classifier can be obtained; on the other hand, the algorithm model is not an end-to-end model, and a plurality of models and a plurality of decomposition steps are adopted, so that the computational complexity and the time complexity of the model are increased. The standard U-Net network is a lightweight end-to-end network with a coding and decoding structure, because the coding model and the decoding model use the same symmetrical structure, and image processing is performed in a down-sampling mode in the coding stage, and image processing is performed in an up-sampling mode in the decoding stage, so that the whole network shape presents a U-shaped structure, which is called as the U-Net network. The unique mirror image operation of the U-Net network and the simple network model have quick and effective segmentation on a small amount of data, but the difficulty of network design is increased due to the fact that the input size and the output size are not equal, meanwhile, the feature graph merging and cutting operation is complicated, and the picture edge cutting and feature graph mismatching exist in an up-sampling mode, so that improvement needs to be conducted on the U-Net network to achieve a better segmentation effect.
Disclosure of Invention
The present invention is directed to solve at least one of the problems of the prior art, and provides a method, an apparatus and a medium for segmenting a tongue image, which overcome the disadvantages of the prior art.
The technical scheme of the invention comprises a segmentation method of a tongue picture image, which is characterized by comprising the following steps: acquiring a tongue picture, and performing standard color processing on the tongue picture to obtain a first picture; creating a VGG16 classification regression network, classifying a first picture, acquiring tongue information of the first picture, performing frame selection on the tongue information and performing color space conversion processing on the first picture, and further performing tongue picture detection to obtain a second picture comprising a rough tongue picture position; determining a corresponding threshold range of the second picture according to a Y-Cb-Cr color space conversion algorithm, intersecting the threshold range with a frame selection range, and performing tongue picture cutting processing to obtain a third picture; creating an SRGAN super-resolution generation network, and performing multi-detail reduction on the third picture to generate a fourth image with higher definition; and (3) creating a neural connection network based on an encoder-decoder structure, training and testing the fourth picture, and outputting a complete segmentation tongue picture.
The segmentation method of the tongue picture comprises the following steps of: creating a BP neural connection network, training the tongue picture by taking a standard color card value as a known input, wherein the known input color card value is taken as a supervision value, and executing supervision training to obtain a trained parameter value; initializing the BP neural connection network through the parameter values, inputting the tongue picture for calculation, and outputting a corrected first picture, wherein the first picture is the corrected tongue picture.
The tongue picture segmentation method comprises the following steps of training a tongue picture by taking a standard color card value as a known input: taking 24 colors of the standard color card values as sample values of the BP neural connection network, performing a plurality of hidden layer calculations on RGB channel values of each sample value, and outputting the corrected RGB channel values; comparing the corrected RGB channel value with the standard color card value, calculating an error, performing back propagation, and correcting all parameter values on a path according to an LM algorithm; and circulating the previous step until the errors of all the parameter values are reduced to be within the set range.
The segmentation method of the tongue picture image, wherein the BP neural connection network comprises: the standard color processing network learns the nonlinear relation between neural networks through the difference between the 24-color real color chip values and the color chip values in different scene pictures, and trains a weight model of the network by updating gradient values through back propagation; training the network through the VGG16 classification regression network through the existing tongue picture labels, and training the network through the existing data through a picture augmentation strategy to obtain results of 5 dimensions, namely whether the results are tongue pictures and position information; adopting Y-Cb-Cr color space conversion algorithm processing, including roughly determining the position information of the tongue picture through a color clustering strategy; generating a network and a tongue segmentation network through the SRGAN super resolution; using a ReLU (rectified Linear Unit) function as an activation function of the BP neural connection network, wherein the ReLU function is set as:
Figure BDA0003192170090000031
Figure BDA0003192170090000032
and x represents the output result of each forward propagation, the output result is mapped into a fixed range through an activation function, and the output result is used as the output function of the BP neural connection network through an overall average pooling layer and a Softmax layer.
The tongue picture image segmentation method comprises the following steps of: converting the first picture into a Y-Cb-Cr color space after passing through a VGG16 classification regression network, calculating the similarity of all pixels in the first picture according to the two-dimensional Gaussian distribution of Cb and Cr of the chroma of different tongue colors in the first picture, and calculating the tongue color probability in a way of
Figure BDA0003192170090000041
Figure BDA0003192170090000042
Figure BDA0003192170090000043
Wherein x isaFor the tongue sample value of each pixel a, t denotes the pixel mean value C as a covariance matrix, and m denotes the total number of training samples.
The tongue picture image segmentation method comprises the following steps of: calculating the first image by an AdaBoost algorithm to obtain the second image of the relatively rough tongue picture position, further executing normalization, and extracting a 10x10 pixel area from the detected tongue picture center as a tongue picture color reference, wherein the calculation mode is as follows:
Figure BDA0003192170090000044
Figure BDA0003192170090000045
tm=[Cbm,Crm]
tmexpressing the average value of different components of all the pixel points, and judging the probability value of the pixel points by calculating the Euclidean distance D between the pixel points and the average value, wherein
Figure BDA0003192170090000046
Sorting the Euclidean distances D of all the extracted pixels from small to large, taking the pixel points with the closer distances to obtain an extraction proportion, setting a threshold range according to different extraction proportions, and repeating the steps to obtain t as a finally obtained adaptive parameter; and obtaining the position of the corresponding tongue picture by displaying in different color spaces, and cutting the position of the tongue picture to obtain the third picture comprising the whole complete tongue picture.
The segmentation method of the tongue picture image, wherein the cutting of the position of the tongue picture comprises the following steps: and intersecting the tongue picture range selected by the VGG16 network frame with the tongue picture range identified by the Y-Cb-Cr color space, taking the center point of the intersection as the center of a circle, and cutting by taking the boundary distance of 1.25 times as the radius.
The tongue image segmentation method according to, wherein creating a neural connection network based on an encoder-decoder structure comprises: a standard color processing strategy, wherein the network performs standardized correction on the color of the picture by learning the difference between a standard color card and a color card in the picture; roughly positioning the position of the tongue picture by using a VGG16 classification regression network and a Y-Cb-Cr color space conversion algorithm; the texture definition of the cut tongue image is improved through an SRGAN super-resolution generation network; generating a network model by using the countermeasure, generating similar available data according to the data, and expanding the data set in a data augmentation mode; performing convolution processing and deconvolution processing on the third picture, wherein a same filling mode is used in the convolution processing, and the size of the image is kept unchanged after the image is subjected to convolution operation; in the deconvolution process, extending picture information by using bilinear difference values; meanwhile, a ResNet18 residual module is used as a convolution layer in each convolution block, and 1/2 random abandon processing is performed during convolution processing to obtain a tongue picture comprising tongue texture information and high-level information.
The segmentation method of the tongue picture image, wherein the training and testing of the third picture comprises: in the training stage, the existing data are divided into a training set and a verification set according to a proportion, and final model parameters are obtained through multiple groups of cross verification; and in the testing stage, the tongue picture obtained under the color space conversion is used as the input of the neural connection network, and a complete tongue picture segmentation picture is obtained.
The technical scheme of the invention comprises a segmentation device of a tongue image, which comprises a memory, a processor and a computer program stored in the memory and capable of running on the processor, and is characterized in that the processor realizes any one of the steps of the method when executing the computer program.
An aspect of the present invention is a computer-readable storage medium storing a computer program, wherein the computer program implements any of the method steps when executed by a processor.
The invention has the beneficial effects that: the original tongue picture is subjected to color restoration and color space conversion, so that the acquired tongue picture does not depend on hardware for acquiring the tongue picture, and the reliability of the tongue picture result is improved. And through the conversion of the color space, the position of the complete tongue picture is rapidly and roughly obtained, most of useless information is removed, the calculation pressure of the deep convolutional neural network is reduced, the calculation speed is increased, the accuracy is also improved, the hardware cost is reduced, and the applicability of the model is enhanced.
Drawings
The invention is further described below with reference to the accompanying drawings and examples;
FIG. 1 is a flow chart illustrating segmentation of a tongue image according to an embodiment of the present invention;
FIG. 2 is an overall flow diagram according to an embodiment of the present invention;
FIG. 3 is a flowchart illustrating a BP neural connection network process according to an embodiment of the present invention;
fig. 4 shows a schematic view of an apparatus according to an embodiment of the invention.
Detailed Description
Reference will now be made in detail to the present preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout.
In the description of the present invention, the meaning of a plurality of means is one or more, the meaning of a plurality of means is two or more, and larger, smaller, larger, etc. are understood as excluding the number, and larger, smaller, inner, etc. are understood as including the number.
In the description of the present invention, the consecutive reference numbers of the method steps are for convenience of examination and understanding, and the implementation order between the steps is adjusted without affecting the technical effect achieved by the technical solution of the present invention by combining the whole technical solution of the present invention and the logical relationship between the steps.
In the description of the present invention, unless otherwise explicitly defined, terms such as set, etc. should be broadly construed, and those skilled in the art can reasonably determine the specific meanings of the above terms in the present invention in combination with the detailed contents of the technical solutions.
Fig. 1 is a flowchart illustrating segmentation of a tongue image according to an embodiment of the present invention, the flowchart including: acquiring a tongue picture from acquisition equipment, performing standard color processing on the tongue picture, and outputting a first picture; performing VGG16 classification regression network detection on the first picture to detect whether the tongue picture exists, selecting a position, performing color space conversion processing, and further performing tongue picture detection to obtain a second picture comprising the rough position of the tongue picture; determining a corresponding threshold range by a Y-Cb-Cr color space conversion algorithm of the second picture, intersecting the corresponding threshold range with the frame selection range, taking the center point of the intersection as the circle center, taking the boundary distance of 1.25 times as the radius, and performing tongue picture cutting processing to obtain a third picture; and creating a Connect-Net neural network, training and testing the third picture, and outputting a complete segmentation tongue picture.
Fig. 2 is an overall flow chart of the embodiment of the invention, and the RGB color correction is because the display of the three primary colors of RGB is the same as the hardware device, and different devices have corresponding error values, so in order to make the result objectively evaluated, the picture must be processed with standard color. The calibration picture is divided into two stages, a training stage and a testing stage:
wherein the training phase comprises:
i. training by using standard color card value as known input, training by using color card value, and performing supervised training by using known color card value as supervised value
ii.24 color card for Connect-Net neural network is 24 sample values
Each sample has R, G, B channel values, the final output is R, G, B corrected values after calculation through multiple hidden layers, comparison with 24 color cards, error calculation, back propagation, and correction of all parameter values on the path according to LM (Levenberg-Marquardt algorithm)
Cycling the iii step until the error stabilizes and falls within the desired range
Training VGG16 classification regression network
vi training SRGAN super resolution generation network
Training GAN countermeasure generation network
Training Connect-Net segmentation network
The testing stage comprises the following steps:
i. initializing Connect-Net neural network by trained parameter values
ii, calculating each pixel value of the tongue picture to be corrected, and outputting to obtain a corrected picture
iii, sequentially passing the picture after color correction through VGG16, Y-Cb-Cr color space conversion algorithm, SRGAN super-resolution generation network and Connect-Net segmentation network
The connection-Net segmentation neural network model has fewer training samples and a single output result dimension, so that the network structure is relatively simple to design, in order to prevent the over-fitting situation, a Dropout strategy is used for randomly discarding part of neuron information, a ReLU function is used as an activation function of the network, and Global Average Power and Softmax layers are used as output functions of the connection-Net neural network. Wherein x represents the output result of each forward propagation and is mapped into a fixed range through a formula.
Figure 3
Figure BDA0003192170090000082
Wherein x represents the output result of each forward propagation and is mapped into a fixed range through a formula.
In the RGB color space, the tongue color and the non-tongue color are overlapped on R, G, B three components, and generally, the image needs to be subjected to color space conversion to be converted into a Y-Cb-Cr color space which is irrelevant to brightness and chroma, so as to improve the detection rate. In the Y-Cb-Cr color space, the tongue colors have certain clustering characteristics, and the Cb and Cr distribution of the chroma of different tongue colors is approximately equal to two-dimensional Gaussian distribution. The tongue color probability can be estimated by calculating the similarity of all pixels, and the probability estimation formula is as follows:
Figure BDA0003192170090000083
in the formula: x is the number ofaFor the tongue sample value of each pixel a, t denotes the pixel mean value C as a covariance matrix, and m denotes the total number of training samples.
In order to improve the robustness of the detection model, the following improvement is made to the above formula.
Obtaining a relatively rough tongue picture position through a VGG16 classification regression network algorithm and a Y-Cb-Cr color space algorithm, carrying out normalization, and extracting a 10x10 pixel region from the detected tongue picture center as a tongue picture color reference, then:
Figure BDA0003192170090000084
tm=[Cbm,Crm]
in the formula: t is tmExpressing the average value of different components of all pixel points by calculating the Euclidean between the pixel points and the average valueAnd judging the probability value of the pixel point by the distance D.
Figure BDA0003192170090000085
Sorting the Euclidean distances of all pixels from small to large, taking the pixel points with the closer distances to obtain an extraction ratio sigma, and setting a threshold range according to different ratios.
And repeating the steps to obtain t as the finally obtained adaptive parameter.
And determining a corresponding threshold range through a Y-Cb-Cr color space conversion algorithm, intersecting the threshold range with the frame selection range, taking the center point of the intersection as the center of a circle, taking the boundary distance of 1.25 times as the radius, and performing tongue picture cutting processing to obtain a picture containing the whole complete tongue picture.
FIG. 3 is a flowchart illustrating a BP neural connection network processing according to an embodiment of the present invention, which uses the Connect-Net network to obtain a precise segmentation tongue image in three stages:
stage of designing network
i. A network model is redesigned aiming at the problems;
firstly, generating similar available data according to the existing small amount of data by using a confrontation generation network model, and expanding a data set in a data augmentation mode;
using a same filling mode in the convolution step to enable the image to keep the original size after the convolution operation;
the deconvolution process uses bilinear difference values to expand the picture information;
v. using the ResNet18 residual module as a convolutional layer in each block module, preventing the network from training overfitting by random drop1/2 parametric magnitude operation, and more tongue texture information and high-level information can be obtained.
Training phase
i. Dividing the existing data into a training set and a testing set according to the ratio of (8: 2);
dividing the training set into a training set and a verification set according to the ratio of (5: 1);
and ii, obtaining final model parameters through 5 sets of cross validation.
Testing phase
i. Taking the tongue picture obtained by color space conversion as the input of an improved network;
and ii, obtaining a complete segmentation picture.
Fig. 4 shows a schematic view of an apparatus according to an embodiment of the invention. The apparatus comprises a memory 100 and a processor 200, wherein the processor 200 stores a computer program, and the computer program implements the following method flows when executed by the processor 200: acquiring a tongue picture, and performing standard color processing on the tongue picture to obtain a first picture; establishing a VGG16 classification regression network, classifying the first picture, acquiring tongue information of the first picture, performing frame selection on the tongue information and performing color space conversion processing on the first picture, and further performing tongue picture detection to obtain a second picture comprising a rough tongue picture position; determining a corresponding threshold range of the second picture according to a Y-Cb-Cr color space conversion algorithm, intersecting the threshold range with the frame selection range, and performing tongue picture cutting processing to obtain a third picture; creating an SRGAN super-resolution generation network, and performing multi-detail reduction on the third picture to generate a fourth image with higher definition; and (4) creating a neural connection network based on an encoder-decoder structure, training and testing the fourth picture, and outputting a complete segmentation tongue picture.
It should be recognized that the method steps in embodiments of the present invention may be embodied or carried out by computer hardware, a combination of hardware and software, or by computer instructions stored in a non-transitory computer readable memory. The method may use standard programming techniques. Each program may be implemented in a high level procedural or object oriented programming language to communicate with a computer system. However, the program(s) can be implemented in assembly or machine language, if desired. In any case, the language may be a compiled or interpreted language. Furthermore, the program can be run on a programmed application specific integrated circuit for this purpose.
Further, the operations of processes described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The processes described herein (or variations and/or combinations thereof) may be performed under the control of one or more computer systems configured with executable instructions, and may be implemented as code (e.g., executable instructions, one or more computer programs, or one or more applications) collectively executed on one or more processors, by hardware, or combinations thereof. The computer program includes a plurality of instructions executable by one or more processors.
Further, the method may be implemented in any type of computing platform operatively connected to a suitable interface, including but not limited to a personal computer, mini computer, mainframe, workstation, networked or distributed computing environment, separate or integrated computer platform, or in communication with a charged particle tool or other imaging device, and the like. Aspects of the invention may be embodied in machine-readable code stored on a non-transitory storage medium or device, whether removable or integrated into a computing platform, such as a hard disk, optically read and/or write storage medium, RAM, ROM, or the like, such that it may be read by a programmable computer, which when read by the storage medium or device, is operative to configure and operate the computer to perform the procedures described herein. Further, the machine-readable code, or portions thereof, may be transmitted over a wired or wireless network. The invention described herein includes these and other different types of non-transitory computer-readable storage media when such media include instructions or programs that implement the steps described above in conjunction with a microprocessor or other data processor. The invention also includes the computer itself when programmed according to the methods and techniques described herein.
A computer program can be applied to input data to perform the functions described herein to transform the input data to generate output data that is stored to non-volatile memory. The output information may also be applied to one or more output devices, such as consumers. In a preferred embodiment of the present invention, the transformed data represents physical and tangible objects, including particular visual depictions of physical and tangible objects produced on the consumer.
The embodiments of the present invention have been described in detail with reference to the accompanying drawings, but the present invention is not limited to the above embodiments, and various changes can be made within the knowledge of those skilled in the art without departing from the gist of the present invention.

Claims (11)

1.一种舌象图像的分割方法,其特征在于,该方法包括:1. a segmentation method of tongue image, is characterized in that, this method comprises: 获取舌象图片,对所述舌象图片进行标准色处理,得到第一图片;Obtaining a picture of the tongue image, carrying out standard color processing to the picture of the tongue image, and obtaining the first picture; 创建VGG16分类回归网络,对第一图片进行分类,获取所述第一图的舌头信息,对所述舌头信息进行框选并且对所述第一图片进行颜色空间转换处理,进而执行舌象检测,得到包括有舌象粗略位置的第二图片;Create a VGG16 classification and regression network, classify the first picture, obtain the tongue information of the first picture, frame the tongue information and perform color space conversion processing on the first picture, and then perform tongue image detection, Obtain a second picture including the rough position of the tongue image; 对所述第二图片根据Y-Cb-Cr颜色空间转换算法确定相应的阈值范围并和框选范围做交集,进行舌象裁剪处理,得到第三图片;Determine the corresponding threshold range according to the Y-Cb-Cr color space conversion algorithm for the second picture and make an intersection with the frame selection range, and carry out the tongue image cropping process to obtain the third picture; 创建SRGAN超分辨率生成网络,对第三图片进行多细节还原,生成清晰度较高的第四图像;Create a SRGAN super-resolution generation network, restore the third image with multiple details, and generate a fourth image with higher definition; 创建基于编码器-解码器结构的神经连接网络,对所述第四图片进行训练及测试,输出完整的分割舌象图片。A neural connection network based on an encoder-decoder structure is created, the fourth picture is trained and tested, and a complete segmented tongue picture is output. 2.根据权利要求1所述的舌象图像的分割方法,其特征在于,所述对所述舌象图片进行标准色处理包括:2. The method for segmenting a tongue image according to claim 1, wherein the performing standard color processing on the tongue image comprises: 创建BP神经连接网络,以标准色卡值作为已知输入对所述舌象图片进行训练,其中,已知输入的色卡值作为监督值,执行监督训练,得到已训练的参数值;Create a BP neural connection network, and use the standard color swatch value as a known input to train the tongue image, wherein, the known input color swatch value is used as a supervision value, and supervised training is performed to obtain the trained parameter values; 通过所述参数值初始化所述BP神经连接网络,输入所述舌象图片进行计算,输出校正后的第一图片,所述第一图片为校正后的所述舌象图片。The BP neural connection network is initialized by the parameter value, the tongue image is input for calculation, and the corrected first picture is output, and the first picture is the corrected tongue image. 3.根据权利要求2所述的舌象图像的分割方法,其特征在于,所述以标准色卡值作为已知输入对所述舌象图片进行训练包括:3. The method for segmenting a tongue image according to claim 2, wherein the training on the tongue image with a standard color card value as a known input comprises: 以所述标准色卡值的24色作为所述BP神经连接网络的样本值,对每个所述样本值的RGB通道值执行多个隐层计算,输出校正后的所述RGB通道值;Taking 24 colors of the standard color card values as the sample values of the BP neural connection network, performing multiple hidden layer calculations on the RGB channel values of each of the sample values, and outputting the corrected RGB channel values; 将校正后的所述RGB通道值与所述标准色卡值进行对比,计算误差并进行反向传播,根据LM算法修正路径上所有参数值;Compare the corrected RGB channel value with the standard color card value, calculate the error and carry out back propagation, and correct all parameter values on the path according to the LM algorithm; 循环上一步骤,直至将所有参数值的误差降至设定范围内。Repeat the previous step until the error of all parameter values falls within the set range. 4.根据权利要求2所述的舌象图像的分割方法,其特征在于,所述BP神经连接网络包括:4. The segmentation method of tongue image according to claim 2, wherein the BP neural connection network comprises: 标准色处理网络,通过24色真实色卡值和不同场景图片中色卡值之间的差异,让神经网络学习之间的非线性关系,通过反向传播更新梯度值来训练网络的权重模型;Standard color processing network, through the difference between the 24-color real color card value and the color card value in different scene pictures, let the neural network learn the nonlinear relationship between them, and update the gradient value through back propagation to train the weight model of the network; 通过VGG16分类回归网络通过现有舌象标签来训练网络,通过现有数据通过图片增广策略训练网络,得到5个维度的结果,是否是舌象和位置信息;Through the VGG16 classification and regression network, the network is trained through the existing tongue image labels, and the network is trained through the image augmentation strategy through the existing data, and the results of 5 dimensions are obtained, whether it is the tongue image and location information; 采用Y-Cb-Cr颜色空间转换算法处理,包括通过颜色聚类策略来粗略确定舌象的位置信息;It is processed by Y-Cb-Cr color space conversion algorithm, including roughly determining the position information of tongue image through color clustering strategy; 通过SRGAN超分辨率生成网络和舌象分割网络;Through SRGAN super-resolution generation network and tongue image segmentation network; 使用ReLU(Rectified Linear Unit)函数作为所述BP神经连接网络的激活函数,ReLU函数设置为:Use the ReLU (Rectified Linear Unit) function as the activation function of the BP neural connection network, and the ReLU function is set to:
Figure FDA0003192170080000021
Figure FDA0003192170080000021
Figure FDA0003192170080000022
Figure FDA0003192170080000022
式中x表示每次前向传播的输出结果,通过激活函数将输出结果映射到固定范围内,通过整体平均池化层和Softmax层作为所述BP神经连接网络的输出函数。In the formula, x represents the output result of each forward propagation, and the output result is mapped to a fixed range through the activation function, and the overall average pooling layer and the Softmax layer are used as the output function of the BP neural connection network.
5.根据权利要求1所述的舌象图像的分割方法,其特征在于,所述执行舌象检测包括:5. The method for segmenting a tongue image according to claim 1, wherein the performing tongue image detection comprises: 将所述第一图片通过VGG16分类回归网络后转换至Y-Cb-Cr颜色空间,根据所述第一图片中不同舌色的色度的Cb和Cr的二维高斯分布,计算所述第一图片中所有像素的相似程度,进行舌色概率的计算,计算方式为The first picture is converted to the Y-Cb-Cr color space after passing through the VGG16 classification and regression network, and the first picture is calculated according to the two-dimensional Gaussian distribution of Cb and Cr of the chromaticity of different tongue colors in the first picture. The similarity of all pixels in the picture is used to calculate the tongue color probability. The calculation method is as follows:
Figure FDA0003192170080000023
Figure FDA0003192170080000023
Figure FDA0003192170080000024
Figure FDA0003192170080000024
Figure FDA0003192170080000025
Figure FDA0003192170080000025
其中,xa为每个像素a的舌色样本值,t表示像素平均值C为协方差矩阵,m表示训练样本总数。Among them, x a is the tongue color sample value of each pixel a, t is the pixel average value, C is the covariance matrix, and m is the total number of training samples.
6.根据权利要求4所述的舌象图像的分割方法,其特征在于,所述执行舌象检测还包括:6. The segmentation method of tongue image according to claim 4, wherein the performing tongue image detection further comprises: 通过AdaBoost算法对所述第一图像进行计算,得到相对粗略的舌象位置的所述第二图片,进而执行归一化里,将检测到舌象中心提取10x10像素区域作为舌象颜色基准,计算方式为:The first image is calculated by the AdaBoost algorithm to obtain the second picture of the relatively rough tongue image position, and then normalization is performed, and a 10×10 pixel area is extracted from the center of the detected tongue image as the tongue image color reference. The way is:
Figure FDA0003192170080000031
Figure FDA0003192170080000031
Figure FDA0003192170080000032
Figure FDA0003192170080000032
tm=[Cbm,Crm]t m =[Cb m ,Cr m ] tm表示所有像素点的不同分量的平均值,通过计算像素点到平均值之间的欧式距离D来判断像素点所属的概率值,其中t m represents the average value of different components of all pixel points, and the probability value to which the pixel point belongs is determined by calculating the Euclidean distance D between the pixel point and the average value, where
Figure FDA0003192170080000033
Figure FDA0003192170080000033
将提取的所有像素的欧式距离D进行从小到大排序,取距离较近的像素点,得到提取比例,通过不同提取比例来设置阈值范围,重复以上步骤得到的t为最终得到的自适应参数;The Euclidean distance D of all the extracted pixels is sorted from small to large, and the pixels with the shorter distance are selected to obtain the extraction ratio. The threshold range is set by different extraction ratios, and the t obtained by repeating the above steps is the final adaptive parameter obtained; 通过不同颜色空间下显示得到了得到对应舌象的位置,对舌象的位置进行裁剪,得到包括有整个完整舌象的所述第三图片。The position of the corresponding tongue image is obtained by displaying in different color spaces, and the position of the tongue image is cropped to obtain the third picture including the entire complete tongue image.
7.根据权利要求6所述的舌象图像的分割方法,其特征在于,所述对舌象的位置进行裁剪包括:对VGG16网络框选的舌象范围和Y-Cb-Cr颜色空间识别的舌象范围做交集,取交集中心点做圆心,边界距离1.25倍为半径进行裁剪。7. The method for segmenting a tongue image according to claim 6, wherein the clipping of the position of the tongue image comprises: the range of the tongue image selected by the VGG16 network and the identification of the Y-Cb-Cr color space. The range of the tongue image is used as the intersection, the center of the intersection is taken as the center of the circle, and the boundary distance is 1.25 times the radius for cutting. 8.根据权利要求1所述的舌象图像的分割方法,其特征在于,所述创建基于编码器-解码器结构的神经连接网络包括:8. The method for segmenting a tongue image according to claim 1, wherein the creating a neural connection network based on an encoder-decoder structure comprises: 标准色处理策略,网络通过学习标准色卡和图像中色卡的差异来,将图片的颜色进行标准化矫正;Standard color processing strategy, the network standardizes and corrects the color of the picture by learning the difference between the standard color card and the color card in the image; 通过VGG16分类回归网络和Y-Cb-Cr颜色空间转换算法来粗略定位舌象的位置;Roughly locate the position of tongue image through VGG16 classification and regression network and Y-Cb-Cr color space conversion algorithm; 通过SRGAN超分辨率生成网络来提升裁剪后的舌象的纹理清晰度;Improve the texture clarity of the cropped tongue image through the SRGAN super-resolution generation network; 使用对抗生成网络模型并根据若干数据生成相似可用数据,通过数据增广方式进行数据集的扩充;Use the adversarial generative network model and generate similar available data based on several data, and expand the data set through data augmentation; 对所述第三图片执行卷积处理及反卷积处理,其中所述卷积处理中使用same填充方式,其中图像通过卷积操作后保持大小不变;Performing convolution processing and deconvolution processing on the third picture, wherein the same filling method is used in the convolution processing, and the size of the image remains unchanged after the convolution operation; 反卷积过程使用双线性差值扩展图片信息;The deconvolution process uses bilinear difference to expand the picture information; 同时在每个卷积块中使用ResNet18残差模块作为卷积层,卷积处理时进行1/2的随机舍弃处理,得到包括有舌象纹理信息和高层信息的舌象图片。At the same time, the ResNet18 residual module is used as the convolution layer in each convolution block, and 1/2 of the convolution process is randomly discarded to obtain a tongue image including tongue texture information and high-level information. 9.根据权利要求1所述的舌象图像的分割方法,其特征在于,所述对所述第三图片进行训练及测试包括:9. The segmentation method of tongue image according to claim 1, wherein the third picture is trained and tested comprising: 训练阶段,通过现有数据将其按照比例分为训练集和验证集,通过多组交叉验证得到最终模型参数;In the training phase, the existing data is divided into training set and verification set according to the proportion, and the final model parameters are obtained through multiple sets of cross-validation; 测试阶段,将颜色空间转换下得到的舌象图片作为所述神经连接网络的输入,并得到完整的舌象分割图片。In the testing stage, the tongue image obtained by color space conversion is used as the input of the neural connection network, and a complete tongue image segmentation image is obtained. 10.一种舌象图像的分割装置,该装置包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现权利要求1-9任一所述的方法步骤。10. A device for segmenting a tongue image, the device comprising a memory, a processor and a computer program stored in the memory and running on the processor, wherein the processor executes the computer The program implements the method steps of any one of claims 1-9. 11.一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1-9任一所述的方法步骤。11. A computer-readable storage medium storing a computer program, characterized in that, when the computer program is executed by a processor, the method steps according to any one of claims 1-9 are implemented.
CN202110881511.9A 2021-08-02 2021-08-02 Tongue image segmentation method, device and medium Active CN113781488B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110881511.9A CN113781488B (en) 2021-08-02 2021-08-02 Tongue image segmentation method, device and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110881511.9A CN113781488B (en) 2021-08-02 2021-08-02 Tongue image segmentation method, device and medium

Publications (2)

Publication Number Publication Date
CN113781488A true CN113781488A (en) 2021-12-10
CN113781488B CN113781488B (en) 2025-01-28

Family

ID=78836501

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110881511.9A Active CN113781488B (en) 2021-08-02 2021-08-02 Tongue image segmentation method, device and medium

Country Status (1)

Country Link
CN (1) CN113781488B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115601358A (en) * 2022-12-01 2023-01-13 合肥云诊信息科技有限公司(Cn) Tongue picture image segmentation method under natural light environment
CN116509326A (en) * 2023-04-11 2023-08-01 广东健齿生物科技有限公司 Tongue image multispectral image generation method, device, equipment and storage medium
CN117094966A (en) * 2023-08-21 2023-11-21 青岛美迪康数字工程有限公司 Tongue image identification method and device based on image amplification and computer equipment
CN117197139A (en) * 2023-11-07 2023-12-08 天津市肿瘤医院(天津医科大学肿瘤医院) Tongue diagnosis image multi-label classification method based on AI
CN118629660A (en) * 2024-08-15 2024-09-10 青岛言鼎生物医疗科技有限公司 A method for constructing a health status prediction model based on tongue image characteristics
CN118691608A (en) * 2024-08-23 2024-09-24 北京易康医疗科技有限公司 Tongue image processing method, device, equipment and storage medium

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103106386A (en) * 2011-11-10 2013-05-15 华为技术有限公司 Dynamic self-adaption skin color segmentation method and device
KR20140104558A (en) * 2013-02-19 2014-08-29 대전대학교 산학협력단 New tongue diagnosis model system developed by analyzing the color of the tongue
CN105303152A (en) * 2014-07-15 2016-02-03 中国人民解放军理工大学 Human body re-recognition method
CN106909883A (en) * 2017-01-17 2017-06-30 北京航空航天大学 A kind of modularization hand region detection method and device based on ROS
CN107316307A (en) * 2017-06-27 2017-11-03 北京工业大学 A kind of Chinese medicine tongue image automatic segmentation method based on depth convolutional neural networks
CN107977671A (en) * 2017-10-27 2018-05-01 浙江工业大学 A kind of tongue picture sorting technique based on multitask convolutional neural networks
CN108734108A (en) * 2018-04-24 2018-11-02 浙江工业大学 A kind of fissured tongue recognition methods based on SSD networks
CN109815860A (en) * 2019-01-10 2019-05-28 中国科学院苏州生物医学工程技术研究所 TCM tongue diagnosis image color correction method, electronic equipment, storage medium
CN110136062A (en) * 2019-05-10 2019-08-16 武汉大学 A Super-resolution Reconstruction Method for Joint Semantic Segmentation
US20210038198A1 (en) * 2019-08-07 2021-02-11 Siemens Healthcare Gmbh Shape-based generative adversarial network for segmentation in medical imaging

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103106386A (en) * 2011-11-10 2013-05-15 华为技术有限公司 Dynamic self-adaption skin color segmentation method and device
KR20140104558A (en) * 2013-02-19 2014-08-29 대전대학교 산학협력단 New tongue diagnosis model system developed by analyzing the color of the tongue
CN105303152A (en) * 2014-07-15 2016-02-03 中国人民解放军理工大学 Human body re-recognition method
CN106909883A (en) * 2017-01-17 2017-06-30 北京航空航天大学 A kind of modularization hand region detection method and device based on ROS
CN107316307A (en) * 2017-06-27 2017-11-03 北京工业大学 A kind of Chinese medicine tongue image automatic segmentation method based on depth convolutional neural networks
CN107977671A (en) * 2017-10-27 2018-05-01 浙江工业大学 A kind of tongue picture sorting technique based on multitask convolutional neural networks
CN108734108A (en) * 2018-04-24 2018-11-02 浙江工业大学 A kind of fissured tongue recognition methods based on SSD networks
CN109815860A (en) * 2019-01-10 2019-05-28 中国科学院苏州生物医学工程技术研究所 TCM tongue diagnosis image color correction method, electronic equipment, storage medium
CN110136062A (en) * 2019-05-10 2019-08-16 武汉大学 A Super-resolution Reconstruction Method for Joint Semantic Segmentation
US20210038198A1 (en) * 2019-08-07 2021-02-11 Siemens Healthcare Gmbh Shape-based generative adversarial network for segmentation in medical imaging

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115601358A (en) * 2022-12-01 2023-01-13 合肥云诊信息科技有限公司(Cn) Tongue picture image segmentation method under natural light environment
CN116509326A (en) * 2023-04-11 2023-08-01 广东健齿生物科技有限公司 Tongue image multispectral image generation method, device, equipment and storage medium
CN117094966A (en) * 2023-08-21 2023-11-21 青岛美迪康数字工程有限公司 Tongue image identification method and device based on image amplification and computer equipment
CN117094966B (en) * 2023-08-21 2024-04-05 青岛美迪康数字工程有限公司 Tongue image identification method and device based on image amplification and computer equipment
CN117197139A (en) * 2023-11-07 2023-12-08 天津市肿瘤医院(天津医科大学肿瘤医院) Tongue diagnosis image multi-label classification method based on AI
CN117197139B (en) * 2023-11-07 2024-02-02 天津市肿瘤医院(天津医科大学肿瘤医院) Tongue diagnosis image multi-label classification method based on AI
CN118629660A (en) * 2024-08-15 2024-09-10 青岛言鼎生物医疗科技有限公司 A method for constructing a health status prediction model based on tongue image characteristics
CN118691608A (en) * 2024-08-23 2024-09-24 北京易康医疗科技有限公司 Tongue image processing method, device, equipment and storage medium
CN118691608B (en) * 2024-08-23 2024-12-03 北京易康医疗科技有限公司 Tongue image processing method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN113781488B (en) 2025-01-28

Similar Documents

Publication Publication Date Title
US10482603B1 (en) Medical image segmentation using an integrated edge guidance module and object segmentation network
CN113781488B (en) Tongue image segmentation method, device and medium
CN110120040B (en) Slice image processing method, slice image processing device, computer equipment and storage medium
CN114399510B (en) Skin focus segmentation and classification method and system combining image and clinical metadata
CN111539911B (en) Mouth breathing face recognition method, device and storage medium
US20230377097A1 (en) Laparoscopic image smoke removal method based on generative adversarial network
CN114649092B (en) Auxiliary diagnosis method and device based on semi-supervised learning and multi-scale feature fusion
CN115049602B (en) Optimization Method of Artificial Neural Network Module
Gamage et al. Instance-based segmentation for boundary detection of neuropathic ulcers through Mask-RCNN
Salama et al. Enhancing medical image quality using neutrosophic fuzzy domain and multi-level enhancement transforms: A comparative study for leukemia detection and classification
CN119444700A (en) Colorectal lesion classification method based on KAN network and variable attention mechanism
Wang et al. Optic disc detection based on fully convolutional neural network and structured matrix decomposition
CN118116576A (en) Intelligent case analysis method and system based on deep learning
Codella et al. Segmentation of both diseased and healthy skin from clinical photographs in a primary care setting
CN117315378A (en) Grading judgment method for pneumoconiosis and related equipment
CN116453657A (en) Kidney image processing method and system based on convolutional neural network
CN118429680A (en) Method and system for identifying and predicting tongue picture full-class label
Thokar et al. Medical Image Segmentation for Anomaly Detection Using Deep Learning Techniques
Aydoghmishi Skin Cancer Detection by Deep Learning Algorithms
Wei et al. Application of U-net with variable fractional order gradient descent method in rectal tumor segmentation
Paul et al. Computer-Aided Diagnosis Using Hybrid Technique for Fastened and Accurate Analysis of Tuberculosis Detection with Adaboost and Learning Vector Quantization
US12125200B2 (en) Methods, devices, and systems for determining presence of appendicitis
Bora et al. Lung Cancer Recognition using CNN Algorithm and MATLAB
Yuanbo Application of Object Detection and Image Generation in Capsule Endoscopy Assisted Diagnosis System
Mohammadi Aydoghmishi Skin Cancer Detection by Deep Learning Algorithms

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant