WO2022198898A1 - 图像分类方法和装置及设备 - Google Patents
图像分类方法和装置及设备 Download PDFInfo
- Publication number
- WO2022198898A1 WO2022198898A1 PCT/CN2021/112932 CN2021112932W WO2022198898A1 WO 2022198898 A1 WO2022198898 A1 WO 2022198898A1 CN 2021112932 W CN2021112932 W CN 2021112932W WO 2022198898 A1 WO2022198898 A1 WO 2022198898A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- classified
- signal waveform
- array
- pixel
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/86—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using syntactic or structural representations of the image or video pattern, e.g. symbolic string recognition; using graph matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/22—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
- G06V10/225—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition based on a marking or identifier characterising the area
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/30—Noise filtering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/34—Smoothing or thinning of the pattern; Morphological operations; Skeletonisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/42—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/12—Classification; Matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
Definitions
- the present invention relates to the technical field of image classification and recognition, and in particular, to an image classification method, device and equipment.
- an image recognition model or a currently popular deep learning network model CNN can be used to implement image classification.
- a recognition model needs to be built in advance. In the process of building a recognition model, several stages such as feature extraction, feature coding, space constraints, classifier design, and model fusion are required. The research and development cycle is long, and the requirements for algorithm designers’ own professional skills and algorithm deployment are high.
- the deep learning network model CNN to achieve image classification, although it can liberate manpower and achieve fast and efficient automatic sorting, it takes a lot of time to build a neural network model, label training data sets, train model parameters, and optimize in the early stage.
- the required technical threshold and hardware configuration are high, which makes the reproducibility and generalization ability of the traditional image classification method weak, and it is difficult to achieve out-of-the-box, rapid replication, and generalization to various complex and changeable clinical situations. in a medical setting.
- the present invention provides the following scheme:
- an image classification method comprising:
- the image data set includes images to be classified
- the category of the image to be classified is determined based on the signal waveform diagram.
- any single-channel pixel value array in the standard image select any single-channel pixel value array in the standard image, and draw a corresponding signal waveform diagram based on the pixel value array, including:
- the corresponding signal waveform diagram is drawn according to the row pixel compression array and the column pixel compression array after the smoothing process.
- axis represents the dimension of the summation of pixel values
- the method further includes: taking the logarithm of the smoothed row pixel compression array and the column pixel compression array. The operation of arithmetic processing.
- determining the category of the image to be classified based on the signal waveform diagram includes:
- the category of the image to be classified is determined.
- the method when drawing the corresponding signal waveform diagram based on the pixel value array, the method further includes:
- the processing of the signal waveform graph according to the sharpness of each peak includes: an operation of removing peaks whose sharpness is less than a preset minimum threshold of sharpness.
- it also includes:
- the image to be classified is preprocessed, the marker area in the image to be classified is located, and the category of the image to be classified is determined according to the shape of the located marker area.
- an image classification apparatus including: a data input module, an image interception module, a waveform diagram drawing module, and a first category determination module;
- the data input module is configured to input an image data set to be classified; wherein, the image data set includes images to be classified;
- the image interception module is configured to take the center of the image to be classified as a reference center point, and intercept an image of a preset size including a specific area as a standard image;
- the waveform graph drawing module is configured to select any single-channel pixel value array in the standard image, and draw a corresponding signal waveform graph based on the pixel value array;
- the first category determination module is configured to determine the category of the image to be classified based on the signal waveform diagram.
- an image classification device comprising:
- memory for storing processor-executable instructions
- the processor is configured to implement any of the foregoing methods when executing the executable instructions.
- any single-channel pixel value array is selected from the standard image obtained by the interception, and then the signal waveform diagram is drawn according to the selected single-channel pixel value array, so as to be classified.
- the image is classified into categories, it is divided based on the signal waveform obtained by drawing.
- the image recognition model and deep learning network model CNN can not only complete the classification of various types of images quickly, accurately and efficiently, but also only need to perform the above processing on the images to be classified to generate corresponding signal waveforms.
- the image is enough, and there is no need to collect and label a large amount of sample data, nor to train the recognition model.
- the sample data can be more suitable for various image classification application scenarios, and finally effectively improve the reproducibility and generalization ability of the image classification method.
- FIG. 1 is a flowchart of an image classification method according to an embodiment of the application
- FIG. 3 is a distribution diagram of hue thresholds of different colors in the HSV color space on which images to be classified in an image data set are filtered according to image colors in an image classification method according to an embodiment of the application;
- FIG. 4 is a schematic diagram of the sharpness detection of each peak in a drawn signal waveform diagram in an image classification method according to an embodiment of the present application
- Fig. 5a and Fig. 5b are respectively the signal waveform diagram corresponding to the fundus diagram and the signal waveform diagram corresponding to the outer eye diagram in the image classification method finally drawn in an embodiment of the application;
- FIG. 6 is an effect diagram of an image classification method according to an embodiment of the present application, when an image to be classified is detected as an outer eye diagram according to the shape of the located pupil region;
- FIG. 7 is a structural block diagram of an image classification apparatus according to an embodiment of the application.
- FIG. 8 is a structural block diagram of an image classification apparatus according to an embodiment of the present application.
- FIG. 1 shows a flowchart of an image classification method according to an embodiment of the present application.
- the method includes: step S100 , inputting an image data set to be classified.
- the image data set may include various images such as OCT, eye B-ultrasound, graphic report, FFA image, fundus image collected by fundus camera, and outer eye image.
- Step S200 taking the center of the image to be classified as the reference center point, and intercepting an image of a preset size including a specific area as a standard image.
- step S300 a pixel value array of any single channel in the standard image is selected, and a corresponding signal waveform diagram is drawn based on the pixel value array.
- step S400 the category of the image to be classified is determined based on the obtained signal waveform diagram.
- any single-channel pixel value array is selected from the standard image obtained by the interception, and then according to the selected single-channel pixel value array
- the pixel value array draws a signal waveform diagram, so that when the image to be classified is classified into categories, the division is performed based on the drawn signal waveform diagram.
- the image recognition model and deep learning network model CNN can not only complete the classification of various types of images quickly, accurately and efficiently, but also only need to perform the above processing on the images to be classified to generate corresponding signal waveforms.
- the image is enough, and there is no need to collect and label a large amount of sample data, nor to train the recognition model.
- the sample data can be more suitable for various image classification application scenarios, and finally effectively improve the reproducibility and generalization ability of the image classification method.
- the intercepted specific area is different from the image to be classified contained in the image dataset. associated with the class to which it belongs.
- the images to be classified contained in the image data set are multiple, and the categories of different images to be classified are different, they all belong to the same category.
- the images to be classified included in the image data set all belong to the category of eye detection data, but different images to be classified correspond to different image data under the eye detection images.
- the image to be classified may be a fundus image collected by a fundus camera, or an external eye image, and the image to be classified may also be an eye B ultrasound, OCT, FFA, and the like.
- the images to be classified included in the image data set all belong to the category of chest detection data
- the images to be classified included in the image data set may be data such as chest CT and chest B-ultrasound.
- the images to be classified in the image data set may also be image data collected in other application fields, which will not be described one by one here.
- the image data set may be different image data collected in different application scenarios.
- the images to be classified in the image data set should all belong to different forms or different categories of image data obtained in the same application scenario.
- the specific area to be intercepted can be determined based on the specific application scenario to which each image to be classified in the image data set belongs. Also taking the eye detection data in clinical medical images as an example, the intercepted specific area is to envelop the entire pupil or macular area.
- the preset size can also be set according to the specific application scenario to which the image to be classified in the image data set belongs. The preset sizes set in different application scenarios are different.
- an operation of filtering the images to be classified in the image data set according to the image size and/or image color may be further included.
- the images to be classified in the image data set are filtered according to the two pieces of information of graphic size and image color, the sequence of the images can be flexibly set according to the actual situation.
- step S021 may be used first. Filter the images to be classified in the image data set according to the image size, and filter out small-sized image data such as OCT, eye B-ultrasound, and graphic reports in the image data set. Then, through step S022, the remaining images to be classified in the filtered image data set are filtered again according to the image color, and data such as FFA images are identified and filtered.
- the color recognition principle used can be specifically based on the “hue threshold distribution of different colors in the HSV color space” (see Figure 3). Divide.
- the FFA image is a grayscale image
- the fundus image and the outer eye image are all RGB three-channel color images.
- Using color recognition can further subdivide the three types of images and filter out the FFA images.
- the chromaticity values of the fundus map and the outer eye map are located in the hue interval of ['red2', 'red', 'orange'].
- step S200 can be executed, taking the center of the image to be classified as the reference center point, and intercepting an image of a preset size including a specific area as a standard image.
- the selection of the specific area and the setting of the preset size may be performed according to the specific application scenario to which the image data set belongs.
- the intercepted specific area is to envelop the entire through hole or macular area.
- the default size can be set to a side length of 700px.
- step S300 can be executed to select any single-channel pixel value array in the standard image, and draw a corresponding signal waveform diagram based on the pixel value array.
- the selection can be made according to the background color of the image to be classified currently being identified and divided. That is, in the image classification method of the embodiment of the present application, the selected single channel should be a channel that is close to the background color of the image to be classified that is currently being identified and divided.
- the pixel value array of the single channel R in the standard image can be selected, and then the signal waveform diagram can be generated based on the pixel value array of the selected R channel.
- selecting any single-channel pixel value array in the standard image, and drawing a corresponding signal waveform diagram based on the pixel value array may be implemented in the following manner.
- any single-channel pixel value array in the standard image select any single-channel pixel value array in the standard image. Then, perform row compression and column compression on the row pixel array and the column pixel array in the pixel value array, respectively, to obtain the row pixel compression array and the column pixel compression array. Furthermore, curve smoothing is performed on the row pixel compression array and the column pixel compression array to remove noise points in the row pixel compression array and the column pixel compression array. Finally, the corresponding signal waveform diagram is drawn according to the row pixel compression array and the column pixel compression array after smoothing.
- axis represents the dimension of the summation of pixel values
- curve smoothing can be performed on the row pixel compression array and the column pixel compression array.
- smoothing can be performed by calling savgol_filter in the cipy.signal library to remove noise points in the row pixel compression array and the column pixel compression array.
- calculation formula for performing curve smoothing filtering on the row pixel compression array and the column pixel compression array is as follows:
- hi is the smoothing coefficient
- the operation of performing logarithmic operation processing on the smoothed row pixel compression array and the column pixel compression array is also included.
- logarithmic operations on the row pixel compression array and column pixel compression array after smoothing not only the absolute value of the data can be reduced, which is convenient for calculation, but also the variable scale can be compressed without changing the nature and correlation of the data. , which weakens the collinearity, heteroscedasticity, etc. of the model.
- the plt.plot function in the matplotlib.pyplot library can be directly called when drawing the signal waveform of the compressed pixel value.
- step S400 may be executed to determine the category of the image to be classified based on the signal waveform diagram.
- the percentage of change rate of pixel values in the signal waveform curve, the sharpness of each peak in the signal waveform curve, and the signal waveform can be obtained according to the drawn signal waveform diagram.
- the percentage of column amplitude to row amplitude in the curve, and then according to at least one of the percentage of change rate of pixel values in the signal waveform curve, the sharpness of each peak in the signal waveform graph, and the percentage of column amplitude to row amplitude in the signal waveform curve The identification and determination of the category of the image to be classified is performed.
- the percentage change rate of the pixel value in the signal waveform curve can be calculated based on the drawn signal waveform graph. Its calculation formula is:
- delta is the percentage change of the pixel value in the signal waveform curve
- max(y smooth ) is the maximum pixel value after Savitzky-Golay smoothing filtering
- min(y smooth ) is the minimum pixel value after Savitzky-Golay smoothing filtering .
- the sharpness of each peak in the signal waveform curve can be implemented in the following manner.
- each peak in the signal waveform diagram is detected based on the peak properties.
- the peak detection on the signal waveform graph can be realized by directly calling the find_peaks method in the scipy.signal library.
- the position interval between the identified peak signals in the same cycle is controlled by the parameter distance, and the minimum value that the peak signal needs to meet
- the threshold peak min is calculated using the following formula:
- the detection calculation of the sharpness is performed for each of the detected peaks.
- the peak_prominences method in the scipy.signal library can be called to calculate and detect the prominence of each peak in the signal waveform graph.
- the image classification method according to an embodiment of the present application further includes: an operation of removing a peak with a small sharpness to realize post-processing of the signal waveform, so as to avoid the sharpness Smaller peaks interfere with subsequent waveform recognition. That is, the peaks in the signal waveform curve whose convexity is smaller than the preset convexity minimum threshold are removed, so as to obtain the final signal waveform.
- the value range of the preset minimum threshold of suddenness may be: min(y smooth )+0.35 ⁇ (max(y smooth )-min(y smooth )).
- the calculation of the percentage of the column amplitude to the row amplitude in the signal waveform curve is performed.
- the percentage of col amplitude to row amplitude percent_col_in_row is used as one of the criteria for judging the fundus image and the outer eye image, and the detected obvious peak signal is displayed on the signal waveform.
- the corresponding signal waveform diagrams when the image to be classified is a fundus diagram and the corresponding signal waveform diagram when the image to be classified is an external eye diagram are respectively shown.
- the detected significant peak signals are marked and displayed in the waveform diagrams.
- the percentage of change rate of pixel values in the signal waveform curve obtained from the signal waveform graph, the sharpness of each peak in the signal waveform graph, and the columns in the signal waveform graph can be obtained. At least one of the amplitudes as a percentage of line amplitudes determines the class of the image to be classified.
- the identification and determination are performed according to at least one of the preceding three items of information.
- the signal waveform curve in the signal waveform diagram may be first determined whether the signal waveform curve in the signal waveform diagram is monotonically increasing or monotonically decreasing.
- the monotonicity of the signal waveform can be judged by calculating whether the first derivative of the signal waveform is always greater than or equal to 0 (ie, ⁇ 0), or is always less than or equal to 0 (ie, ⁇ 0). .
- the signal waveform curve When it is calculated that the signal waveform curve is always greater than or equal to 0 or is always less than or equal to 0, it indicates that the signal waveform curve is monotonically increasing or monotonically decreasing, so it can be directly determined that the image to be classified corresponding to the signal waveform graph is a fundus image.
- the to-be-classified image corresponding to the signal waveform image is the fundus image.
- the values of the first preset value and the second preset value can be flexibly set according to actual conditions. That is, the values of the first preset value and the second preset value can be set according to factors such as the image category to be recognized currently, specific application scenarios, and application requirements. In a possible implementation manner, when the currently identified image category is a fundus image or an external eye image, the value of the first preset value may be 6%, and the value of the second preset value may be 0.02.
- the image to be classified corresponding to the signal waveform diagram is an outer eye diagram.
- the values of the third preset value and the fourth preset value can also be set according to factors such as the image category currently to be recognized, specific application scenarios, and application requirements.
- the value of the third preset value may be 40%
- the value of the fourth preset value may be 6% .
- the value of the fifth preset value may be 30%.
- the value of the fifth preset value may also be selected by testing according to factors such as the image category to be recognized in the actual situation, specific application scenarios, and application requirements, which are not specifically limited here.
- the method further includes: step S500 , preprocessing the image to be classified, locating the marker area in the image to be classified, and determining the image to be classified according to the shape of the located marker area.
- the marker area is a marker position used to characterize the attribute of the image to be classified.
- the attribute of the image to be classified refers to the category to which the image belongs.
- the marker area refers to the through-hole area.
- the marker area is a representative location that can characterize the category to which the image belongs. No further examples are provided here.
- the marker area in the image to be classified is located, and the category of the image to be classified is determined according to the shape of the located marker area, it can be specifically implemented in the following manner.
- the preprocessing of the to-be-classified image includes: cropping the to-be-classified image, and cropping the to-be-classified image into a standard image.
- the cropping method can directly adopt the above-mentioned intercepting method, whereby the standard image can be directly read by taking the center of the image to be classified as the reference center point and intercepting an image of a preset size including a specific area.
- the standard image is preprocessed to obtain a black and white binary image.
- the preprocessing of the standard image may include:
- Gaussian filtering is performed on the standard image to remove part of the noise; among them, cv2.GaussianBlur can be used for Gaussian filtering, and the Gaussian kernel size is selected (5, 5).
- grayscale conversion can be performed using cv2.cvtColorc.
- the following steps are performed: detecting each connected region in the binary image.
- the preprocessed binarized image is subjected to closing operation and opening operation in turn, and isolated noise points are filtered, and then the connected area formed by dense pixel points in the binarized image is identified.
- cv2.morphologyEx can be used to achieve accurate identification of connected regions formed by dense pixels in a binary image.
- the connected area with the largest area is selected from the detected connected areas, so as to determine the optimal adaptive clipping size of the image, and then draw a corresponding size on the binarized image according to the determined optimal adaptive clipping size. Rectangle.
- This process can use cv2.contourArea to calculate the area of the connected area.
- the area of the selected connected area should be greater than 20000, and the tolerance area between the connected area and the edge contour of the binarized image should be greater than 2000 pixels.
- the operation of removing noise interference and locating the pupil position is performed.
- the cv2.getStructuringElement method can be used to remove the noise interference around the binary image, and the positioning of the via position can be completed by cv2.morphologyEx.
- the position of the center of the pupil circle may be set within an interval range of [200px, 660px].
- the marker area of the image to be classified that is currently being identified and divided can be located.
- the category of the image to be classified can be determined according to the shape of the located marker area.
- ellipse detection can be performed through cv2.fitEllipse to complete the classification of fundus and outer eye images.
- the image is determined to be an outer eye diagram.
- the predetermined interval range of the short-axis radius of the pupil may be: [82px, 700px].
- the image data in the process of identifying and determining the to-be-classified image in the image data set according to the signal waveform diagram, can be determined in combination with the shape of the marker area in the image. It focuses on the distinction of 93.7% of the images, which greatly improves the accuracy of image classification. For different eye positions, different eye shapes, different lesions, different shooting angles, and different exposure and saturation All images can be accurately distinguished and recognized, which effectively improves the flexibility and robustness of the image classification method.
- the present application further provides an image classification apparatus. Since the working principle of the image classification device provided in the present application is the same as or similar to the principle of the image classification method of the present application, the repeated places will not be repeated.
- the image classification apparatus 100 includes: a data input module 110 , an image interception module 120 , a waveform diagram drawing module 130 and a first category determination module 140 .
- the data input module 110 is configured to input an image data set to be classified; wherein, the image data set includes images to be classified.
- the image intercepting module 120 is configured to take the center of the image to be classified as a reference center point, and intercept an image of a preset size including a specific area as a standard image.
- the waveform graph drawing module 130 is configured to select any single-channel pixel value array in the standard image, and draw a corresponding signal waveform graph based on the pixel value array.
- the first category determination module 140 is configured to determine the category of the image to be classified based on the signal waveform diagram.
- a second category determination module (not shown in the figure) is also included.
- the second category determination module is configured to preprocess the image to be classified, locate the marker area in the image to be classified, and determine the category of the image to be classified according to the shape of the located marker area.
- an image classification device 200 is also provided.
- the image classification apparatus 200 includes a processor 210 and a memory 220 for storing instructions executable by the processor 210 .
- the processor 210 is configured to implement any of the aforementioned image classification methods when executing the executable instructions.
- the number of processors 210 may be one or more.
- the image classification apparatus 200 in this embodiment of the present application may further include an input device 230 and an output device 240 .
- the processor 210, the memory 220, the input device 230, and the output device 240 may be connected through a bus, or may be connected in other ways, which are not specifically limited here.
- the memory 220 can be used to store software programs, computer-executable programs, and various modules, such as programs or modules corresponding to the image classification method in the embodiments of the present application.
- the processor 210 executes various functional applications and data processing of the image classification apparatus 200 by running software programs or modules stored in the memory 220 .
- the input device 230 may be used to receive input numbers or signals. Wherein, the signal may be the generation of a key signal related to user setting and function control of the device/terminal/server.
- the output device 240 may include a display device such as a display screen.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Molecular Biology (AREA)
- Mathematical Physics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Ophthalmology & Optometry (AREA)
- Human Computer Interaction (AREA)
- Image Analysis (AREA)
- Eye Examination Apparatus (AREA)
Abstract
Description
Claims (10)
- 一种图像分类方法,其特征在于,包括:输入待分类的影像数据集;其中,所述影像数据集中包括有待分类图像;以所述待分类图像的中心为基准中心点,截取包括特定区域的预设尺寸的图像作为标准图像;选取所述标准图像中任一单通道的像素值数组,并基于所述像素值数组绘制相应的信号波形图;基于所述信号波形图确定所述待分类图像的类别。
- 根据权利要求1所述的方法,其特征在于,输入待分类的影像数据集之后,还包括:根据图像尺寸和/或图像颜色对所述影像数据集中的待分类图像进行过滤的操作。
- 根据权利要求1所述的方法,其特征在于,选取所述标准图像中任一单通道的像素值数组,并基于所述像素值数组绘制相应的信号波形图,包括:选取所述标准图像中的任一单通道的像素值数组;对所述像素值数组中的行像素数组和列像素数组分别进行行压缩和列压缩,得到行像素压缩数组和列像素压缩数组;对所述行像素压缩数组和所述列像素压缩数组进行曲线平滑处理;根据平滑处理后的所述行像素压缩数组和所述列像素压缩数组绘制相应的所述信号波形图。
- 根据权利要求3所述的方法,其特征在于,对所述像素值数组中的行像素数组和列像素数组分别进行行压缩和列压缩时,采用如下公式进行:y_row=img_r.sum(axis=0)#像素值行压缩;y_col=img_r.sum(axis=1)#像素值列压缩;其中,axis表征像素值求和的维度,axis=0表示各行的像素值求和,axis=1表示各列的像素值求和。
- 根据权利要求3所述的方法,其特征在于,对所述行像素压缩数组和所述列像素压缩数组进行曲线平滑处理后,还包括:对平滑处理后的 行像素压缩数组和列像素压缩数组进行取对数运算处理的操作。
- 根据权利要求1至5任一项所述的方法,其特征在于,基于所述信号波形图确定所述待分类图像的类别,包括:根据所述信号波形图中的信号波形曲线中像素值的变化率百分比、所述信号波形图中的各峰值的突度和所述信号波形曲线中列振幅占行振幅的百分比中的至少一种确定所述待分类图像的类别。
- 根据权利要求1或3所述的方法,其特征在于,基于所述像素值数组绘制相应的信号波形图时,还包括:根据峰属性检测所述信号波形图中的各峰值,计算出所述信号波形图中的各峰值的突度,根据各峰值的突度对所述信号波形图进行处理;其中,根据各峰值的突度对所述信号波形图进行处理包括:移除突度小于预设突度最小阈值的波峰的操作。
- 根据权利要求1至5任一项所述的方法,其特征在于,还包括:对所述待分类图像进行预处理,定位出所述待分类图像中的标志区域,并根据定位出的所述标志区域的形状确定所述待分类图像的类别。
- 一种图像分类装置,其特征在于,包括:数据输入模块、图像截取模块、波形图绘制模块和第一类别确定模块;所述数据输入模块,被配置为输入待分类的影像数据集;其中,所述影像数据集中包括有待分类图像;所述图像截取模块,被配置为以所述待分类图像的中心为基准中心点,截取包括特定区域的预设尺寸的图像作为标准图像;所述波形图绘制模块,被配置为选取所述标准图像中任一单通道的像素值数组,并基于所述像素值数组绘制相应的信号波形图;所述第一类别确定模块,被配置为基于所述信号波形图确定所述待分类图像的类别。
- 一种图像分类设备,其特征在于,包括:处理器;用于存储处理器可执行指令的存储器;其中,所述处理器被配置为执行所述可执行指令时实现权利要求1至8中任意一项所述的方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/642,568 US12293576B2 (en) | 2021-03-26 | 2021-08-17 | Determining type of to-be-classified image based on signal waveform graph |
JP2022527909A JP7305046B2 (ja) | 2021-03-26 | 2021-08-17 | 画像分類方法及び装置並びに機器 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110322703.6A CN112801049B (zh) | 2021-03-26 | 2021-03-26 | 图像分类方法和装置及设备 |
CN202110322703.6 | 2021-03-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022198898A1 true WO2022198898A1 (zh) | 2022-09-29 |
Family
ID=75815763
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/112932 WO2022198898A1 (zh) | 2021-03-26 | 2021-08-17 | 图像分类方法和装置及设备 |
Country Status (4)
Country | Link |
---|---|
US (1) | US12293576B2 (zh) |
JP (1) | JP7305046B2 (zh) |
CN (1) | CN112801049B (zh) |
WO (1) | WO2022198898A1 (zh) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116071725A (zh) * | 2023-03-06 | 2023-05-05 | 四川蜀道新能源科技发展有限公司 | 一种路面标线识别方法及系统 |
CN116320459A (zh) * | 2023-01-08 | 2023-06-23 | 南阳理工学院 | 一种基于人工智能的计算机网络通信数据处理方法及系统 |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112801049B (zh) * | 2021-03-26 | 2021-07-23 | 北京至真互联网技术有限公司 | 图像分类方法和装置及设备 |
US12100152B1 (en) * | 2023-01-30 | 2024-09-24 | BelleTorus Corporation | Compute system with acne diagnostic mechanism and method of operation thereof |
CN116979991B (zh) * | 2023-07-24 | 2025-07-22 | 电子科技大学 | 一种多跳速情况下跳频信号预分选方法 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080085048A1 (en) * | 2006-10-05 | 2008-04-10 | Department Of The Navy | Robotic gesture recognition system |
CN106446936A (zh) * | 2016-09-06 | 2017-02-22 | 哈尔滨工业大学 | 基于卷积神经网络的空谱联合数据转波形图的高光谱数据分类方法 |
CN108596169A (zh) * | 2018-03-12 | 2018-09-28 | 北京建筑大学 | 基于视频流图像的分块信号转换与目标检测方法及装置 |
CN108921076A (zh) * | 2018-09-21 | 2018-11-30 | 南京信息工程大学 | 基于图像的道面裂缝病害自适应恒虚警检测方法 |
CN109359693A (zh) * | 2018-10-24 | 2019-02-19 | 国网上海市电力公司 | 一种电能质量扰动分类方法 |
CN112801049A (zh) * | 2021-03-26 | 2021-05-14 | 北京至真互联网技术有限公司 | 图像分类方法和装置及设备 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5117353B2 (ja) * | 2008-11-07 | 2013-01-16 | オリンパス株式会社 | 画像処理装置、画像処理プログラムおよび画像処理方法 |
CN111814564B (zh) * | 2020-06-09 | 2024-07-12 | 广州视源电子科技股份有限公司 | 基于多光谱图像的活体检测方法、装置、设备和存储介质 |
-
2021
- 2021-03-26 CN CN202110322703.6A patent/CN112801049B/zh active Active
- 2021-08-17 JP JP2022527909A patent/JP7305046B2/ja active Active
- 2021-08-17 WO PCT/CN2021/112932 patent/WO2022198898A1/zh active IP Right Grant
- 2021-08-17 US US17/642,568 patent/US12293576B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080085048A1 (en) * | 2006-10-05 | 2008-04-10 | Department Of The Navy | Robotic gesture recognition system |
CN106446936A (zh) * | 2016-09-06 | 2017-02-22 | 哈尔滨工业大学 | 基于卷积神经网络的空谱联合数据转波形图的高光谱数据分类方法 |
CN108596169A (zh) * | 2018-03-12 | 2018-09-28 | 北京建筑大学 | 基于视频流图像的分块信号转换与目标检测方法及装置 |
CN108921076A (zh) * | 2018-09-21 | 2018-11-30 | 南京信息工程大学 | 基于图像的道面裂缝病害自适应恒虚警检测方法 |
CN109359693A (zh) * | 2018-10-24 | 2019-02-19 | 国网上海市电力公司 | 一种电能质量扰动分类方法 |
CN112801049A (zh) * | 2021-03-26 | 2021-05-14 | 北京至真互联网技术有限公司 | 图像分类方法和装置及设备 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116320459A (zh) * | 2023-01-08 | 2023-06-23 | 南阳理工学院 | 一种基于人工智能的计算机网络通信数据处理方法及系统 |
CN116320459B (zh) * | 2023-01-08 | 2024-01-23 | 南阳理工学院 | 一种基于人工智能的计算机网络通信数据处理方法及系统 |
CN116071725A (zh) * | 2023-03-06 | 2023-05-05 | 四川蜀道新能源科技发展有限公司 | 一种路面标线识别方法及系统 |
CN116071725B (zh) * | 2023-03-06 | 2023-08-08 | 四川蜀道新能源科技发展有限公司 | 一种路面标线识别方法及系统 |
Also Published As
Publication number | Publication date |
---|---|
US12293576B2 (en) | 2025-05-06 |
JP2023522511A (ja) | 2023-05-31 |
CN112801049B (zh) | 2021-07-23 |
US20240046632A1 (en) | 2024-02-08 |
JP7305046B2 (ja) | 2023-07-07 |
CN112801049A (zh) | 2021-05-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022198898A1 (zh) | 图像分类方法和装置及设备 | |
US11681418B2 (en) | Multi-sample whole slide image processing in digital pathology via multi-resolution registration and machine learning | |
CN106023151B (zh) | 一种开放环境下中医舌象目标检测方法 | |
CN107038704B (zh) | 视网膜图像渗出区域分割方法、装置和计算设备 | |
CN102096802A (zh) | 人脸检测方法及装置 | |
KR102826732B1 (ko) | 안저 영상 식별 방법 및 장치와 기기 | |
CN106355599A (zh) | 基于非荧光眼底图像的视网膜血管自动分割方法 | |
CN106355584A (zh) | 基于局部熵确定阈值的眼底图像微动脉瘤自动检测方法 | |
CN118172608B (zh) | 一种幽门杆菌的细菌培养图像分析方法及系统 | |
CN118097141A (zh) | 一种基于产科影像的图像分割方法及系统 | |
CN117576121A (zh) | 一种显微镜扫描区域自动分割方法、系统、设备及介质 | |
CN113989588A (zh) | 一种基于自学习的五边形绘图测试智能评价系统及方法 | |
AU2020103713A4 (en) | Digital imaging methods and system for processing agar plate images for automated diagnostics | |
CN118898775A (zh) | 一种水下图像结构病害全自动标注方法及系统 | |
CN113052234A (zh) | 一种基于图像特征和深度学习技术的玉石分类方法 | |
CN118135620A (zh) | 基于病理切片图像的肝癌微血管侵犯区域的识别方法及系统 | |
JP2008084109A (ja) | 目開閉判定装置及び目開閉判定方法 | |
CN108921171B (zh) | 一种骨关节x线片自动识别分级方法 | |
CN113706515B (zh) | 舌像异常确定方法、装置、计算机设备和存储介质 | |
CN107358224B (zh) | 一种白内障手术中虹膜外轮廓检测的方法 | |
Cloppet et al. | Adaptive fuzzy model for blur estimation on document images | |
Jalil et al. | Iris localization using colour segmentation and circular Hough transform | |
CN110543802A (zh) | 眼底图像中左右眼识别方法与装置 | |
CN113139552B (zh) | 一种小麦抽穗期识别方法、装置、电子设备及存储介质 | |
CN114445364B (zh) | 眼底图像微动脉瘤区域检测方法及其成像方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 17642568 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2022527909 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21932510 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21932510 Country of ref document: EP Kind code of ref document: A1 |
|
WWG | Wipo information: grant in national office |
Ref document number: 17642568 Country of ref document: US |