[go: up one dir, main page]

CN119295806A - Chironomid larvae identification system based on deep learning and image recognition - Google Patents

Chironomid larvae identification system based on deep learning and image recognition Download PDF

Info

Publication number
CN119295806A
CN119295806A CN202411318584.7A CN202411318584A CN119295806A CN 119295806 A CN119295806 A CN 119295806A CN 202411318584 A CN202411318584 A CN 202411318584A CN 119295806 A CN119295806 A CN 119295806A
Authority
CN
China
Prior art keywords
image
sample
classification
deep learning
larvae
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202411318584.7A
Other languages
Chinese (zh)
Inventor
王俊如
雷琦
彭玉
郭旭升
杨铁柱
李斌
方标
刘邦淼
党昊
刘泽远
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ecological Environment Monitoring And Scientific Research Center Of Yangtze River Basin Ecological Environment Supervision And Administration Bureau Ministry Of Ecological Environment
Xinyang Agriculture and Forestry University
Original Assignee
Ecological Environment Monitoring And Scientific Research Center Of Yangtze River Basin Ecological Environment Supervision And Administration Bureau Ministry Of Ecological Environment
Xinyang Agriculture and Forestry University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ecological Environment Monitoring And Scientific Research Center Of Yangtze River Basin Ecological Environment Supervision And Administration Bureau Ministry Of Ecological Environment, Xinyang Agriculture and Forestry University filed Critical Ecological Environment Monitoring And Scientific Research Center Of Yangtze River Basin Ecological Environment Supervision And Administration Bureau Ministry Of Ecological Environment
Priority to CN202411318584.7A priority Critical patent/CN119295806A/en
Publication of CN119295806A publication Critical patent/CN119295806A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/60Editing figures and text; Combining figures or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • G06V20/698Matching; Classification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Multimedia (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Microscoopes, Condenser (AREA)

Abstract

本申请涉及生物识别和环境监测领域,公开了基于深度学习和图像识别的摇蚊幼虫识别系统,包括图像采集模块、自动化样品处理装置、图像叠加处理模块、深度学习识别模块和种类自动分类与统计模块,图像采集模块通过光学显微镜和数字相机获取多焦平面的摇蚊幼虫图像,自动化样品处理装置实现样品的自动加载操作,图像叠加处理模块对多焦平面图像进行自适应融合,生成包含多个深度信息的全焦面图像,深度学习识别模块提取摇蚊幼虫的微细结构特征,自动识别并分类不同种类的摇蚊幼虫,种类自动分类与统计模块根据分类结果进行统计分析,并生成种类分布和时间序列变化的报告,本发明提高了摇蚊幼虫识别的精度和效率,适用于水生态环境的自动化监测。

The present application relates to the field of biometrics and environmental monitoring, and discloses a chironomid larvae identification system based on deep learning and image recognition, including an image acquisition module, an automated sample processing device, an image overlay processing module, a deep learning identification module, and an automatic species classification and statistics module. The image acquisition module acquires chironomid larvae images of multiple focal planes through an optical microscope and a digital camera, the automated sample processing device realizes automatic sample loading operation, the image overlay processing module adaptively fuses the multi-focal plane images, and generates a full-focal surface image containing multiple depth information, the deep learning identification module extracts the microstructural features of the chironomid larvae, and automatically identifies and classifies different species of chironomid larvae, the automatic species classification and statistics module performs statistical analysis based on the classification results, and generates a report on species distribution and time series changes. The present invention improves the accuracy and efficiency of chironomid larvae identification, and is suitable for automated monitoring of aquatic ecological environments.

Description

Chironomus larva recognition system based on deep learning and image recognition
Technical Field
The invention relates to the technical field of biological recognition and environmental monitoring, in particular to a midge larva recognition system based on deep learning and image recognition.
Background
In the fields of biological monitoring and environmental monitoring, the identification of the species of chironomus larvae is one of the important indicators for assessing the health condition of the water ecosystem. Currently, traditional chironomus larva identification relies primarily on manual manipulation, including manual handling of samples, microscopic observation, and classification based on morphological features. The method is time-consuming and labor-consuming, has high requirements on professional knowledge and experience of operators, is prone to subjective errors, and is difficult to ensure identification accuracy and efficiency. In addition, since the midge larvae are tiny, the tiny structures (such as chin plate, tail bristles and the like) of the midge larvae are difficult to comprehensively present in a single focal plane, and all key features are difficult to capture by the traditional method, so that the identification is inaccurate.
The existing automatic image recognition technology is applied in certain fields, but in the aspect of the midge larva recognition, the existing technology cannot effectively combine a multi-focal-plane image processing and deep learning algorithm because of complex sample structure and unstable image quality, so that the problems of sample processing automation, image definition, recognition accuracy and the like are difficult to solve at the same time. Therefore, the prior art still has a plurality of defects in the aspects of processing efficiency, identification precision, automation degree and the like, and is difficult to meet the requirements of rapid and accurate identification in water ecological monitoring.
Aiming at the problems, the invention provides a midge larva recognition system based on deep learning and image recognition technology, which aims to overcome the limitations of the traditional method, realize automation and intellectualization of sample processing, image acquisition and recognition processes and improve recognition precision and efficiency.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a midge larva recognition system based on deep learning and image recognition, solves the problems of low sample processing efficiency, poor recognition precision and difficult automation in the traditional midge larva recognition method, and realizes the intellectualization and high efficiency of sample processing, image acquisition and classification recognition.
In order to achieve the aim, the invention is realized by the following technical scheme that the midge larva recognition system based on deep learning and image recognition comprises:
The image acquisition module comprises an optical microscope with a transmission light source, a microscope objective lens and a digital camera, wherein the digital camera is arranged on the microscope and is connected with a computer through a data transmission interface;
The automated sample processing device comprises a sample loading system and a triaxial electric object stage, wherein the sample loading system conveys a midge larva sample to an imaging position, and the triaxial electric object stage adjusts the XYZ axis position of the sample for automatic image acquisition under multiple fields of view and multiple focal planes;
The image superposition processing module adopts self-adaptive multi-scale image fusion to carry out high-precision image superposition on the acquired images of different focal planes and generate a full focal plane image containing a plurality of depth information from bottom to top;
The deep learning recognition module is used for automatically recognizing the microstructure features of the midge larvae, including but not limited to chin plate, inner cheilose, eyespot and tail seta features, by training and extracting features of the superimposed full focal plane images based on the convolutional neural network model so as to realize classification of midge larva types of different subfamilies and genera;
The automatic classification and statistics module of the species automatically identifies and classifies the species of the midge larva by combining the output of the deep learning model, provides statistical analysis, and stores the result in a database to support the automatic monitoring requirements of biological monitoring and aquatic ecological environment.
The invention also provides a midge larva recognition method based on deep learning and image recognition, which comprises the following steps:
an image acquisition step, namely acquiring multi-focal plane images of the midge larvae through an image acquisition module, wherein the image acquisition module comprises a transmission light source, an optical microscope and a high-resolution digital camera, and the digital camera is controlled to automatically acquire image sequences under different focal planes;
An image superimposition processing step of performing image superimposition processing on the acquired multi-focal-plane image, the image superimposition processing including:
Decomposing each focal plane image into a plurality of scale layers through multi-scale Laplacian pyramid decomposition;
Calculating fusion weight according to the local gradient, contrast and motion compensation result of the image, and fusing a plurality of focal plane images to generate a full focal plane image;
the feature extraction step, namely inputting the superimposed full focal plane image into a deep learning recognition module, wherein the deep learning recognition module extracts the features of the midge larvae based on a convolutional neural network, and the feature extraction process comprises convolutional, activating and pooling operations;
classifying, namely classifying the extracted features through a full-connection layer of the deep learning recognition module and a Softmax classifier, and outputting a class label of the midge larva and a corresponding classification confidence;
A classification result storage step of storing the classification result together with the image number, the time stamp and the detection position information to a database;
And a statistical analysis step, wherein the classification result is statistically analyzed through a type automatic classification and statistics module, the number of the midge larvae of different types and the time sequence change thereof are calculated, a type distribution and time trend chart is generated, and a report is automatically generated.
The invention provides a midge larva recognition system based on deep learning and image recognition. The device comprises the following
The beneficial effects are that:
1. According to the invention, through the multi-focal plane image acquisition and image superposition processing technology, the full-focal plane image is generated, the problem that a single focal plane image cannot capture all key structures of the midge larvae is effectively solved, and the identification precision of the midge larvae microstructure is remarkably improved.
2. According to the invention, an automatic sample processing device is adopted to realize automatic loading, positioning, tabletting and image acquisition of the sample, so that errors caused by manual operation are reduced, the stability of the sample and the consistency of image acquisition are ensured, and the stability of the identification system is improved.
3. According to the invention, through the deep learning recognition module, the characteristic is automatically extracted from the full focal plane image by utilizing the convolutional neural network, and the midge larvae are efficiently classified. Compared with the traditional manual classification method, the deep learning can perform category classification more quickly and accurately, and the processing efficiency is greatly improved.
4. By carrying out confidence screening on the classification results, the invention ensures that only the classification results with high confidence are used for automatic classification and statistics, thereby improving the reliability of the classification results of the system. Meanwhile, for the result with low confidence, a manual examination mechanism is provided, and the classification accuracy is further improved.
5. The automatic classification and statistics module of the invention can carry out statistics of the number of species and time sequence analysis in real time, and display the dynamic change condition of the midge larva population. The method provides timely and accurate data support for long-term monitoring and research of the water ecological environment, and is convenient for environmental change analysis and trend prediction.
6. The system can automatically generate the report containing the classification result, the statistical analysis and the time sequence change trend, supports the export of various formats, is convenient for the user to archive and analyze, reduces the workload of manual data arrangement and report writing, and improves the automation degree of data processing.
Drawings
FIG. 1 is a schematic diagram of a system architecture according to the present invention;
FIG. 2 is a schematic diagram of an image acquisition module and an automated sample processing apparatus according to the present invention;
FIG. 3 is a schematic view of a slide glass trough structure of the present invention;
FIG. 4 is a schematic diagram of a computer controlled connection according to the present invention;
FIG. 5 is a schematic diagram of a picture stacking processing module according to the present invention;
FIG. 6 is a schematic diagram of a deep learning recognition module according to the present invention;
FIG. 7 is a schematic diagram of an automatic classification and statistics module according to the present invention;
FIG. 8 is a schematic flow chart of the method of the present invention.
The device comprises a 10-optical microscope, 11-transmission light source, 12-objective lens, 13-objective lens, 14-digital camera, 15-triaxial electric objective lens, 20-computer, 30-automatic sample processing device, 31-sample liquid bottle, 32-sample tube, 33-peristaltic pump, 34-sample tube, 35-sample containing groove, 36-sample discharging tube, 37-waste liquid barrel, 40-slide groove, 50-image superposition processing module, 51-multi-focal plane image acquisition unit, 52-multi-scale image decomposition unit, 53-adaptive weighted fusion unit, 54-image fusion unit, 55-image reconstruction unit, 60-deep learning identification module, 61-input preprocessing unit, 62-convolution characteristic extraction unit, 63-characteristic classification unit, 64-training and optimizing unit, 65-database interface unit, 70-type automatic classification and statistics module, 71-classification result storage unit, 72-confidence screening unit, 73-type statistics unit, 74-time sequence analysis unit, 75-report generation unit.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and fully with reference to the accompanying drawings, in which it is evident that the embodiments described are only some, but not all embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Referring to fig. 1-7, the invention provides a midge larva recognition system based on deep learning and image recognition, which can realize automatic recognition, classification and statistical analysis of midge larva. The system realizes efficient and accurate detection and identification of the midge larvae through combination of automatic sample processing, image acquisition, image superposition processing, deep learning identification and automatic classification and statistical analysis modules of the species, and provides technical support for biological monitoring and aquatic ecological environment monitoring. The following detailed description of the invention refers to the accompanying drawings:
The image acquisition module is one of the core components of the whole system and is used for acquiring multi-focal plane images of the midge larvae. The module is combined with a high-resolution digital camera through precise optical imaging equipment, so that high-precision imaging of the microstructure of the midge larva is ensured, and high-quality image data is provided for subsequent image processing and deep learning identification.
The optical microscope 10 is a core imaging device in an image acquisition module for high magnification observation of a chironomid larva sample. The optical microscope mainly comprises a transmission light source 11, an objective 12 and an objective table 13, and ensures the visibility and high-precision imaging of a sample. The concrete explanation is as follows:
the transmissive light source 11 is used to provide uniform and stable illumination, ensuring that transmitted light is formed beneath the sample, making each microstructure of the sample clearly visible. The brightness of the transmission light source 11 can be adjusted according to the requirements of different samples, so that the influence of the too strong or the too weak light on the image quality is avoided.
Objective 12 the optical microscope 10 is equipped with an objective 12 of various magnifications, e.g. 10-fold, 40-fold, 100-fold, the user being able to select the appropriate objective magnification depending on the sample size. Objective 12 has high optical transmittance and is capable of magnifying microscopic structures of the midge larvae, such as the chin plate, tail bristles and inner labial bar, ensuring a clear characteristic.
A digital camera 14 is mounted above the microscope's objective lens for capturing an image of the sample imaged by the optical microscope. The camera has high resolution, typically 2000 ten thousand pixels or more, and is capable of recording detailed images of the midge larvae in a high definition manner. The specific functions are as follows:
resolution-high resolution cameras can ensure that each fine feature of the midge larvae is captured, especially under high magnification, ensuring that the images are clear and undistorted.
The camera is provided with an automatic focusing function, so that the focal length can be automatically adjusted when shooting in multiple focal planes, the sample is ensured to be in an optimal imaging state on each focal plane, and blurred images are avoided.
And the camera supports a high frame rate continuous shooting mode, so that continuous image acquisition of different focal planes is completed in a short time, and the working efficiency is improved. More than 5 frames of images are shot per second, so that the integrity of the sample is ensured.
The digital camera 14 is connected with a computer through a high-speed data transmission interface (such as a USB 3.0 or higher standard interface) to transmit the collected image data in real time, so that the high-efficiency processing of the data is ensured. The data transmission speed can reach more than 5Gbps, and the stability of image transmission is ensured.
The computer 20 serves as a core control unit of the image acquisition module and is responsible for managing the entire image acquisition process. The computer is built with special control software which can interact with the digital camera 14 and the optical microscope 10, automatically control various parameters in the image acquisition process, and store and manage the acquired images. The functions of the method include:
And in the automatic control, through presetting acquisition parameters, the computer 20 can automatically control the operations of exposure, focusing, shooting speed and the like of the camera, so that manual intervention is reduced, and the precision and efficiency are improved.
Real-time image display, in which the computer can display the image in real time in the acquisition process, so that an operator can monitor the imaging quality and carry out necessary parameter adjustment.
And (3) storing the images, namely automatically storing the acquired images to local or network storage equipment, and naming file names by time stamps or sample numbers so as to facilitate subsequent retrieval and processing. The system supports multiple image format stores including RAW, JPEG, and TIFF formats.
The multi-focal plane automatic control unit is connected with the computer 20 and the three-axis electric object stage 15 and is responsible for automatically acquiring images under the multi-focal plane. By adjusting the Z axis of the microscope, images of the sample at different focal lengths can be continuously acquired. The multi-focal plane automatic control unit functions as follows:
The Z axis is automatically adjusted, and the unit can accurately adjust the position of the sample on the Z axis, so that the focuses of images collected under different focal lengths are different, and information of different depths of the sample is obtained. The adjustment range can be precisely controlled at the micron level, and each step is moved by 1-2 microns, so that multi-level detail images are ensured to be acquired.
And setting a focal length range, wherein a user can preset the focal length range and the shooting interval through a software interface, and the system automatically adjusts the focal length according to the settings and completes image acquisition. Typically, a depth range of 0 to hundreds of microns may be covered to ensure complete recording of the various layers of information of the sample.
The XYZ three-axis motorized stage 15 is used to fix the sample and make fine adjustments in XYZ directions. In image acquisition, the stage mainly performs the following functions:
X and Y movement the stage can be moved in the horizontal direction (X and Y axes) to ensure that the entire sample area is scanned. For large-size samples, scanning imaging of the full area can be accomplished by multi-field image stitching techniques.
And the Z direction moves, namely, the object stage is matched with the multi-focal-plane automatic control unit, and the height of the sample is accurately adjusted along the Z axis direction, so that the camera can acquire images under different focal planes. The Z-axis adjusting precision reaches the micron level, so that each layer of focal plane can clearly present the structural information of the sample.
The software control interface is used as a user interaction platform of the computer system and is responsible for visual control of the whole image acquisition process. An operator may complete a series of acquisition settings and operations through an interface, including:
and the exposure and focusing settings are that an operator can manually or automatically adjust parameters such as exposure time, focal length, light brightness and the like of the camera to ensure the best quality of image acquisition.
And the acquisition mode selection is that the system supports a plurality of image acquisition modes, including single-view multi-focal-plane acquisition and multi-view scanning acquisition, and a user can select a proper mode according to requirements.
Image previewing and monitoring, namely previewing the acquired image in real time, and timely adjusting acquisition parameters to avoid acquisition errors caused by image blurring or improper exposure.
The whole flow of the image acquisition module is as follows:
1. the operator sets acquisition parameters including focal length range, exposure time and shooting interval through a software control interface.
2. The sample is adjusted to a proper position by an XYZ three-axis electric stage 15, and the optical microscope 10 is matched with the transmission light source 11 to provide stable illumination.
3. The digital camera 14 begins multi-focal plane image acquisition at different focal lengths and the images are transmitted in real time to the computer 20 via the data transmission interface.
4. The computer 20 manages and stores the acquired images, and an operator can monitor the acquisition process in real time through a software interface to ensure the image quality.
The image acquisition module realizes multi-level and high-precision image acquisition of the midge larvae through the organic combination of the precise optical microscope 10, the high-resolution digital camera 14, the three-axis electric object stage 15 and the computer 20. The components are mutually matched, so that complete depth information of the sample is obtained under different focal planes, and high-quality image data is provided for subsequent image processing and identification.
The automated sample processing device 30 is an important component in the system and is responsible for the transport, positioning and fixing of the chironomus larva samples to ensure that the samples are in a stable state during microscopic imaging and can be accurately collected under multiple fields of view and multiple focal planes. Since the body length of each midge larva is usually 1-2 cm, the system can perform certain number control in a single visual field so as to ensure the stable positioning and imaging quality of the sample. Through automatic adjustment and various processing operations to the sample position, errors in manual operation are avoided, and accuracy and efficiency of imaging and identification processes are greatly improved.
The sample loading system is responsible for automatically conveying the chironomus larva samples from the storage container to the imaging area, and ensuring that the number of the samples in the field of view is controlled to be 1 during each imaging, so as to avoid that a plurality of samples interfere with the imaging quality. The specific structure and functions include:
Sample feeding device the chironomus larva sample is transported from the sample bottle to the imaging position by a feeding hose 32. The system has the function of controlling sample flow, can accurately adjust the sample injection speed of a sample, and ensures that only one midge larva individual is contained in each view field. The sample feeding hose 32 is connected to a sample liquid bottle 31 and a peristaltic pump 33, and the sample is transported under controlled liquid flow to an imaging location above the stage.
Peristaltic pump 33 is used to precisely control the flow rate of a liquid sample to avoid too fast a flow of the sample affecting the positioning or imaging of the sample. By adjustable control of peristaltic pump 33, the system is able to adjust the liquid flow rate according to the type and density of the sample, ensuring that the sample is stably and continuously transferred to the imaging area. The peristaltic pump head applies proper compression to the sample introduction hose 32 to push the liquid flow without direct contact with the liquid, thereby avoiding contamination or sample damage. In addition, as the body length of the midge larva is 1-2 cm, the peristaltic pump can control the sample feeding speed of the sample according to preset parameters, so that only 1 complete individual is captured each time within the visual field limit. Peristaltic pumps control sample flow at a low rate through a hose, preventing multiple samples from entering the imaging region.
The sample processing sample holding groove 35 is a key area for flowing and collecting samples on the object stage, and the interior of the groove is connected with the sample feeding hose 34 through the peristaltic pump 33 to inject the chironomus larva sample, so that the imaging and detection of a microscope are facilitated. And injecting a midge larva sample into the groove through a peristaltic pump, and controlling the flow and the number of the samples through software. Specific features of the sample processing well include:
Liquid control after filling the sample processing well with liquid sample by peristaltic pump 33, the system will adjust the liquid flow rate and amount as needed to ensure that the sample remains properly observed in the well. Liquid control has a large impact on the stability of the sample and proper flow rates facilitate proper positioning of the sample within the microscope field of view.
The waste liquid treatment system is used for collecting waste liquid generated in the imaging process, and ensures the cleaning of experimental environment and the safe treatment of samples. The functions of the method include:
The waste liquid guiding device is connected with the discharging hose 36, and the sample waste liquid is led into the waste liquid treatment system. The discharge hose 36 is connected to the sample container 35, and introduces the excess liquid into the waste liquid tank 37, thereby preventing the sample from overflowing or contaminating the stage.
Workflow of automated sample processing device:
1. The sample loading system transfers the chironomus larva sample from the sample storage bottle to the sample processing well 35 by peristaltic pump 33, ensuring that only one individual at a time in the field of view enters the imaging region.
2. The sample is precisely moved in the XYZ direction through the triaxial electric object stage 15, and is automatically adjusted to different fields of view and focal planes, so that the image acquisition of the sample under multiple fields of view and multiple focal planes is completed.
3. And the waste liquid generated in the collection process is discharged through the sample waste liquid treatment system, so that the cleaning of the experimental environment is ensured.
Automated sample processing device 30 achieves a fully automated operation from sample loading to imaging by integrating a sample loading system, peristaltic pump 33, tri-axial motorized stage 15, and waste treatment system. And ensure that only 1 individual is contained in the field of view in each imaging process, avoiding a plurality of samples from interfering with the image quality. The device ensures the stability and accuracy of the sample in the whole imaging process, avoids errors in manual operation, and provides a reliable sample processing basis for high-precision midge larva identification and classification.
In some embodiments, the upper surface structure of the tri-axial motorized stage 15 of the automated sample processing device 30 is optimally designed to accommodate different types of sample processing requirements. To meet the special application of manual operation, especially in the case of needing to perform manual operation on the chironomus larvas, a slide groove 40 for placing slides is added in the system, and a flow design for manually sucking sampling liquid, placing slides and tabletting is provided. The following is a detailed description of this embodiment.
Structural design of slide groove
A slide groove 40 is provided on the upper surface of the three-axis motorized stage 15 for firmly holding a slide desired under a microscope. The slide groove is designed into a groove structure conforming to the size of a standard microscope slide, and has the following functions:
the slide is stably placed, the slide groove 40 can fix the position of the slide, so that slide or shift of the slide in the moving process of the objective table is avoided, and the sample is ensured to be stable in the tabletting and imaging processes.
In some operations, the user can process the sample solution in the sample well 35 by manually sucking it up and dropping it onto the surface of the slide. The operation is suitable for a sample liquid processing scene needing manual control, and comprises the following steps:
1. sample liquid sucking, namely, a user sucks a proper amount of sample liquid from the sample processing sample containing groove 35 through a manual suction pipe. The manual suction tube has reasonable volume design, can absorb a proper amount of liquid at one time, and avoids excessive sample liquid from being injected on a slide.
2. Slide placement a user places a standard sized microscope slide into slide well 40. The slide well 40 has a suitable depth and boundaries to ensure that the slide is firmly placed within the well.
3. And the sample liquid is injected, namely the user uniformly drops the sucked sample liquid on the glass slide, and the sample liquid is uniformly distributed on the surface of the glass slide, so that the fixing of a sample and the subsequent tabletting operation are facilitated.
Manual tabletting process of sample liquid
After the sample liquid is instilled on the glass slide, the system can carry out tabletting treatment on the sample liquid through a tabletting device, so that the chironomus larva sample is uniformly and stably attached to the glass slide. The manual tabletting comprises the following specific processes:
Sample spreading, namely, after the sample drops are dropped into a slide, the chironomus larva sample is distributed on the surface of the slide. The user can more uniformly distribute the sample in the central region of the slide by properly swinging or tilting the slide, thereby facilitating subsequent observation and imaging.
And (3) manual tabletting, namely, carrying out tabletting treatment on the samples by a user through a system or manually adjusting the strength of a tabletting device. The sheeting apparatus applies uniform pressure to cause the chironomus larva sample to adhere evenly to the slide, particularly for micro-structures such as chin plates and tail bristles, and the sheeting operation helps to reveal these critical parts of the sample more clearly.
Recognition process after sample liquid tablet
After manual compression is completed, the system automatically enters an image acquisition and recognition stage. At this point, the sample on the slide has been fixed and leveled, ensuring that high resolution microscopic imaging is enabled and that the microstructure of the sample is identified. The specific identification process is as follows:
Automatic positioning, namely, a slide placed in a slide groove can be adjusted along with the XYZ direction of the three-axis electric object stage 15 to enter the view field of a microscope, and the system ensures that a key area of a sample is positioned at the center of a microscope lens through the accurate movement of the XYZ axis.
Multi-focal plane image acquisition the system performs multi-focal plane image acquisition on the pressed sample by combining the digital camera 14 and the optical microscope 10. By adjusting the Z-axis direction, the system can acquire image information of the sample at different depths, and ensure that all key structures (such as chin plate, tail bristles and the like) of the sample can be imaged clearly in a focal range.
And deep learning identification, namely carrying out feature extraction and automatic classification on the acquired image through an image processing module and a deep learning identification module. The deep learning model can automatically classify the types and subfamilies of the midge larvae according to different microstructures of the samples, and store the results in a database.
Advantages and application scenarios of manual operation
The manual aspiration and compression mode of operation provides flexibility for certain specific experimental requirements. The following are the advantages and applicable scenarios of this embodiment:
The operation is flexible, through manual suction and tabletting, a user can carry out fine control on sample processing, and the method is particularly suitable for processing special or precious samples, and sample damage or error possibly generated in the full-automatic process is avoided.
The manual operation flow is suitable for the scene requiring special treatment of the sample, and if different fixing agents or coloring agents are required to be added into different samples, the manual operation enables the sample treatment to be more flexible.
In the scene of repeatedly treating a sample, adjusting the concentration of the sample liquid or testing different experimental conditions, the manual operation can be matched with an automatic device for use, so that the experimental effect is optimized.
The embodiment provides a flexible manual operation flow for the system, and particularly designs the functions of a slide groove, a manual suction tube and a tabletting operation, so that a user can manually suck sampling liquid, instill a slide and carry out the tabletting operation. The operation mode provides more controllability and flexibility for the sample processing process, and is suitable for experimental scenes needing fine control.
The image overlay processing module 50 is one of the key components of the system and is responsible for fusing and processing data from the multi-focal plane images to generate a high resolution image with a complete depth of focus. Through the processing of the module, the sample images of the midge larvae at different depths can be precisely aligned and overlapped to generate an image of a full focal plane. The quality of the image is greatly improved, and key characteristics of the midge larvae, such as chin plates, inner lip bars, tail bristles and the like, can be ensured, so that the midge larvae can be clearly displayed in a single image.
The multi-focal plane image acquisition unit 51 acquires the images of the midge larvae in different focal planes through the combination of the XYZ three-axis electric stage 15 and the digital camera 14. Each group of image sequences contains information of different depths, and can capture the micro-structures of the midge larvae under different focal lengths.
The digital camera 14 and the microscope work cooperatively, and the system can continuously shoot on different focal planes by adjusting the height of the Z axis. Each image corresponds to the layers of the sample under different depths, the image acquisition process has high frame rate, a large number of image sequences are ensured to be acquired in a short time, and image blurring caused by movement of the sample or liquid interference is avoided.
And (3) data synchronization and caching, namely caching the acquired image into a memory in real time by the system in the acquisition process, and transmitting the acquired image to a processing module through a high-speed data interface. Image data transmission and processing are performed in parallel to ensure the high efficiency and real-time performance of the system.
The multi-scale image decomposition unit 52 is configured to decompose the multi-focal plane image into image layers with different scales by using a laplacian pyramid or gaussian pyramid method. The purpose of this step is to extract different frequency information of the image and provide data support for subsequent image fusion.
Laplacian pyramid decomposition-the unit decomposes each focal plane image into multiple scale levels using the Laplacian pyramid algorithm. Firstly, an image generates a Gaussian pyramid through a Gaussian filter, and then a difference between adjacent scale images is calculated to generate a Laplacian pyramid. The formula is as follows:
Wherein, Representing a laplacian pyramid representation of the ith focal plane image at the ith scale,Is a gaussian pyramid representation of the image.
High frequency and low frequency feature separation-by image decomposition, the low frequency (smooth region) and high frequency (edge and texture detail) information of the image can be effectively separated. The high frequency portion typically represents details and sharp edges in the image, and the low frequency portion represents illumination and a wide range of hue changes. This separation provides a basis for subsequent image fusion so that images of different focal planes can be optimally fused according to the sharpness of a particular region.
The adaptive weighted fusion unit 53 dynamically adjusts the fusion weight of each focal plane image according to the local definition, contrast and motion compensation result of each focal plane image, so as to ensure that the key detail area is optimally displayed in the final full focal plane image.
Local gradient calculation the unit evaluates the sharpness of the image by calculating the local gradient of each focal plane image. The larger the gradient, the more rich the edge and texture information of the image, indicating that the region is more sharp in the focal plane image. The gradient is calculated as follows:
Wherein, Is the gradient value of the ith focal plane image at pixel point x, y.
Local contrast calculation the system evaluates the sharpness of the image by calculating the local contrast of the image in addition to the gradient. The local contrast reflects the brightness difference of the image in a certain area and can reveal whether the image has enough visual information.
Motion compensation-if slight motion (e.g., liquid disturbance) is generated in the sample or environment during multi-focal plane image acquisition, the system corrects for these movements by motion compensation techniques. By an optical flow method or a phase correlation method, the system can estimate the motion vector of the image and compensate, and the formula is as follows:
Where v (x, y) is the motion vector of the sample at time t, I (x, y, t) is the image gray value at time t, and the motion vector is used for correction before image superimposition.
And the self-adaptive weighted fusion unit calculates the weighted value of each focal plane image according to the local gradient, the contrast and the motion compensation result. The specific calculation formula is as follows:
Wherein w i (x, y) is the weight of the ith focal plane image, For the local gradient of the image, C i (x, y) is the local contrast, M i (x, y) is the motion compensated image quality weight, and α and β are adjustment coefficients.
The image fusion unit 54 is responsible for fusing the weighted images of different scales. For each decomposition scale, the system fuses the Laplacian pyramid layers of each focal plane image according to the weighted values to generate a full focal plane image containing multiple layers of information.
And in the multi-scale fusion process, on each scale, calculating a final fusion image according to the weighted value of each focal plane image. The fusion formula is as follows:
Wherein, Is a fused image at the first scale.
And (3) layer-by-layer fusion, namely, the fusion process starts from the lowest scale and gradually expands to high-frequency information, so that the clearest image content is ensured to be reserved on each frequency segment. In the low frequency scale, more is the fusion of illumination and color information, while in the high frequency scale, the system focuses more on detail information such as edges and textures.
The image reconstruction unit 55 is responsible for recombining the fused multi-scale images into a complete full-focus image. Through the operation of the inverse laplacian pyramid, the system can restore the size and resolution of the original image, generating a high-resolution image with complete depth of focus.
And in the image reconstruction process, the fused Laplacian pyramid is subjected to inverse transformation layer by layer, and a final full focal plane image is reconstructed. The reconstruction process is as follows:
Wherein, I fused (x, y) is the final full focal plane image after fusion, and L is the number of decomposed scales.
Feature enhancement the system can also feature enhance key structures (such as chin plate, inner cheilose and tail seta of the midge larva) in the image reconstruction process. By locally reinforcing the high frequency information, the system ensures that these micro structures are clearly shown in the final image.
Preferential fusion of feature regions
The system is particularly optimized for specific structural areas of the midge larvae. For key characteristic areas such as chin plates and tail bristles, the system gives higher weight to the areas in the fusion process, so that the areas are more prominent in the full focal plane image, and the recognition accuracy of a subsequent deep learning model is improved.
Workflow of the image superimposition processing module 50:
1. And acquiring multiple focal plane images by the system through a triaxial electric object stage and a digital camera under different focal lengths to form an image sequence.
2. And (3) multi-scale image decomposition, namely performing multi-scale decomposition on each image to generate a Laplacian pyramid and a Gaussian pyramid, and separating image information with different frequencies.
3. And (3) weighting fusion, namely dynamically adjusting the fusion weight of each focal plane image through the local gradient, the contrast and the motion compensation result, and carrying out multi-scale fusion.
4. And (3) reconstructing the image, namely reconstructing the multi-scale image into a complete full focal plane image through the inverse Laplacian pyramid process, so as to ensure that all key structures are clearly displayed.
The image superposition processing module 50 generates high-quality full-focal-plane images through multi-focal-plane image acquisition, decomposition, weighted fusion and reconstruction, and ensures that structural features of the midge larva sample at different depths can be clearly presented at the same time. The multi-scale fusion and feature region priority processing strategy of the module not only improves the overall quality of the image, but also provides high-quality data input for subsequent deep learning identification, and improves the identification accuracy and efficiency of the system.
The deep learning identification module 60 is an intelligent core of the system and is responsible for automatically identifying and classifying the full-focus image generated by the image superposition processing module. Through the multi-level feature extraction capability of a Convolutional Neural Network (CNN), the module can identify the microstructure of the midge larvae and realize accurate category classification according to the features of the midge larvae. In order to improve the robustness and the accuracy of the system, the deep learning identification module combines a plurality of sub-modules such as data preprocessing, feature extraction, classification prediction, model training, optimization algorithm and the like.
The input preprocessing unit 61 is responsible for data cleaning and preprocessing of the full-focus image output by the image superimposition processing module. In order to meet the input requirements of the deep learning model, the image needs to be subjected to operations such as normalization, enhancement, size adjustment and the like.
And (3) normalizing the image pixel value to the [0,1] interval, thereby being beneficial to accelerating the model training process and improving the convergence of the model.
Image resizing-to accommodate the input layer size of convolutional neural networks, the image needs to be resized, with a common input size of 224 x 224 or 512 x 512. The system uses bilinear interpolation to zoom the image, and ensures that detail information is reserved in the zooming process.
And in the model training stage, the data enhancement technology can generate more diversified samples, and the robustness of the model is enhanced. Common data enhancement operations include:
random rotation-image rotation within the range of [ -15 °, +15° ], preventing the model from excessively relying on a specific directionality.
Random cropping-a portion of the image is cropped randomly to simulate the diversity of the image data.
Random overturn, namely horizontally or vertically overturning the image, and enhancing the generalization capability of the model on samples at different positions.
The convolution feature extraction unit 62 is a core part of the deep learning identification module, and is responsible for extracting a high-level feature representation from an input full-focus image. The unit is based on a Convolutional Neural Network (CNN) architecture, and microstructure features of the midge larvae, such as chin plates, tail setae, inner cheilose and the like, are gradually extracted through a plurality of convolutional layers and pooling layers.
Convolution layer the convolution layer performs a sliding convolution operation on the image through a plurality of convolution kernels (typically 3×3 or 5×5 convolution kernels), extracts local features of the image, and generates a feature map (feature maps). The formula of the convolution operation is as follows:
Wherein I normalized (x, y) is the normalized full focal plane image, w (I, j) is the weight of the convolution kernel, F (x, y) is the convolution result, and b is the bias.
Activation function the convolution results are processed after each convolution layer using a nonlinear activation function (e.g., reLU) to enhance the model's ability to express nonlinear features. The activation formula of the ReLU is:
A(x,y)=max(0,F(x,y))
where a (x, y) is the output after activation.
And the pooling layer is used for performing down-dimension and down-sampling on the activated feature map so as to reduce the calculated amount and enhance the spatial invariance of the features. The usual pooling operation is maximum pooling, and the formula is as follows:
where P is the size of the pooling window, typically 2 x 2.
Stacking of convolutional layers deep convolutional neural networks are typically composed of a stack of multiple convolutional layers and pooled layers, with the network being able to extract more complex, higher-level features as the number of layers increases. For example, shallow convolutions may extract the edge and texture information of the midge larvae, while deep convolutions capture more semantic features such as microscopic shapes of the chin and tail bristles.
The feature classification unit 63 is responsible for further processing the high-level features processed by the convolution feature extraction unit, and outputting the classification result of the midge larvae. The feature classification unit comprises a full connection layer and a Softmax classifier, and comprises the following specific steps:
And the full connection layer is used for flattening the feature map output by the convolution feature extraction unit into a one-dimensional vector and transmitting the one-dimensional vector to the full connection layer (Fully Connected Layer) so as to map the high-dimensional feature into the classification space. The calculation formula of the full connection layer is as follows:
z=W·x+b
wherein W is the weight matrix of the full connection layer, x is the input feature vector, b is the bias, and z is the output feature representation.
The Softmax classifier is used for converting the output of the full-connection layer into probability distribution of various midge larvae, and the formula is as follows:
wherein z j is the score of category j, N is the number of chironomus larva categories, and p (y=j|x) is the probability that the input image belongs to category j.
The training and optimizing unit 64 is used for training and parameter optimization of the model, and updates the weights in the network through a back propagation algorithm, so that the model can learn the characteristics of the midge larvae better. The unit comprises the following key parts:
loss function the loss function of the model adopts cross entropy loss for measuring the difference between the model prediction result and the real label. The definition of cross entropy loss is:
Where y i is the true label of the sample, and p i is the prediction probability of the model output.
And the optimizer is an adaptive learning rate optimization algorithm, wherein in order to accelerate model convergence, the system adopts an Adam optimizer.
Back propagation-in each iteration, the system calculates the gradient by a back propagation algorithm and updates the network weights using an optimizer to reduce the value of the loss function. The back propagation ensures that the network can continuously correct errors, so that the prediction result of the model is more accurate.
The database interface unit 65 is responsible for storing the classification result and related information (such as image number, time stamp, classification confidence, etc.) to the database, so as to facilitate subsequent statistical analysis and data query. The unit has the following functions:
the classification result is stored, namely the result of each classification, including the type, confidence and related metadata (such as detection time and sample position) of the midge larvae are stored in a database.
Query and update the system supports query and update of historical classification results, and users can search classification records of some samples or update wrong classification results according to requirements.
Work flow of the deep learning identification module:
And the input preprocessing, namely normalizing and data enhancing the full focal plane image through an input preprocessing unit 61, and ensuring the image to adapt to the input requirement of the convolutional neural network.
Feature extraction by convolution feature extraction unit 62, the high-level features of the midge larvae are extracted layer by layer using a plurality of convolution layers and pooling layers.
Feature classification, namely mapping the extracted features into a classification space by using a full connection layer and a Softmax classifier in a feature classification unit 63, and outputting the types and the classification confidence of the midge larvae.
Training and optimizing the network by training and optimizing unit 64, updating model weights using back propagation and Adam optimizers, and continuously improving the classification accuracy of the model.
The result is stored, the classified result is stored in the database through the database interface unit 65, so that the subsequent inquiry and statistical analysis are convenient.
The deep learning recognition module can accurately recognize microstructure features of the midge larvae through the organic combination of data preprocessing, convolution feature extraction, classification prediction and optimization training, and realize efficient category classification. The model has strong feature extraction and classification capability, and classification accuracy and robustness are improved by continuously optimizing model weights. The database interface unit ensures the storage and inquiry of the classification result and provides data support for the subsequent statistical analysis.
The automatic classification and statistics module 70 is a core functional module in the whole system, which is responsible for receiving the output of the deep learning identification module and performing the classification and statistical analysis of the midge larvae. The module is combined with the classification result of the deep learning model to automatically classify and count different types of midge larvae, provide real-time type analysis and dynamic change monitoring, and generate a related report. Through the module, a user can comprehensively know the distribution situation and the number change trend of the midge larvae, and data support is provided for water ecological environment monitoring and biodiversity research.
The classification result storage unit 71 is responsible for storing the classification result output by the deep learning identification module in a database, and associating with metadata such as image numbers, time stamps, positions, and the like. The unit ensures the durability and retrievability of the classification results, facilitating subsequent queries and analysis.
And (3) recording classification results, wherein the classification results comprise information such as class labels, confidence levels and the like of the midge larvae, and the classification results of all samples are stored in a database together with metadata such as corresponding image IDs, detection time, sampling positions and the like to generate complete record entries. The entry is stored in a structured manner, so that the system can conveniently and rapidly inquire.
Metadata association-in addition to classification labels, the system automatically associates detected related information, such as image numbers, time stamps, detection places and other metadata, to classification results. These metadata can provide background information for sample collection, particularly in environmental monitoring tasks, which can be analyzed according to different sampling points and times.
And (3) data retrieval and updating, wherein the system supports the user to inquire and update the classification result. If the identification result needs to be manually adjusted or confirmed, the user can edit and update the classification result of the specific sample through the query system, so that the accuracy of the data is ensured.
The confidence level screening unit 72 is configured to screen the classification result according to the confidence level output by the deep learning model, and ensure the reliability of the classification result. For samples with low confidence, the system will mark them as an uncertain classification and allow the user to manually review or further process.
Confidence threshold setting the unit allows the user to set a threshold for classification confidence. The classification result with lower confidence will be labeled "uncertain" and may be processed by further manual review or model optimization. For example, when the confidence level of the classification result is below the set threshold pthresholdpthreshold, the system marks the result as uncertain and the screening process is defined as:
Where y i is a classification label, p i is a classification confidence, and p threshold is a preset confidence threshold.
And a manual review mechanism, wherein for the classification result marked as uncertain, the system provides a manual review function, and a user can review the sample image manually and manually adjust the classification result according to the requirement. This mechanism helps to improve the overall accuracy of the classification result.
The species statistics unit 73 is used for performing statistical analysis on the stored classification result to provide distribution conditions of various midge larvae in the detection sample. The unit can quickly calculate the quantity proportion of different types and generate corresponding statistical charts.
And counting the number of the larvae of the midge of different types according to the classification result, and calculating the number and the proportion of the larvae in the detection sample. The statistical formula is as follows:
Where 1 (y j=yi) is the indicator function, returning 1 when y j=yi, otherwise returning 0, m being the number of samples.
The statistical result shows that the statistical unit can generate various charts including pie charts, bar charts and the like, and the quantity distribution situation of various midge larvae is shown. These charts provide a clear visual data presentation to the user, helping the user to quickly understand the composition of the species in the current sample.
The time sequence analysis unit 74 is used for performing time sequence analysis on the classification result of the midge larvae, and displaying the quantity change trend of various midge larvae in different time periods, and is particularly suitable for long-term monitoring tasks and environmental change analysis.
And (3) processing the time sequence data, namely grouping and counting the number of the midge larvae according to time stamp information in the classification result by the system to generate time sequence data of the midge larvae. The user can set time windows such as day, week, month, etc. so as to flexibly analyze the dynamic changes of the midge larva population in different time periods.
And (3) trend analysis and visualization, wherein the system displays the time sequence analysis results in the form of a line graph and the like to reflect the change trend of the quantity of various midge larvae along with time. The time sequence chart can clearly show the growth, reduction or stability of the population, and helps a user to know the influence of environmental change on the midge larva population.
The report generating unit 75 is responsible for automatically generating detailed report files based on the classification results and the statistical analysis. The report contains the species distribution, time series analysis result and other statistical data of the midge larvae, and supports the report to be exported in various formats, so that the user can conveniently archive or further analyze the report.
Report content the automatically generated report includes the following information:
and summarizing the classification result, namely counting the number and the duty ratio of various midge larvae and providing the type composition of the current sample.
And the time change trend is that the dynamic change of various midge larvae in different time periods is displayed through a time sequence analysis result, and the long-term change trend of population quantity is reflected.
Confidence statistics, namely displaying confidence distribution of various classification results and helping users to know the reliability of the classification results.
Report format export the report generation unit supports export of the report into a variety of formats, which the user may choose to save in PDF, excel or CSV formats. PDF format reports facilitate printing and archiving, while Excel and CSV formats facilitate further data analysis and processing by users.
Workflow of the category automatic classification and statistics module 70:
The classification result of the deep learning identification module 60 is stored in the database through the classification result storage unit 71 and is associated with the related metadata, so that the integrity and traceability of the classification result are ensured.
Confidence screening and processing the system screens the classification results by a confidence screening unit 72, and for classification results with lower confidence, the system marks them as "uncertain classification" and provides manual review functionality.
The system counts the number of various midge larvae through the species counting unit 73 and generates a chart of species distribution, so that the user can visually check the chart conveniently.
Time series analysis by the time series analysis unit 74, the system displays the trend of the quantity of various midge larvae along with time, and helps the user to know the dynamic change of the sample along with time.
Report generation and export the report generation unit 75 automatically generates a complete report file based on the classification results and statistical analysis, supporting multiple format exports.
The automatic classification and statistics module provides powerful automatic analysis capability and user interaction function through automatic storage of classification results, confidence level screening, class statistics and time sequence analysis. The module provides a complete solution for automatic classification and statistical analysis of the midge larva species, and can meet the requirements of multiple application scenes such as environmental monitoring, biodiversity research and the like. Through report generation and a user interface, a user can easily acquire, view and export analysis results, so that related works such as water ecological monitoring are supported.
In general, the invention provides an automatic midge larva recognition system by combining deep learning and image recognition technology, which can realize multi-focal-plane image acquisition, image superposition processing, microstructure feature extraction and automatic classification of species. The system has higher recognition precision and processing efficiency, can effectively reduce manual intervention, improves automatic recognition and classification precision of the midge larvae, and meets the requirements of water ecological environment monitoring and biodiversity research. The system also provides a real-time statistical analysis and report generation function, and provides comprehensive and timely data support for environmental monitoring.
Referring to fig. 8, the invention also provides a method for identifying midge larvae based on deep learning and image identification, which comprises the following steps:
s1, acquiring multi-focal plane images of the midge larvae through an image acquisition module, wherein the image acquisition module comprises a transmission light source, an optical microscope and a high-resolution digital camera, and the digital camera is controlled to automatically acquire image sequences under different focal planes;
s2, performing image superposition processing on the acquired multi-focal-plane image, wherein the image superposition processing comprises the following steps:
Decomposing each focal plane image into a plurality of scale layers through multi-scale Laplacian pyramid decomposition;
Calculating fusion weight according to the local gradient, contrast and motion compensation result of the image, and fusing a plurality of focal plane images to generate a full focal plane image;
S3, inputting the superimposed full-focus image into a deep learning recognition module, wherein the deep learning recognition module extracts the characteristics of the midge larvae based on a convolutional neural network, and the characteristic extraction process comprises convolution, activation and pooling operations;
S4, classifying the extracted features through a full-connection layer of the deep learning identification module and a Softmax classifier, and outputting a type tag of the midge larva and a corresponding classification confidence;
s5, storing the classification result, the image number, the time stamp and the detection position information into a database;
S6, a statistical analysis step, namely carrying out statistical analysis on classification results through a type automatic classification and statistics module, calculating the number of the midge larvae of different types and time sequence changes thereof, generating a type distribution and time trend chart, and automatically generating a report.
The principle of the process flow of the method of the present invention is fully described in the foregoing system embodiment, and is not repeated herein.
Although embodiments of the present invention have been shown and described, it will be understood by those skilled in the art that various changes, modifications, substitutions and alterations can be made therein without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (10)

1.基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,包括:1. A chironomid larvae identification system based on deep learning and image recognition, characterized by comprising: 图像采集模块,包括带有透射光源的光学显微镜、显微镜物镜和数字相机,所述数字相机安装于显微镜上,并通过数据传输接口与计算机连接;所述图像采集模块受计算机软件控制,在不同焦平面下自动采集摇蚊幼虫的多层次图像;An image acquisition module includes an optical microscope with a transmission light source, a microscope objective lens and a digital camera, wherein the digital camera is installed on the microscope and connected to a computer via a data transmission interface; the image acquisition module is controlled by computer software and automatically acquires multi-level images of chironomid larvae at different focal planes; 自动化样品处理装置,包括样品加载系统和三轴电动载物台,所述样品加载系统将摇蚊幼虫样本输送至成像位置,三轴电动载物台调整样本的XYZ轴位置,用于多视野和多焦平面下的自动图像采集;An automated sample handling device, comprising a sample loading system and a three-axis motorized stage, wherein the sample loading system transports the chironomid larvae sample to an imaging position, and the three-axis motorized stage adjusts the XYZ axis position of the sample for automatic image acquisition in multiple fields of view and multiple focal planes; 图像叠加处理模块,所述图像叠加处理模块采用自适应多尺度图像融合,对所获取的不同焦距平面的图像进行高精度图像叠加,生成包含从下至上多个深度信息的全焦面图像;An image stacking processing module, wherein the image stacking processing module uses adaptive multi-scale image fusion to perform high-precision image stacking on the acquired images of different focal length planes to generate a full-focus plane image containing multiple depth information from bottom to top; 深度学习识别模块,基于卷积神经网络模型,通过对叠加后生成的全焦面图像进行训练和特征提取,自动识别摇蚊幼虫的微细结构特征,包括但不限于颏板、内唇栉、眼点和尾刚毛特征,以实现对不同亚科和属的摇蚊幼虫种类的分类;The deep learning recognition module, based on the convolutional neural network model, automatically identifies the microstructural features of the chironomid larvae, including but not limited to the chin plate, inner labial comb, eyespots and caudal setae, by training and extracting features from the full-focus images generated after superposition, so as to achieve the classification of chironomid larvae of different subfamilies and genera; 种类自动分类与统计模块,所述种类自动分类与统计模块通过结合深度学习模型的输出,自动识别和分类摇蚊幼虫的种类,并提供统计分析,结果存储于数据库,支持生物监测与水生态环境的自动化监测需求。The module for automatic species classification and statistics automatically identifies and classifies the species of chironomid larvae by combining the output of the deep learning model, and provides statistical analysis. The results are stored in a database to support the needs of automated monitoring of biological monitoring and aquatic ecological environment. 2.根据权利要求1所述的基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,所述样品加载系统包括蠕动泵,所述蠕动泵通过进样软管与样液瓶连接,所述蠕动泵通过送样软管连接在三轴电动载物台的成像位置,所述三轴电动载物台的成像位置通过排样软管与废液桶连接。2. According to the deep learning and image recognition based chironomid larvae identification system of claim 1, it is characterized in that the sample loading system includes a peristaltic pump, the peristaltic pump is connected to the sample liquid bottle through a sampling hose, the peristaltic pump is connected to the imaging position of the three-axis electric stage through a sample delivery hose, and the imaging position of the three-axis electric stage is connected to the waste liquid barrel through a sample discharge hose. 3.根据权利要求2所述的基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,所述三轴电动载物台中部设置有盛样槽,所述送样软管和排样软管均嵌于三轴电动载物台内,且端部连接在盛样槽内。3. According to the deep learning and image recognition based chironomid larvae identification system of claim 2, it is characterized in that a sample holding slot is provided in the middle of the three-axis electric stage, the sample delivery hose and the sample discharge hose are both embedded in the three-axis electric stage, and the ends are connected to the sample holding slot. 4.根据权利要求3所述的基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,所述三轴电动载物台上表面位于盛样槽外沿设置有玻片槽,用于放置玻片,压片摇蚊幼虫进行识别。4. The chironomid larvae identification system based on deep learning and image recognition according to claim 3 is characterized in that a glass slide groove is provided on the upper surface of the three-axis electric stage at the outer edge of the sample holding groove for placing glass slides to press the chironomid larvae for identification. 5.根据权利要求1所述的基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,所述图像叠加处理模块包括:5. The chironomid larvae identification system based on deep learning and image recognition according to claim 1, characterized in that the image overlay processing module comprises: 多焦平面图像采集单元,用于从不同的焦平面获取摇蚊幼虫的图像序列,所述图像序列包含多个不同深度的图像;A multi-focal plane image acquisition unit, used for acquiring an image sequence of the chironomid larvae from different focal planes, wherein the image sequence includes multiple images at different depths; 多尺度图像分解单元,用于对所获取的图像序列中的每个图像进行拉普拉斯金字塔分解,生成多个不同尺度的图像层次,分解过程为:The multi-scale image decomposition unit is used to perform Laplace pyramid decomposition on each image in the acquired image sequence to generate multiple image levels of different scales. The decomposition process is as follows: 其中,表示第i个焦平面图像在第l个尺度的拉普拉斯金字塔表示,是图像的高斯金字塔表示;in, represents the Laplacian pyramid representation of the i-th focal plane image at the l-th scale, is the Gaussian pyramid representation of the image; 自适应加权融合单元,用于根据图像的局部梯度、对比度以及运动补偿结果计算融合权重,所述权重计算公式为:The adaptive weighted fusion unit is used to calculate the fusion weight according to the local gradient, contrast and motion compensation results of the image. The weight calculation formula is: 其中,wi(x,y)为第i个焦平面图像的权重,为图像的局部梯度,Ci(x,y)为局部对比度,Mi(x,y)为运动补偿后的图像质量权重;Among them, w i (x, y) is the weight of the i-th focal plane image, is the local gradient of the image, Ci (x, y) is the local contrast, and Mi (x, y) is the image quality weight after motion compensation; 图像融合单元,用于将经自适应加权计算后的各尺度图像进行融合,所述融合过程为:The image fusion unit is used to fuse the images of each scale after adaptive weighted calculation. The fusion process is as follows: 其中,为第l个尺度下的融合图像;in, is the fused image at the lth scale; 图像重构单元,用于通过逆拉普拉斯金字塔过程重构最终的全焦面图像,重构公式为:The image reconstruction unit is used to reconstruct the final full-focus image through an inverse Laplace pyramid process. The reconstruction formula is: 其中,Ifused(x,y)为融合后的最终全焦面图像,L为分解的尺度数量。Among them, I fused (x, y) is the final full-focus plane image after fusion, and L is the number of decomposed scales. 6.根据权利要求5所述的基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,所述自适应加权融合单元还包括光流法运动补偿单元,用于在多焦平面图像叠加前检测并校正因液体环境中的样本运动引起的位移,所述运动补偿通过以下公式计算:6. The chironomid larvae identification system based on deep learning and image recognition according to claim 5, characterized in that the adaptive weighted fusion unit also includes an optical flow motion compensation unit, which is used to detect and correct the displacement caused by the sample movement in the liquid environment before the multi-focal plane image is superimposed, and the motion compensation is calculated by the following formula: 其中,v(x,y)为样本在时间t时的运动矢量,I(x,y,t)为时刻t时的图像灰度值,运动矢量用于在图像叠加前进行校正。Among them, v(x,y) is the motion vector of the sample at time t, I(x,y,t) is the image grayscale value at time t, and the motion vector is used for correction before image superposition. 7.根据权利要求6所述的基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,所述图像重构单元在全焦面图像重构过程中采用局部特征增强技术,通过拉普拉斯算子对摇蚊幼虫的特定微观结构进行细节增强,增强后的重构公式为:7. The chironomid larvae identification system based on deep learning and image recognition according to claim 6 is characterized in that the image reconstruction unit adopts local feature enhancement technology in the process of full-focal plane image reconstruction, and enhances the details of the specific microstructure of the chironomid larvae through the Laplace operator, and the enhanced reconstruction formula is: Ienhanced(x,y)=Ifused(x,y)+λ·ΔIfused(x,y)I enhanced (x,y)=I fused (x,y)+λ·ΔI fused (x,y) 其中,ΔIfused(x,y)为全焦面图像的拉普拉斯算子结果,λ为控制增强强度的参数。Wherein, ΔI fused (x, y) is the Laplace operator result of the full-focus plane image, and λ is the parameter that controls the enhancement intensity. 8.根据权利要求1所述的基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,所述深度学习识别模块包括:8. The chironomid larvae identification system based on deep learning and image recognition according to claim 1, characterized in that the deep learning identification module comprises: 输入预处理单元,用于将由图像叠加处理模块生成的全焦面图像进行归一化处理和数据增强处理;An input preprocessing unit is used to perform normalization processing and data enhancement processing on the full-focus plane image generated by the image superposition processing module; 卷积特征提取单元,用于通过卷积神经网络提取图像中的高层次特征,卷积过程为:The convolution feature extraction unit is used to extract high-level features from images through a convolutional neural network. The convolution process is: 其中,Inormalized(x,y)为归一化后的全焦面图像,w(i,j)为卷积核的权重,F(x,y)为卷积结果,b为偏置;Among them, I normalized (x, y) is the normalized full-focal plane image, w(i, j) is the weight of the convolution kernel, F(x, y) is the convolution result, and b is the bias; 激活单元,用于对卷积结果应用ReLU激活函数,激活函数的计算公式为:The activation unit is used to apply the ReLU activation function to the convolution result. The calculation formula of the activation function is: A(x,y)=max(0,F(x,y))A(x,y)=max(0,F(x,y)) 其中,A(x,y)为激活后的输出;Among them, A(x,y) is the output after activation; 池化单元,用于通过最大池化操作对激活后的特征图进行降维和下采样处理,池化计算公式为:The pooling unit is used to reduce the dimension and downsample the activated feature map through the maximum pooling operation. The pooling calculation formula is: 其中,P为池化窗口的大小;Where P is the size of the pooling window; 全连接层单元,用于将池化后的特征图展平为一维特征向量并映射到目标分类空间,全连接层的计算公式为:The fully connected layer unit is used to flatten the pooled feature map into a one-dimensional feature vector and map it to the target classification space. The calculation formula of the fully connected layer is: z=W·x+bz=W·x+b 其中,W为全连接层的权重矩阵,x为输入特征向量,b为偏置;Among them, W is the weight matrix of the fully connected layer, x is the input feature vector, and b is the bias; Softmax分类单元,用于将全连接层的输出转换为各类摇蚊幼虫的概率分布,Softmax分类器的计算公式为:The Softmax classification unit is used to convert the output of the fully connected layer into the probability distribution of various types of chironomid larvae. The calculation formula of the Softmax classifier is: 其中,zj为类别j的得分,N为类别数量,p(y=j|x)为输入图像属于类别j的概率。Where zj is the score of category j, N is the number of categories, and p(y=j|x) is the probability that the input image belongs to category j. 9.根据权利要求1所述的基于深度学习和图像识别的摇蚊幼虫识别系统,其特征在于,其特征在于,所述种类自动分类与统计模块包括:9. The chironomid larvae identification system based on deep learning and image recognition according to claim 1, characterized in that the automatic species classification and statistics module comprises: 分类结果存储单元,用于接收深度学习识别模块的分类结果,所述分类结果包括摇蚊幼虫的种类标签和置信度,并将分类结果与对应的图像编号、时间戳、检测位置元数据一同存储至数据库;A classification result storage unit, used to receive the classification result of the deep learning recognition module, the classification result including the species label and confidence of the chironomid larvae, and store the classification result together with the corresponding image number, timestamp, and detection position metadata in a database; 置信度筛选单元,用于根据分类结果中的置信度进行筛选,当分类置信度低于预设阈值时,将该结果标记为不确定分类,所述筛选过程定义为:The confidence screening unit is used to screen according to the confidence in the classification result. When the classification confidence is lower than a preset threshold, the result is marked as an uncertain classification. The screening process is defined as: 其中,yi为分类标签,pi为分类置信度,pthreshold为预设置信度阈值;Among them, yi is the classification label, pi is the classification confidence, and pthreshold is the preset confidence threshold; 种类统计单元,用于对存储的分类结果进行统计分析,计算不同种类摇蚊幼虫的数量及其比例,统计过程定义为:The species statistics unit is used to perform statistical analysis on the stored classification results and calculate the number and proportion of different species of chironomid larvae. The statistical process is defined as: 其中,1(yj=yi)为指示函数,当yj=yi时返回1,否则返回0,m为样本数量;Where 1(y j = y i ) is an indicator function, which returns 1 when y j = y i and returns 0 otherwise, and m is the number of samples; 时间序列分析单元,用于基于历史数据对摇蚊幼虫的种类检测结果进行时间序列分析,生成不同种类摇蚊幼虫数量随时间变化的趋势图,并对检测数据进行动态监测;A time series analysis unit, used to perform time series analysis on the detection results of the species of chironomid larvae based on historical data, generate a trend graph of the number of different species of chironomid larvae changing over time, and dynamically monitor the detection data; 报告生成单元,用于根据种类统计和时间序列分析结果自动生成报告,所述报告包括种类分布图、时间变化趋势图,并支持将报告导出为PDF或Excel格式。The report generation unit is used to automatically generate a report based on the category statistics and time series analysis results. The report includes a category distribution diagram, a time change trend diagram, and supports exporting the report to PDF or Excel format. 10.基于深度学习和图像识别的摇蚊幼虫识别方法,使用如权利要求1-9任一项所述的系统,其特征在于,包括以下步骤:10. A method for identifying chironomid larvae based on deep learning and image recognition, using the system according to any one of claims 1 to 9, characterized in that it comprises the following steps: 图像采集步骤,通过图像采集模块获取摇蚊幼虫的多焦平面图像,所述图像采集模块包括透射光源、光学显微镜和高分辨率数字相机,控制数字相机在不同焦平面下自动采集图像序列;An image acquisition step, obtaining multi-focal plane images of the chironomid larvae through an image acquisition module, wherein the image acquisition module includes a transmitted light source, an optical microscope and a high-resolution digital camera, and controls the digital camera to automatically acquire image sequences at different focal planes; 图像叠加处理步骤,对所获取的多焦平面图像进行图像叠加处理,所述图像叠加处理包括:The image stacking processing step is to perform image stacking processing on the acquired multi-focal plane images, and the image stacking processing includes: 通过多尺度拉普拉斯金字塔分解将各焦平面图像分解为多个尺度层次;Each focal plane image is decomposed into multiple scale levels through multi-scale Laplacian pyramid decomposition; 根据图像的局部梯度、对比度以及运动补偿结果计算融合权重,将多个焦平面图像融合生成全焦面图像;The fusion weight is calculated according to the local gradient, contrast and motion compensation results of the image, and multiple focal plane images are fused to generate a full focal plane image; 特征提取步骤,将叠加后的全焦面图像输入深度学习识别模块,所述深度学习识别模块基于卷积神经网络提取摇蚊幼虫的特征,特征提取过程包括卷积、激活、池化操作;A feature extraction step, inputting the superimposed full-focus surface image into a deep learning recognition module, wherein the deep learning recognition module extracts features of the chironomid larvae based on a convolutional neural network, and the feature extraction process includes convolution, activation, and pooling operations; 分类步骤,通过深度学习识别模块的全连接层和Softmax分类器对提取的特征进行分类,输出摇蚊幼虫的种类标签和对应的分类置信度;In the classification step, the extracted features are classified through the fully connected layer and Softmax classifier of the deep learning recognition module, and the species label of the midge larvae and the corresponding classification confidence are output; 分类结果存储步骤,将分类结果连同图像编号、时间戳、检测位置信息存储至数据库;a classification result storage step, storing the classification result together with the image number, timestamp, and detection position information into a database; 统计分析步骤,通过种类自动分类与统计模块对分类结果进行统计分析,计算不同种类摇蚊幼虫的数量及其时间序列变化,生成种类分布和时间趋势图,并自动生成报告。In the statistical analysis step, the classification results are statistically analyzed through the automatic species classification and statistical module, the number of different species of chironomid larvae and their time series changes are calculated, the species distribution and time trend graphs are generated, and a report is automatically generated.
CN202411318584.7A 2024-09-20 2024-09-20 Chironomid larvae identification system based on deep learning and image recognition Pending CN119295806A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202411318584.7A CN119295806A (en) 2024-09-20 2024-09-20 Chironomid larvae identification system based on deep learning and image recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202411318584.7A CN119295806A (en) 2024-09-20 2024-09-20 Chironomid larvae identification system based on deep learning and image recognition

Publications (1)

Publication Number Publication Date
CN119295806A true CN119295806A (en) 2025-01-10

Family

ID=94162076

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202411318584.7A Pending CN119295806A (en) 2024-09-20 2024-09-20 Chironomid larvae identification system based on deep learning and image recognition

Country Status (1)

Country Link
CN (1) CN119295806A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119533337A (en) * 2025-01-17 2025-02-28 华东交通大学 Monocular structured light three-dimensional imaging method and system based on multi-frequency polarization projection

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105744159A (en) * 2016-02-15 2016-07-06 努比亚技术有限公司 Image synthesizing method and device
CN108020509A (en) * 2017-12-12 2018-05-11 佛山科学技术学院 The method and its device of a kind of optical projection tomography
CN114202490A (en) * 2021-11-23 2022-03-18 深圳大学 Wear particle surface reconstruction method based on multi-focus image and related device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105744159A (en) * 2016-02-15 2016-07-06 努比亚技术有限公司 Image synthesizing method and device
CN108020509A (en) * 2017-12-12 2018-05-11 佛山科学技术学院 The method and its device of a kind of optical projection tomography
CN114202490A (en) * 2021-11-23 2022-03-18 深圳大学 Wear particle surface reconstruction method based on multi-focus image and related device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
朱敏慧等: "基于深度学习的白纹伊蚊卵粒识别模型初步研究", 《寄生虫与医学昆虫学报》, vol. 31, no. 1, 14 May 2024 (2024-05-14), pages 12 - 18 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119533337A (en) * 2025-01-17 2025-02-28 华东交通大学 Monocular structured light three-dimensional imaging method and system based on multi-frequency polarization projection

Similar Documents

Publication Publication Date Title
CN110879999B (en) Micro microscopic image acquisition device based on mobile phone and image splicing and identifying method
CN113392775B (en) Sugarcane seedling automatic identification and counting method based on deep neural network
CN110794569B (en) Cell micro microscopic image acquisition device and image identification method
CA2436043C (en) Focusable virtual microscopy apparatus and method
US8116543B2 (en) System for and method of intelligently directed segmentation analysis for automated microscope systems
AU2019370154B2 (en) Digital imaging system and method
CN111443028A (en) Automatic monitoring equipment and method for floating algae based on AI technology
CN119295806A (en) Chironomid larvae identification system based on deep learning and image recognition
WO2006023675A2 (en) Microscopy system having automatic and interactive modes for forming a magnified mosaic image and associated method
CN102122066A (en) Full-automatic double-ccd photosensitive element digital microscope
EP4064194B1 (en) Method for analyzing yeasts
CN113130049A (en) Intelligent pathological image diagnosis system based on cloud service
EP4063836A1 (en) Algae analysis method
CN100392403C (en) A method for automatically counting the number of white blood cells in blood microscopic images
CN116129096B (en) Digital measurement system and method for potato root length
CN107680152A (en) Target surface topography measurement method and apparatus based on image procossing
CN115760837A (en) A crystal quality assessment method and system based on deep neural network
GB2610426A (en) System and method for identifying and counting biological species
US20210233647A1 (en) Digital imaging system and method
CN114774253A (en) Portable spore capturing and detecting device and detecting method thereof
CN202110328U (en) Full-automatic double charge coupled device (CCD) sensitization element digital microscope
CN112446421A (en) Silkworm cocoon counting and identifying method based on machine vision
CN112992336A (en) Intelligent pathological diagnosis system
CN111402242B (en) Medical image refinement processing system and method
CN119579585B (en) Fish in-situ video processing method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination