图像处理方法、图像处理装置以及存储介质Image processing method, image processing device and storage medium
本公开要求于2019年9月29日递交的中国专利申请第201910935208.5号的优先权,在此全文引用上述中国专利申请公开的内容以作为本公开的一部分。This disclosure claims the priority of the Chinese patent application No. 201910935208.5 filed on September 29, 2019, and the contents of the above-mentioned Chinese patent application are quoted here in full as a part of this disclosure.
技术领域Technical field
本公开的实施例涉及一种图像处理方法、图像处理装置以及存储介质。The embodiments of the present disclosure relate to an image processing method, an image processing device, and a storage medium.
背景技术Background technique
图像分类是指根据一定的分类规则将输入图像自动分到一组预定义类别中。例如,根据图像中包含的语义信息,可以对输入图像进行对象分类、场景分类等。例如,可以识别输入图像中包含的预设的目标对象,并根据识别的对象进行分类。又例如,也可以根据输入图像中的语义信息将具有相似内容的图像划分成相同的类别。Image classification refers to automatically classifying input images into a set of predefined categories according to certain classification rules. For example, according to the semantic information contained in the image, the input image can be classified into objects and scenes. For example, a preset target object contained in the input image can be recognized and classified according to the recognized object. For another example, images with similar content can also be classified into the same category according to semantic information in the input image.
发明内容Summary of the invention
本公开至少一实施例提供一种图像处理装置,包括:深度特征提取器,配置为获取待识别图像的深度特征,所述待识别图像为医学图像;专家特征提取器,配置为获取所述待识别图像的专家特征;融合处理器,配置为融合所述深度特征以及所述专家特征,以获得所述待识别图像的融合特征;分类处理器,配置为根据所述待识别图像的融合特征对所述待识别图像进行分类。At least one embodiment of the present disclosure provides an image processing device, including: a depth feature extractor configured to obtain the depth feature of an image to be recognized, the image to be recognized is a medical image; and an expert feature extractor, configured to obtain the Identify the expert features of the image; a fusion processor configured to fuse the depth feature and the expert feature to obtain the fusion feature of the image to be recognized; the classification processor is configured to pair according to the fusion feature of the image to be recognized The image to be recognized is classified.
例如,本公开至少一实施例提供的图像处理装置,还包括:无监督特征提取器,配置为获取所述待识别图像的无监督特征;所述融合处理器还配置为融合所述深度特征、所述专家特征以及所述无监督特征,以获得所述待识别图像的融合特征。For example, the image processing device provided by at least one embodiment of the present disclosure further includes: an unsupervised feature extractor configured to obtain unsupervised features of the image to be recognized; the fusion processor is further configured to fuse the depth features, The expert feature and the unsupervised feature are used to obtain the fusion feature of the image to be recognized.
例如,在本公开至少一实施例提供的图像处理装置中,所述深度特征提取器还配置为利用深度神经网络获取所述待识别图像的深度特征。For example, in the image processing device provided by at least one embodiment of the present disclosure, the depth feature extractor is further configured to obtain the depth feature of the image to be recognized by using a deep neural network.
例如,在本公开至少一实施例提供的图像处理装置中,所述专家特征提取器还配置为基于根据医学图像数据获得的经验公式、规则和特征值,提取所述待识别图像的专家特征。For example, in the image processing device provided by at least one embodiment of the present disclosure, the expert feature extractor is further configured to extract the expert features of the image to be recognized based on empirical formulas, rules, and feature values obtained from medical image data.
例如,在本公开至少一实施例提供的图像处理装置中,所述专家特征的类别包括统计、形态、时域和频域中的至少之一。For example, in the image processing device provided by at least one embodiment of the present disclosure, the category of the expert feature includes at least one of statistics, morphology, time domain, and frequency domain.
例如,在本公开至少一实施例提供的图像处理装置中,所述无监督特征提取器还配置为在所述基于无监督特征提取器获取所述待识别图像的无监督特征之前,利用主成分分析法、随机投影法和序列自动编码器中的至少之一训练得到所述无监督特征提取器。For example, in the image processing device provided by at least one embodiment of the present disclosure, the unsupervised feature extractor is further configured to use principal components before acquiring the unsupervised features of the image to be recognized based on the unsupervised feature extractor. At least one of the analysis method, the random projection method and the sequence autoencoder is trained to obtain the unsupervised feature extractor.
例如,在本公开至少一实施例提供的图像处理装置中,所述融合处理器还配置为:拼接所述深度特征、所述专家特征以及所述无监督特征,以获得所述融合特征。For example, in the image processing device provided by at least one embodiment of the present disclosure, the fusion processor is further configured to splice the depth feature, the expert feature, and the unsupervised feature to obtain the fusion feature.
例如,在本公开至少一实施例提供的图像处理装置中,所述融合处理器还配置为:分 别对所述深度特征、所述专家特征与所述无监督特征进行全局池化操作和均值池化操作,以分别获取所述深度特征的全局向量和均值向量、所述专家特征的全局向量和均值向量以及所述无监督特征的全局向量和均值向量;拼接所述深度特征的全局向量和均值向量的至少之一、所述专家特征的全局向量和均值向量的至少之一以及所述无监督特征的全局向量和均值向量的至少之一,以获得所述融合特征。For example, in the image processing device provided by at least one embodiment of the present disclosure, the fusion processor is further configured to: perform a global pooling operation and an average pooling operation on the depth feature, the expert feature, and the unsupervised feature, respectively. Operation to obtain the global vector and mean vector of the depth feature, the global vector and mean vector of the expert feature, and the global vector and mean vector of the unsupervised feature; splicing the global vector and mean of the depth feature At least one of the vectors, at least one of the global vector and the mean vector of the expert feature, and at least one of the global vector and the mean vector of the unsupervised feature, to obtain the fusion feature.
例如,在本公开至少一实施例提供的图像处理装置中,所述分类处理器还配置为:根据所述待识别图像的融合特征判断所述待识别图像是否包含房颤特征。For example, in the image processing device provided by at least one embodiment of the present disclosure, the classification processor is further configured to determine whether the image to be identified contains atrial fibrillation features according to the fusion feature of the image to be identified.
本公开至少一实施例提供一种图像处理方法,包括:基于深度特征提取器获取待识别图像的深度特征,所述待识别图像为医学图像;基于专家特征提取器获取所述待识别图像的专家特征;融合所述深度特征以及所述专家特征,以获得所述待识别图像的融合特征;根据所述待识别图像的融合特征对所述待识别图像进行分类。At least one embodiment of the present disclosure provides an image processing method, including: obtaining a depth feature of an image to be recognized based on a depth feature extractor, where the image to be recognized is a medical image; and an expert who obtains the image to be recognized based on an expert feature extractor Feature; fusion of the depth feature and the expert feature to obtain the fusion feature of the image to be recognized; classify the image to be recognized according to the fusion feature of the image to be recognized.
例如,本公开至少一实施例提供的图像处理方法,还包括基于无监督特征提取器获取所述待识别图像的无监督特征;融合所述深度特征、所述专家特征以及所述无监督特征,以获得所述待识别图像的融合特征。For example, the image processing method provided by at least one embodiment of the present disclosure further includes obtaining unsupervised features of the image to be recognized based on an unsupervised feature extractor; fusing the depth feature, the expert feature, and the unsupervised feature, To obtain the fusion feature of the image to be recognized.
例如,在本公开至少一实施例提供的图像处理方法中,融合所述深度特征、所述专家特征以及所述无监督特征,以获得所述待识别图像的融合特征,包括:拼接所述深度特征、所述专家特征以及所述无监督特征,以获得所述融合特征。For example, in the image processing method provided by at least one embodiment of the present disclosure, fusing the depth feature, the expert feature, and the unsupervised feature to obtain the fusion feature of the image to be recognized includes: stitching the depth Feature, the expert feature, and the unsupervised feature to obtain the fusion feature.
例如,在本公开至少一实施例提供的图像处理方法中,融合所述深度特征、所述专家特征以及所述无监督特征,以获得所述待识别图像的融合特征,包括:分别对所述深度特征、所述专家特征与所述无监督特征进行全局池化操作和均值池化操作,以分别获取所述深度特征的全局向量和均值向量、所述专家特征的全局向量和均值向量以及所述无监督特征的全局向量和均值向量;拼接所述深度特征的全局向量和均值向量的至少之一、所述专家特征的全局向量和均值向量的至少之一以及所述无监督特征的全局向量和均值向量的至少之一,以获得所述融合特征。For example, in the image processing method provided by at least one embodiment of the present disclosure, fusing the depth feature, the expert feature, and the unsupervised feature to obtain the fusion feature of the image to be recognized includes: The depth feature, the expert feature and the unsupervised feature perform a global pooling operation and an average pooling operation to obtain the global vector and the mean vector of the depth feature, the global vector and the mean vector of the expert feature, and the total The global vector and the mean vector of the unsupervised feature; splicing at least one of the global vector and the mean vector of the depth feature, at least one of the global vector and the mean vector of the expert feature, and the global vector of the unsupervised feature And at least one of the mean vector to obtain the fusion feature.
本公开至少一实施例还提供一种图像处理装置,包括:处理器;存储器;一个或多个计算机程序模块,所述一个或多个计算机程序模块被存储在所述存储器中并被配置为由所述处理器执行,所述一个或多个计算机程序模块包括用于执行实现本公开任一实施例提供的图像处理方法的指令。At least one embodiment of the present disclosure further provides an image processing device, including: a processor; a memory; one or more computer program modules, the one or more computer program modules are stored in the memory and configured to be configured by Executed by the processor, the one or more computer program modules include instructions for executing the image processing method provided by any embodiment of the present disclosure.
本公开至少一实施例还提供一种存储介质,所述存储介质存储有计算机可读指令,当所述计算机可读指令由处理器执行时可以执行本公开任一实施例提供的图像处理方法。At least one embodiment of the present disclosure further provides a storage medium that stores computer-readable instructions, and when the computer-readable instructions are executed by a processor, the image processing method provided in any embodiment of the present disclosure can be executed.
附图说明Description of the drawings
为了更清楚地说明本公开实施例的技术方案,下面将对实施例的附图作简单地介绍,显而易见地,下面描述中的附图仅仅涉及本公开的一些实施例,而非对本公开的限制。In order to explain the technical solutions of the embodiments of the present disclosure more clearly, the following will briefly introduce the drawings of the embodiments. Obviously, the drawings in the following description only refer to some embodiments of the present disclosure, rather than limiting the present disclosure. .
图1A为本公开至少一实施例提供的一种图像处理方法的流程图;FIG. 1A is a flowchart of an image processing method provided by at least one embodiment of the present disclosure;
图1B示出了根据本公开实施例的图像处理系统的示例性的场景图;FIG. 1B shows an exemplary scene diagram of an image processing system according to an embodiment of the present disclosure;
图2为本公开至少一实施例提供的一种深度特征的提取示意图;FIG. 2 is a schematic diagram of extracting a depth feature provided by at least one embodiment of the present disclosure;
图3A为本公开至少一实施例提供的一种融合操作的流程图;3A is a flowchart of a fusion operation provided by at least one embodiment of the present disclosure;
图3B为本公开至少一实施例提供的一种融合操作的示意图;FIG. 3B is a schematic diagram of a fusion operation provided by at least one embodiment of the present disclosure;
图4为本公开至少一实施例提供的另一种图像处理方法的流程图;4 is a flowchart of another image processing method provided by at least one embodiment of the present disclosure;
图5为本公开至少一实施例提供的另一种融合操作的示意图;FIG. 5 is a schematic diagram of another fusion operation provided by at least one embodiment of the present disclosure;
图6为本公开至少一实施例提供的一种图像处理装置的示意框图;FIG. 6 is a schematic block diagram of an image processing apparatus provided by at least one embodiment of the present disclosure;
图7为本公开至少一实施例提供的另一种图像处理装置的示意框图;FIG. 7 is a schematic block diagram of another image processing apparatus provided by at least one embodiment of the present disclosure;
图8为本公开至少一实施例提供的另一种图像处理装置的示意框图;FIG. 8 is a schematic block diagram of another image processing apparatus provided by at least one embodiment of the present disclosure;
图9为本公开至少一实施例提供的一种电子设备的示意图;以及FIG. 9 is a schematic diagram of an electronic device provided by at least one embodiment of the present disclosure; and
图10为本公开至少一实施例提供的一种存储介质的示意图。FIG. 10 is a schematic diagram of a storage medium provided by at least one embodiment of the present disclosure.
具体实施方式detailed description
为使本公开实施例的目的、技术方案和优点更加清楚,下面将结合本公开实施例的附图,对本公开实施例的技术方案进行清楚、完整地描述。显然,所描述的实施例是本公开的一部分实施例,而不是全部的实施例。基于所描述的本公开的实施例,本领域普通技术人员在无需创造性劳动的前提下所获得的所有其他实施例,都属于本公开保护的范围。In order to make the objectives, technical solutions, and advantages of the embodiments of the present disclosure clearer, the technical solutions of the embodiments of the present disclosure will be described clearly and completely in conjunction with the accompanying drawings of the embodiments of the present disclosure. Obviously, the described embodiments are part of the embodiments of the present disclosure, rather than all of the embodiments. Based on the described embodiments of the present disclosure, all other embodiments obtained by those of ordinary skill in the art without creative labor are within the protection scope of the present disclosure.
除非另外定义,本公开使用的技术术语或者科学术语应当为本公开所属领域内具有一般技能的人士所理解的通常意义。本公开中使用的“第一”、“第二”以及类似的词语并不表示任何顺序、数量或者重要性,而只是用来区分不同的组成部分。同样,“一个”、“一”或者“该”等类似词语也不表示数量限制,而是表示存在至少一个。“包括”或者“包含”等类似的词语意指出现该词前面的元件或者物件涵盖出现在该词后面列举的元件或者物件及其等同,而不排除其他元件或者物件。“连接”或者“相连”等类似的词语并非限定于物理的或者机械的连接,而是可以包括电性的连接,不管是直接的还是间接的。“上”、“下”、“左”、“右”等仅用于表示相对位置关系,当被描述对象的绝对位置改变后,则该相对位置关系也可能相应地改变。Unless otherwise defined, the technical terms or scientific terms used in the present disclosure shall have the usual meanings understood by those with ordinary skills in the field to which this disclosure belongs. The "first", "second" and similar words used in the present disclosure do not indicate any order, quantity or importance, but are only used to distinguish different components. Similarly, similar words such as "a", "one" or "the" do not mean a quantity limit, but mean that there is at least one. "Include" or "include" and other similar words mean that the elements or items appearing before the word cover the elements or items listed after the word and their equivalents, but do not exclude other elements or items. Similar words such as "connected" or "connected" are not limited to physical or mechanical connections, but may include electrical connections, whether direct or indirect. "Up", "Down", "Left", "Right", etc. are only used to indicate the relative position relationship. When the absolute position of the described object changes, the relative position relationship may also change accordingly.
心电图(electrocardiogram,ECG)被广泛应用于各类心脏病的诊断,虽然目前医用心电图机和可穿戴的自动监测设备等医疗设备都具备一些基础心电图自动分析功能(如自动测量波形参数、节律参数等),但房颤等部分心律失常类型,由于医疗设备的自动分析诊断的错误率较高,房颤等部分心律失常的解读与诊断目前仍然主要依靠医学专家完成。Electrocardiogram (ECG) is widely used in the diagnosis of various heart diseases, although medical equipment such as current medical electrocardiographs and wearable automatic monitoring equipment all have some basic electrocardiogram automatic analysis functions (such as automatic measurement of waveform parameters, rhythm parameters, etc.) ), but for some types of arrhythmia such as atrial fibrillation, due to the high error rate of automatic analysis and diagnosis of medical equipment, the interpretation and diagnosis of some arrhythmia such as atrial fibrillation are still mainly completed by medical experts.
按照特征提取方法的不同,现有的房颤识别方法可以包括基于特征工程的方法和基于深度学习的方法。传统的房颤识别方法基本上都采用基于特征工程的方法,根据房颤分析机理的不同,基于特征工程的方法又可分为基于心房活动分析的方法、基于心室反应分析的方法、结合心房活动和心室反应的方法。基于心房活动性分析的方法主要关注房颤的P波消失特点或TQ间期上F波的出现。基于心房活动性分析的方法主要关注心房活动变化 产生的心电数据形态变化,如果心电信号数据的分辨率高、且几乎没有噪声污染,基于心房活动分析的房颤检测器可以实现高精度检测,但在具有一定噪声干扰且难以进行复杂降噪操作的实时场景之上则受影响较大。基于心室反应分析的方法主要关注基于QRS检测的心拍之间的时间间隔(RR间期长度)变化。RR间期主要是根据心电信号数据中波动幅度最大的R波的波峰位置确定,基于心室反应分析的方法受噪声的干扰能够比基于心房活动分析的方法小很多,也更适用于实时的房颤诊断问题。结合心房活动和心室反应的方法可以通过结合周期性的独立信号来提供更强的性能。结合心房活动和心室反应的方法包括:结合P波形态相似性度量和PR间期变异性的RR间期马尔可夫模型以及结合RR间期不规则、P波缺失、F波出现的模糊逻辑分类方法。According to different feature extraction methods, the existing atrial fibrillation recognition methods can include a method based on feature engineering and a method based on deep learning. Traditional atrial fibrillation recognition methods basically use feature engineering-based methods. According to the different analysis mechanisms of atrial fibrillation, feature engineering-based methods can be divided into methods based on atrial activity analysis, methods based on ventricular response analysis, and combined with atrial activity And the way the ventricles react. The method based on atrial activity analysis mainly focuses on the disappearance of P waves in atrial fibrillation or the appearance of F waves in the TQ interval. The method based on atrial activity analysis mainly focuses on changes in the shape of the ECG data caused by changes in atrial activity. If the resolution of the ECG signal data is high and there is almost no noise pollution, the atrial fibrillation detector based on atrial activity analysis can achieve high-precision detection , But it will be more affected in real-time scenes with certain noise interference and difficult to perform complex noise reduction operations. The method based on ventricular response analysis mainly focuses on the change of the time interval (RR interval length) between heart beats detected by QRS. The RR interval is mainly determined based on the peak position of the R wave with the largest fluctuation in the ECG signal data. The method based on ventricular response analysis can be much less interfered by noise than the method based on atrial activity analysis, and it is also more suitable for real-time atrium Diagnose the problem with tremor. Methods that combine atrial activity and ventricular response can provide greater performance by combining periodic independent signals. The methods of combining atrial activity and ventricular response include: RR interval Markov model combining P wave morphological similarity measurement and PR interval variability, and fuzzy logic classification combining RR interval irregularity, P wave absence, and F wave appearance method.
基于特征工程的方法与领域专家知识密切相关,因此也可以叫做基于领域知识的方法。现有的基于特征工程的房颤识别模型的研究在适用性上受到限制,仅能对部分心律失常进行分类。由于不同患者之间的波形变化复杂且许多非房颤信号表现出的特征可能类似于房颤信号的特征(例如不规则的RR间期等),而人工设计的专家特征需针对性尝试设计特征,虽然针对特定有限种类的心律失常能取得相对较好的识别效果,但是很难在各类心律失常混杂的复杂情况下都能准确区分出房颤和其他心律失常类型。原因主要包括两个方面:一方面由于很难保证提取了所有的特征,导致房颤识别模型在特征提取阶段可能丢弃了很多关键信息;另一方面,心电信号数据不可避免地包含工频干扰、电极接触噪声、人为运动、肌电干扰、基线漂移和心电信号的幅值变化、设备仪器噪声等大量噪声。在基于特征工程的方法中,由于很难从包含噪声的心电图中准确地进行参数测量和P波、T波、S波、F波等波形识别,因此基于特征工程的方法很容易受到噪声污染的影响。The method based on feature engineering is closely related to the knowledge of domain experts, so it can also be called the method based on domain knowledge. The existing researches on atrial fibrillation recognition model based on feature engineering are limited in applicability and can only classify some arrhythmias. Because the waveform changes between different patients are complex and the characteristics of many non-AF signals may be similar to the characteristics of atrial fibrillation signals (such as irregular RR intervals, etc.), artificially designed expert features need to be designed specifically. Although a relatively good recognition effect can be achieved for a specific limited type of arrhythmia, it is difficult to accurately distinguish atrial fibrillation from other types of arrhythmia under the complicated situation where various types of arrhythmia are mixed. The reasons mainly include two aspects: on the one hand, it is difficult to ensure that all the features are extracted, resulting in the atrial fibrillation recognition model may discard a lot of key information in the feature extraction stage; on the other hand, the ECG signal data inevitably contains power frequency interference , Electrode contact noise, human movement, EMG interference, baseline drift and ECG signal amplitude changes, equipment noise and other large amounts of noise. In the method based on feature engineering, it is difficult to accurately measure the parameters and recognize the waveforms such as P wave, T wave, S wave and F wave from the electrocardiogram containing noise. Therefore, the method based on feature engineering is very susceptible to noise pollution. influences.
由于深度学习强大的自动提取数据特征的能力,深度神经网络模型在生物医学信号中的应用受到了人们的广泛关注。Due to the powerful ability of deep learning to automatically extract data features, the application of deep neural network models in biomedical signals has received extensive attention.
近年来,基于深度学习的方法在心电图上的房颤检测中取得了成功。然而,这些方法仍然具有较高的错误诊断率,只有大约66%的房颤能够从各类心律失常混杂的心电图数据中正确识别。由于实际应用领域内已经积累了大量宝贵的领域知识,现阶段的深度神经网络还无法取代这些领域知识。因此,研究如何将深度神经网络与领域知识相结合,提升房颤自动检测的精度,是一个非常有价值的研究问题,也是现阶段很多方法考虑不足的地方所在。In recent years, methods based on deep learning have achieved success in the detection of atrial fibrillation on the electrocardiogram. However, these methods still have a high false diagnosis rate, and only about 66% of atrial fibrillation can be correctly identified from ECG data mixed with various arrhythmias. Since a large amount of valuable domain knowledge has been accumulated in the actual application field, the deep neural network at this stage cannot replace these domain knowledge. Therefore, studying how to combine deep neural networks with domain knowledge to improve the accuracy of automatic atrial fibrillation detection is a very valuable research problem, and it is also a place where many methods are not considered at this stage.
一般说来,深度神经网络模型的质量很大程度上取决于训练样本的质量,训练样本的类别越准确、内容越全面,训练得到的房颤识别模型的质量就越高。但是在实际应用中,全面且准确的训练样本是很难得到的。在心律失常识别应用中,由于心电图是连续采集的且人体的心电信号十分微弱,训练样本不可避免地包含各种噪声,这些包含噪声的训练样本将对最终的识别结果产生重要影响。由此,在房颤检测中,深度神经网络模型具有较高的错误率,一方面是由于训练数据量的不足;另一方面,心电图样本中的噪声片段(例如,噪声片段可以包括噪声和除当前心电图样本所属的心律失常类型之外的其他心律片段)导 致的语义模糊也是导致心律失常识别模型准确率不高的主要原因,利用深度学习技术从带有噪声数据片段的心电信号数据中自动学习特征将导致错误的特征被映射到当前的心律失常类型的数据分布上,从而导致深度神经网络模型的质量变差。Generally speaking, the quality of a deep neural network model largely depends on the quality of the training samples. The more accurate the types of training samples and the more comprehensive the content, the higher the quality of the atrial fibrillation recognition model obtained by training. However, in practical applications, it is difficult to obtain comprehensive and accurate training samples. In the application of arrhythmia recognition, since the ECG is continuously collected and the human body's ECG signal is very weak, the training samples inevitably contain various noises, and these training samples containing noise will have an important impact on the final recognition results. Therefore, in the detection of atrial fibrillation, the deep neural network model has a higher error rate. On the one hand, it is due to the insufficient amount of training data; on the other hand, the noise segment in the ECG sample (for example, the noise segment can include noise and The semantic ambiguity caused by the current ECG sample belongs to other arrhythmia types) is also the main reason for the low accuracy of the arrhythmia recognition model. Deep learning technology is used to automatically extract the ECG signal data from the noisy data segment. Learning features will cause the wrong features to be mapped to the data distribution of the current arrhythmia type, resulting in the deterioration of the quality of the deep neural network model.
因此,在实际房颤检测的应用场景下,现有的方法由于对于建模视角、领域知识等方面挑战带来的问题考虑不足,导致深度学习方法由于缺乏领域知识而很难被实际领域所接受,同时面对实际场景下的低数据质量问题时基于领域知识的房颤检测方法精确度低的局限性,使得研究场景下的房颤检测成果始终很难真正应用在实际场景中。Therefore, in the actual application scenario of atrial fibrillation detection, the existing methods have insufficient consideration of the problems caused by the modeling perspective, domain knowledge and other challenges, resulting in deep learning methods that are difficult to be accepted by the actual domain due to the lack of domain knowledge. At the same time, when faced with the problem of low data quality in actual scenarios, the domain knowledge-based atrial fibrillation detection method has the limitation of low accuracy, which makes it difficult to truly apply the results of atrial fibrillation detection in research scenarios in actual scenarios.
本公开至少一实施例提供一种图像处理方法,该图像处理方法包括:基于深度特征提取器获取待识别图像的深度特征,所述待识别图像为医学图像;基于专家特征提取器获取所述待识别图像的专家特征;融合所述深度特征以及所述专家特征,以获得所述待识别图像的融合特征;根据所述待识别图像的融合特征对所述待识别图像进行分类。At least one embodiment of the present disclosure provides an image processing method. The image processing method includes: acquiring a depth feature of an image to be recognized based on a depth feature extractor, where the image to be recognized is a medical image; and acquiring the image to be recognized based on an expert feature extractor. Identify the expert features of the image; fuse the depth feature and the expert feature to obtain the fusion feature of the image to be identified; classify the image to be identified according to the fusion feature of the image to be identified.
本公开一些实施例还提供对应于上述图像处理方法的图像处理装置和存储介质。Some embodiments of the present disclosure also provide an image processing device and a storage medium corresponding to the above-mentioned image processing method.
本公开上述实施例提供的图像处理方法、图像处理装置和存储介质,通过基于领域知识的专家特征的表示与提取,以及基于深度神经网络的深度特征的表示与提取,并采用统一的框架对专家特征和深度特征进行表示与融合,以达到提升房颤自动检测精度的目的,从而为实时、动态地房颤识别与诊断提供高精度的辅助方法,帮助医生及时诊断和准确发现患者的房颤的发生,帮助病人及时了解病情的变化,从而提高医疗质量、降低心脏性猝死等危及生命情况的发生率,最终减少给家庭和社会带来的健康和经济负担。The image processing method, image processing device, and storage medium provided by the above-mentioned embodiments of the present disclosure use the representation and extraction of expert features based on domain knowledge, and the representation and extraction of deep features based on deep neural networks, and adopt a unified framework for experts Features and depth features are expressed and fused to achieve the purpose of improving the accuracy of automatic detection of atrial fibrillation, thereby providing high-precision auxiliary methods for real-time and dynamic atrial fibrillation recognition and diagnosis, helping doctors to diagnose and accurately discover the patient’s atrial fibrillation in time To help patients understand the changes in their condition in time, thereby improving the quality of medical care, reducing the incidence of life-threatening conditions such as sudden cardiac death, and ultimately reducing the health and economic burdens brought to the family and society.
下面结合附图对本公开的实施例及其示例进行详细说明。The embodiments and examples of the present disclosure will be described in detail below with reference to the accompanying drawings.
本公开至少一实施例提供一种图像处理方法,图1A为该图像处理方法的一个示例的流程图。该图像处理方法可以以软件、硬件、固件或其任意组合的方式实现,由例如手机、笔记本电脑、桌面电脑、网络服务器、数码相机等设备中的处理器加载并执行,可以实现对待识别图像的分类,以用于后续步骤的处理。图1B示出了根据本公开实施例的图像处理系统的示例性的场景图。下面,参考图1A和图1B对本公开至少一实施例提供的图像处理方法进行说明。如图1B所示,该图像处理方法包括步骤S110至步骤S140。At least one embodiment of the present disclosure provides an image processing method, and FIG. 1A is a flowchart of an example of the image processing method. The image processing method can be implemented in the form of software, hardware, firmware or any combination thereof. It is loaded and executed by processors in devices such as mobile phones, notebook computers, desktop computers, network servers, digital cameras, etc., and can realize the image processing to be recognized. Classification for processing in subsequent steps. FIG. 1B shows an exemplary scene diagram of an image processing system according to an embodiment of the present disclosure. Hereinafter, the image processing method provided by at least one embodiment of the present disclosure will be described with reference to FIG. 1A and FIG. 1B. As shown in FIG. 1B, the image processing method includes steps S110 to S140.
步骤S110:基于深度特征提取器获取待识别图像的深度特征。Step S110: Obtain the depth feature of the image to be recognized based on the depth feature extractor.
步骤S120:基于专家特征提取器获取待识别图像的专家特征。Step S120: Obtain expert features of the image to be recognized based on the expert feature extractor.
步骤S130:融合深度特征以及专家特征,以获得待识别图像的融合特征。Step S130: Fusion of the depth feature and the expert feature to obtain the fusion feature of the image to be recognized.
步骤S140:根据待识别图像的融合特征对待识别图像进行分类。Step S140: Classify the image to be recognized according to the fusion feature of the image to be recognized.
例如,该待识别图像为医学图像。这里所说的医学图像可以是例如通过CT、MRI、超声、X光、核素显像(如SPECT、PET)等方法采集的医学图像,也可以是例如心电图、脑电图、光学摄影等显示人体生理信息的图像。在本公开的实施例中,以该医学图像为心电图像为例进行说明,本公开的实施例对此不作限制。在心电图中,房颤信号的特点是P波完全消失,P波被F波完全替代,并且心率绝对不齐(即任意相邻的两个RR间期均不相同)。For example, the image to be recognized is a medical image. The medical images mentioned here can be, for example, medical images collected by CT, MRI, ultrasound, X-ray, radionuclide imaging (such as SPECT, PET), etc., or can be displays such as electrocardiogram, electroencephalogram, optical photography, etc. Images of human body physiological information. In the embodiment of the present disclosure, the medical image is an electrocardiogram image as an example for description, which is not limited in the embodiment of the present disclosure. In the electrocardiogram, the characteristic of atrial fibrillation signal is that the P wave completely disappears, the P wave is completely replaced by the F wave, and the heart rate is absolutely uneven (that is, any two adjacent RR intervals are not the same).
对于步骤S110,例如,在一些示例中,该深度特征提取器可以实现为深度神经网络(例如,包括全连接层),该步骤S110包括利用深度神经网络获取待识别图像的深度特征。例如,该深度神经网络可以是卷积神经网络,如Inception系列网络(例如Googlenet等)、VGG系列网络、Resnet系列网络等中的任意一个或上述网络中任意一个的至少一部分,本公开的实施例对此不作限制。For step S110, for example, in some examples, the deep feature extractor may be implemented as a deep neural network (for example, including a fully connected layer), and this step S110 includes using the deep neural network to obtain the depth features of the image to be recognized. For example, the deep neural network may be a convolutional neural network, such as any one of the Inception series network (such as Googlenet, etc.), the VGG series network, the Resnet series network, etc., or at least a part of any one of the foregoing networks. Embodiments of the present disclosure There is no restriction on this.
在一些应用中,该深度神经网络输出的深度特征的向量维数较高,专家特征或无监督特征等向量的维数较低,由于深度特征的向量维数与专家特征或无监督特征等向量的维数相差较大,使得专家特征或无监督特征等在最终的融合特征中的作用微乎其微,从而在融合特征中体现不出它们的优势,从而达不到融合特征的目的,不能提升房颤检测的精度。因此,为了减小深度特征的向量维数与专家特征或无监督特征等向量的维数之间的差距,该深度特征提取器还可以包括与深度神经网络连接的激活函数层(例如,Identity层)。通过该Identity层可以对深度神经网络提取的深度特征进行降维,以得到降维后的深度特征,以与专家特征或无监督特征等的维数匹配,从而可以解决上述问题,提高融合特征的多样性,以达到提升房颤自动检测精度的目的。In some applications, the vector dimensionality of the deep features output by the deep neural network is relatively high, and the dimensionality of vectors such as expert features or unsupervised features is low, because the vector dimensionality of deep features is the same as that of expert features or unsupervised features. The dimensionality of the fusion is quite different, so that expert features or unsupervised features have little effect in the final fusion feature, so their advantages are not reflected in the fusion feature, and the purpose of the fusion feature is not achieved, and the atrial fibrillation cannot be improved. The accuracy of detection. Therefore, in order to reduce the gap between the vector dimension of deep features and the dimension of vectors such as expert features or unsupervised features, the deep feature extractor may also include an activation function layer connected to a deep neural network (for example, an Identity layer). ). Through the Identity layer, the depth features extracted by the deep neural network can be reduced to obtain the reduced depth features, which can be matched with the dimensions of expert features or unsupervised features, so as to solve the above problems and improve the integration of features. Diversity, in order to achieve the purpose of improving the accuracy of automatic detection of atrial fibrillation.
例如,Identity层中的Identity激活函数则可以在输出分类结果之前输出提取的深度特征(例如,降维后的深度特征)。For example, the Identity activation function in the Identity layer can output the extracted depth features (for example, the depth features after dimensionality reduction) before outputting the classification results.
例如,该深度神经网络中的参数可以在图2中的训练阶段S1中训练得到。For example, the parameters in the deep neural network can be obtained by training in the training stage S1 in FIG. 2.
例如,在训练阶段S1,该深度神经网络可以连接到分类器。例如,该分类器为Softmax分类器或SVM(支持向量机,Support Vector Machine)分类器等,在本公开的实施例中以Softmax分类器为例进行说明,本公开的实施例对此不作限制。分类器可以根据提取的特征对输入深度神经网络的输入数据进行分类。分类器的分类结果经过输出层输出以作为该深度神经网络模型的最终输出。For example, in the training phase S1, the deep neural network can be connected to the classifier. For example, the classifier is a Softmax classifier or an SVM (Support Vector Machine) classifier, etc. In the embodiments of the present disclosure, the Softmax classifier is taken as an example for description, which is not limited in the embodiments of the present disclosure. The classifier can classify the input data of the input deep neural network according to the extracted features. The classification result of the classifier is output through the output layer as the final output of the deep neural network model.
图2为本公开至少一实施例提供的一种深度特征的提取示意图。如图2所示,为了对任何架构的深度神经网络进行改造,使其能够自动提取深度特征,该深度特征的提取算法包括两个阶段:训练阶段和提取阶段。FIG. 2 is a schematic diagram of extracting a depth feature provided by at least one embodiment of the present disclosure. As shown in Figure 2, in order to transform the deep neural network of any architecture so that it can automatically extract deep features, the deep feature extraction algorithm includes two phases: training phase and extraction phase.
如图2所示,在训练阶段S1,首先基于某个具体的任务(例如,判断是否为房颤)训练一个深度神经网络,并保存该深度神经网络的架构和权值。As shown in Fig. 2, in the training stage S1, a deep neural network is first trained based on a specific task (for example, judging whether it is atrial fibrillation), and the architecture and weights of the deep neural network are saved.
在训练阶段S1,向该深度神经网络输入有标签(例如,标签包括是否为房颤)的数据,该深度神经网络可以输出深度特征,并将该深度特征输出至Softmax层,以在Softmax层(即,Softmax分类器所在的层)基于该深度神经网络提取的深度特征输出分类结果(例如,该待识别图像属于预设类别(例如,房颤)的预测概率),确定与预测结果对应的标签,即是否为房颤。In the training stage S1, data with labels (for example, whether the label includes atrial fibrillation) is input to the deep neural network, the deep neural network can output deep features, and output the deep features to the Softmax layer, so as to be in the Softmax layer ( That is, the layer where the Softmax classifier is located) outputs the classification result based on the depth features extracted by the deep neural network (for example, the predicted probability that the image to be recognized belongs to a preset category (for example, atrial fibrillation)), and determines the label corresponding to the prediction result , That is, whether it is atrial fibrillation.
例如,在训练过程中,该深度神经网络的训练过程中还可以包括优化器,优化器中的优化函数可以根据系统损失函数计算得到的系统损失值以计算该深度神经网络的参数的误差值,并根据该误差值对待训练的深度神经网络的参数进行修正,从而可以使得深度神 经网络输出较准确的深度特征。例如,优化函数可以采用随机梯度下降(stochastic gradient descent,SGD)算法、批量梯度下降(batch gradient descent,BGD)算法等计算该深度神经网络的参数的误差值。For example, in the training process, the training process of the deep neural network may also include an optimizer, and the optimization function in the optimizer may calculate the error value of the parameters of the deep neural network according to the system loss value calculated by the system loss function, And according to the error value, the parameters of the deep neural network to be trained are corrected, so that the deep neural network can output more accurate depth features. For example, the optimization function may use a stochastic gradient descent (SGD) algorithm, a batch gradient descent (BGD) algorithm, etc., to calculate the error value of the parameters of the deep neural network.
例如,Softmax层是用于输出分类结果(例如,判断是否为房颤)的回归函数层。For example, the Softmax layer is a regression function layer for outputting classification results (for example, determining whether it is atrial fibrillation).
因此,在提取阶段S2,将与深度神经网络连接的最后一层Softmax层替代为Identity层(例如,输出层,采用Identity激活函数),并且保持该深度神经网络中其它的全连接层的网络结构和权值不变,即利用训练好的深度神经网络提取深度特征,这样可以得到精度较高的深度特征,并在Identity层对其进行降维以输出与专家特征或无监督特征匹配的的深度特征。此时,向该深度神经网络输入新数据,就可以从Identity层输出深度特征,从而可以实现深度特征的表示与提取。Therefore, in the extraction stage S2, the last Softmax layer connected to the deep neural network is replaced by the Identity layer (for example, the output layer, using the Identity activation function), and the network structure of the other fully connected layers in the deep neural network is maintained The sum weight is unchanged, that is, the deep features are extracted by the trained deep neural network, which can obtain the high-precision deep features, and reduce the dimensionality in the Identity layer to output the depth that matches the expert features or unsupervised features feature. At this time, by inputting new data to the deep neural network, deep features can be output from the Identity layer, so that the representation and extraction of deep features can be realized.
在该示例中,该深度特征的提取可以不依赖于深度神经网络的特殊架构,即,可以采用能够实现特征提取的任意网络架构而不限于一种网络。In this example, the deep feature extraction may not rely on the special architecture of the deep neural network, that is, any network architecture that can realize feature extraction can be used and is not limited to one type of network.
需要注意是,上述深度特征提取的方法不限于上述神经网络,还可以通过例如HOG+SVM等本领域内的常规方法实现,本公开的实施例对此不作限制。It should be noted that the above-mentioned deep feature extraction method is not limited to the above-mentioned neural network, and can also be implemented by conventional methods in the art, such as HOG+SVM, which is not limited in the embodiments of the present disclosure.
另外,由于深度学习技术在增加了表征学习能力的同时也减少了可解释性能力。它们往往被当成黑盒模型来使用,即,仅能获得具体的输出结果(例如,是否为房颤),而无法有效解释在中间过程中产生的深度特征等数据,例如,无法解释哪一个深度特征对应哪个区域的房颤等。但在医疗领域,针对一个分析识别结果进行适当的解释对于医生与患者都是极其重要的,而目前的基于深度学习的房颤识别技术所提取的深度特征无法有效指导和辅助医生的诊断决策。针对该技术问题,本公开实施例中可以通过对深度特征进行相应地标注或结合具有可解释性的专家特征等方法来实现,从而提升房颤识别的精度。In addition, because the deep learning technology increases the representation learning ability, it also reduces the interpretability ability. They are often used as black box models, that is, they can only obtain specific output results (for example, whether it is atrial fibrillation), and cannot effectively explain the depth characteristics and other data generated in the intermediate process, for example, cannot explain which depth Which area of the feature corresponds to atrial fibrillation, etc. However, in the medical field, an appropriate interpretation of an analysis and recognition result is extremely important for both doctors and patients, and the current deep learning-based atrial fibrillation recognition technology extracts deep features that cannot effectively guide and assist doctors in diagnosis decisions. In view of this technical problem, the embodiments of the present disclosure can be implemented by correspondingly annotating depth features or combining interpretable expert features, etc., so as to improve the accuracy of atrial fibrillation recognition.
对于步骤S120,例如,在一些示例中,该步骤S120包括:基于根据医学图像数据获得的经验公式、规则和特征值,提取待识别图像的专家特征。例如,该医学图像数据可以包括心电图数据。For step S120, for example, in some examples, step S120 includes: extracting expert features of the image to be recognized based on empirical formulas, rules, and feature values obtained from medical image data. For example, the medical image data may include electrocardiogram data.
例如,在基于领域知识的房颤自动检测场景中,领域专家会根据自己的知识体系,基于数据总结出“经验公式”、“规则”和“特征值”等,并基于它们进行房颤检测与识别,实际上,这些经验公式、规则和特征值,都可以被表示为数值向量C∈R
d(其中,R
d表示维数为d的取值空间,该公式表示数值向量C属于该取值空间)的形式,进而可以被视为专家特征。
For example, in the scene of automatic detection of atrial fibrillation based on domain knowledge, domain experts will sum up "empirical formulas", "rules", and "eigenvalues" based on data based on their own knowledge system, and perform AF detection based on them. Recognition, in fact, these empirical formulas, rules, and eigenvalues can all be expressed as a numeric vector C∈R d (where R d represents a value space with dimension d, and this formula indicates that the numeric vector C belongs to this value The form of space) can then be regarded as an expert feature.
例如,该专家特征的类别可以包括统计、形态、时域和频域的至少之一,当然还可以包括其他的类别,本公开的实施例对此不作限制。For example, the category of the expert feature may include at least one of statistics, morphology, time domain, and frequency domain, and of course may also include other categories, which are not limited in the embodiments of the present disclosure.
具体地,统计类专家特征包括例如均值(Mean)、最大值(Maximum)、最小值(Minimum)、方差(Variance)、偏度(Skewness)、峰度(Kurtosis)、分位数(Percentile)和阈值(Threshold)等统计数值。例如,在心电监护时序数据上如果最大值差异较大,提示病人可能患有房颤。Specifically, statistical expert features include, for example, Mean, Maximum, Minimum, Variance, Skewness, Kurtosis, Percentile, and Statistical values such as Threshold. For example, if the maximum difference in the ECG monitoring time series data is large, it indicates that the patient may have atrial fibrillation.
例如,形态的定义与具体领域有直接关系,例如,在房颤检测领域中,在心电监护时序数据出现P波段消失的形态时,则提示病人可能患有房颤。For example, the definition of morphology is directly related to specific areas. For example, in the field of atrial fibrillation detection, when the P-band disappears in the timing data of the ECG monitoring, it indicates that the patient may have atrial fibrillation.
由于很多时序数据具有一定的周期性,时域的分析注重时序数据在时间维度上的节律特性,例如,心电监护时序数据的RR间期如果不规则,提示病人可能患有房颤。Since many time series data have certain periodicity, time-domain analysis pays attention to the rhythmic characteristics of time series data in the time dimension. For example, if the RR interval of the time series data of ECG monitoring is irregular, it indicates that the patient may have atrial fibrillation.
例如,时序数据在频域上也会展现出一定的特性,例如,在心电监护时序数据中,如果大于50赫兹(Hz)的能量过高,则提示病人身体的肌肉电流较强,即在心电监护时序数据中讯在肌电干扰等噪声,会对分析结果产生影响,此时,房颤识别的精度不高。For example, time series data will also show certain characteristics in the frequency domain. For example, in the time series data of ECG monitoring, if the energy greater than 50 Hz is too high, it indicates that the muscle current of the patient’s body is strong, that is, in the ECG monitoring time series data. In the monitoring sequence data, noises such as electromyographic interference will affect the analysis results. At this time, the accuracy of atrial fibrillation recognition is not high.
对于步骤S130,在一些示例中,该步骤S130可以包括:拼接深度特征和专家特征,以获得融合特征。For step S130, in some examples, the step S130 may include: stitching depth features and expert features to obtain fusion features.
例如,在一些示例中,为了融合不同来源的特征,使用全局池化(Max Pooling)和均值池化(Mean Pooling)操作。For example, in some examples, in order to integrate features from different sources, global pooling (Max Pooling) and mean pooling (Mean Pooling) operations are used.
图3A为本公开至少一实施例提供的一种融合操作的流程图。也就是说,图3A为图1B中所示的步骤S130的至少一个示例的流程图。例如,在图3A所示的示例中,该融合操作包括步骤S1311至步骤S1312。图3B为本公开至少一实施例提供的一种融合操作的示意图。下面,结合图3A和图3B对本公开至少一实施例提供的融合操作进行详细地介绍。FIG. 3A is a flowchart of a fusion operation provided by at least one embodiment of the present disclosure. That is, FIG. 3A is a flowchart of at least one example of step S130 shown in FIG. 1B. For example, in the example shown in FIG. 3A, the fusion operation includes step S1311 to step S1312. FIG. 3B is a schematic diagram of a fusion operation provided by at least one embodiment of the present disclosure. Hereinafter, the fusion operation provided by at least one embodiment of the present disclosure will be described in detail with reference to FIG. 3A and FIG. 3B.
步骤S1311:分别对深度特征和专家特征进行全局池化操作和均值池化操作,以分别获取深度特征的全局向量和均值向量以及专家特征的全局向量和均值向量。Step S1311: Perform a global pooling operation and an average pooling operation on the depth feature and the expert feature respectively to obtain the global vector and the mean vector of the depth feature and the global vector and the mean vector of the expert feature respectively.
例如,如图3B所示,深度特征提取器110提取的深度特征和专家特征提取器提取的专家特征分别表示为向量矩阵V=[v
0,...,v
i,...,v
n-1],其中,v
i∈R
d,表示向量v
i在该取值空间R
d内,V∈R
n×d,表示向量矩阵V在该取值空间R
n×d内,其中,R
n×d表示维数为n*d的取值空间。
For example, as shown in FIG expert feature depth wherein the depth of the feature extractor 110 extracts and 3B expert feature extractor extracts were represented as a vector matrix V = [v 0, ..., v i, ..., v n -1 ], where, v i ∈ R d , means that the vector v i is in the value space R d , and V ∈ R n×d , means that the vector matrix V is in the value space R n×d , where R n×d represents the value space of dimension n*d.
例如,对深度特征和专家特征分别进行全局池化操作(例如,最大池化)和均值池化操作,以分别获取深度特征的全局向量v
max和均值向量v
mean以及专家特征的全局向量v
max和均值向量v
mean,其中,v
max∈R
d和v
mean∈R
d。例如,全局向量v
max是在向量矩阵V中例如按列计算每列向量的最大值得到的,均值向量v
mean是在向量矩阵V中例如按列计算每列向量的平均值得到的。
For example, the depth of features and specialist features global pool operation (e.g., maximum pooling) and mean cell operations, respectively, to obtain the global vector v max and mean vectors v mean and the specialist feature depth features are global vector v max And the mean vector v mean , where v max ∈ R d and v mean ∈ R d . For example, the global vector v max is obtained by calculating the maximum value of each column vector in the vector matrix V, for example, and the mean vector v mean is obtained by calculating the average value of each column vector in the vector matrix V, for example, by column.
步骤S1312:拼接深度特征的全局向量和均值向量的至少之一以及专家特征的全局向量和均值向量的至少之一,以获得融合特征。Step S1312: Splice at least one of the global vector and the mean vector of the depth feature and at least one of the global vector and the mean vector of the expert feature to obtain a fusion feature.
例如,拼接所有的不同来源的特征的全局向量和均值向量,例如,拼接来源于深度特征提取器110的深度特征的全局向量和均值向量和以及来源于专家特征提取器120的专家特征的全局向量v
max和均值向量v
mean,以获得最终的融合特征,为房颤检测任务提供易用且准确的多源融合特征,从而提供准确的房颤检测的辅助方法。
For example, concatenate the global vector and the mean vector of all features from different sources, for example, concatenate the global vector and the mean vector of the depth feature from the deep feature extractor 110 and the global vector of the expert feature from the expert feature extractor 120 v max and the mean vector v mean to obtain the final fusion feature, provide easy-to-use and accurate multi-source fusion features for the atrial fibrillation detection task, thereby providing an accurate auxiliary method for atrial fibrillation detection.
例如,在本公开的实施例中,拼接操作为将不同来源的特征的全局向量和均值向量合并为1个向量,例如,1个1维向量。For example, in the embodiment of the present disclosure, the splicing operation is to merge the global vector and the mean vector of features from different sources into one vector, for example, one one-dimensional vector.
需要注意的是,还可以仅拼接深度特征的全局向量和专家特征的全局向量,或仅拼接深度特征的均值向量和专家特征的均值向量,或仅拼接深度特征的全局向量和专家特征的均值向量,或仅拼接深度特征的均值向量和专家特征的全局向量以获得融合特征,本公开的实施例对此不作限制。It should be noted that it is also possible to splice only the global vector of the depth feature and the global vector of the expert feature, or only the mean vector of the depth feature and the mean vector of the expert feature, or only the global vector of the depth feature and the mean vector of the expert feature. , Or only concatenate the mean vector of the depth feature and the global vector of the expert feature to obtain the fusion feature, which is not limited in the embodiment of the present disclosure.
例如,在另一些示例中,可以直接对不同来源的特征进行拼接。For example, in other examples, features from different sources can be spliced directly.
例如,在一些示例中,深度特征和专家特征可以具有多个通道。例如,深度特征可以是尺寸为H1*W1*C1的张量,其中H1可以是深度特征在第一方向(例如长度方向)上的尺寸,W1可以是深度特征在第二方向(例如宽度方向)上的尺寸,H1、W1可以是以像素数量为单位的尺寸,C1可以是深度特征的通道数。专家特征可以是尺寸为H2*W2*C2的张量,其中H2可以是专家特征在第一方向(例如长度方向)上的尺寸,W2可以是专家特征在第二方向(例如宽度方向)上的尺寸,H2、W2可以是以像素数量为单位的尺寸,C2可以是专家特征的通道数。这里,C1、C2是大于1的整数。For example, in some examples, the depth feature and the expert feature may have multiple channels. For example, the depth feature can be a tensor of size H1*W1*C1, where H1 can be the size of the depth feature in the first direction (for example, the length direction), and W1 can be the depth feature in the second direction (for example, the width direction) The above size, H1 and W1 can be the size in units of the number of pixels, and C1 can be the number of channels of the depth feature. The expert feature can be a tensor of size H2*W2*C2, where H2 can be the size of the expert feature in the first direction (for example, the length direction), and W2 can be the expert feature in the second direction (for example, the width direction) Size, H2 and W2 can be the size in units of the number of pixels, and C2 can be the number of channels of expert features. Here, C1 and C2 are integers greater than 1.
例如,深度特征可以具有100个通道,例如,每个通道中的深度特征为一个1行*M列(M为大于1的整数)的一维向量,该100个通道的深度特征可以组合为一个100行*M列的向量矩阵。例如,专家特征也可以具有100个通道,例如,该100个通道的专家特征可以组合为一个100行*M列的向量矩阵。通过拼接深度特征和专家特征可以得到一个200个通道的融合特征,例如,该200个通道的融合特征为一个200行*M列的向量矩阵。该具有200个通道的融合特征融合有深度特征和专家特征的信息。For example, the depth feature can have 100 channels. For example, the depth feature in each channel is a one-dimensional vector of 1 row*M column (M is an integer greater than 1), and the depth features of the 100 channels can be combined into one A vector matrix of 100 rows*M columns. For example, the expert features may also have 100 channels. For example, the expert features of the 100 channels may be combined into a vector matrix with 100 rows*M columns. A fusion feature of 200 channels can be obtained by splicing depth features and expert features. For example, the fusion feature of 200 channels is a vector matrix of 200 rows*M columns. The fusion feature with 200 channels integrates the information of depth feature and expert feature.
需要注意的是,上述深度特征和专家特征的通道数量仅是示例性的,可以根据具体实施例设置,本公开的实施例对此不作限制。It should be noted that the number of channels of the above-mentioned depth feature and expert feature are only exemplary, and can be set according to specific embodiments, which are not limited in the embodiments of the present disclosure.
例如,可以提供融合处理器,并通过该融合处理器根据融合深度特征以及专家特征,以获得待识别图像的融合特征;例如,也可以通过中央处理单元(CPU)、图像处理器(GPU)、张量处理器(TPU)、现场可编程逻辑门阵列(FPGA)或者具有数据处理能力和/或指令执行能力的其它形式的处理单元以及相应计算机指令来实现该融合处理器。例如,该处理单元可以为通用处理器或专用处理器,可以是基于X86或ARM架构的处理器等。For example, a fusion processor can be provided, and the fusion processor can obtain the fusion feature of the image to be recognized according to the fusion depth feature and the expert feature; for example, the central processing unit (CPU), image processor (GPU), A tensor processor (TPU), a field programmable logic gate array (FPGA), or other forms of processing units with data processing capabilities and/or instruction execution capabilities and corresponding computer instructions implement the fusion processor. For example, the processing unit may be a general-purpose processor or a special-purpose processor, and may be a processor based on the X86 or ARM architecture.
在本公开的实施例中,通过上述融合操作,使得融合特征融合有深度特征和专家特征的信息,可以将深度神经网络与领域知识结合,以克服深度神经网络的由于缺乏样本数据而导致的训练量不足的问题以及由于缺乏领域知识房颤识别精度低而很难被实际领域接受的问题,还可以克服面对实际场景下的低数据质量问题时,基于领域知识的房颤检测方法的精确度低等问题,从而可以提升房颤自动检测的精度。In the embodiment of the present disclosure, through the above fusion operation, the fusion feature is fused with the information of the depth feature and the expert feature, and the deep neural network can be combined with domain knowledge to overcome the training of the deep neural network due to lack of sample data. The problem of insufficient quantity and the problem that atrial fibrillation is difficult to be accepted by the actual field due to lack of domain knowledge and low accuracy in recognition of atrial fibrillation. It can also overcome the problem of low data quality in actual scenarios. The accuracy of the detection method of atrial fibrillation based on domain knowledge Low-level problems, which can improve the accuracy of automatic detection of atrial fibrillation.
具体地,在本公开的实施例中,基于深度神经网络的房颤识别方法对噪声的自适应能力比较好,由于深度特征是自动提取而非人工提取,因此可以克服与基于领域知识的方法很难在各类心律失常混杂的复杂情况下都能准确区分出房颤和其他心律失常类型的问题。但是,基于深度神经网络的房颤识别方法需要大规模的有良好标注的学习样本,这是非常 困难的,因此,在本公开的实施例中将对良好标注样本量依赖性不强的专家特征(基于领域知识或特征工程的房颤识别方法)纳入建模过程中,使得深度特征和专家特征融合,从而能够在有良好标注的学习样本量不足的情况下依然能够提供较可靠的房颤预测精度。Specifically, in the embodiments of the present disclosure, the atrial fibrillation recognition method based on the deep neural network has better adaptive ability to noise. Since the depth features are automatically extracted rather than manually extracted, it can overcome the difficulty of the method based on domain knowledge. It is difficult to accurately distinguish between atrial fibrillation and other types of arrhythmia in the complex situation where various arrhythmia are mixed. However, the deep neural network-based atrial fibrillation recognition method requires large-scale well-annotated learning samples, which is very difficult. Therefore, in the embodiments of the present disclosure, expert features that are not strongly dependent on the amount of well-annotated samples will be used. (Atrial fibrillation recognition method based on domain knowledge or feature engineering) is incorporated into the modeling process to integrate deep features and expert features, so that it can still provide a more reliable prediction of atrial fibrillation even when the number of well-labeled learning samples is insufficient Accuracy.
图4为本公开至少一实施例提供的另一种图像处理方法的流程图。如图4所示,在图1所示的示例的基础上,该图像处理方法还包括步骤S150。FIG. 4 is a flowchart of another image processing method provided by at least one embodiment of the present disclosure. As shown in FIG. 4, based on the example shown in FIG. 1, the image processing method further includes step S150.
步骤S150:基于无监督特征提取器获取待识别图像的无监督特征。Step S150: Obtain unsupervised features of the image to be recognized based on the unsupervised feature extractor.
例如,在该示例中,在实际应用中获取标记数据比较困难的场景下,基于无监督特征提取器获取待识别图像的无监督特征之前,利用主成分分析法、随机投影法以及序列自动编码器中的至少之一无监督学习方法训练无监督特征提取器,学习无监督特征的自动表示与提取,将变换的低维数据提取为无监督特征。For example, in this example, in a scene where it is difficult to obtain labeled data in practical applications, before obtaining the unsupervised features of the image to be recognized based on the unsupervised feature extractor, the principal component analysis method, random projection method and sequence autoencoder are used At least one of the unsupervised learning methods trains an unsupervised feature extractor, learns automatic representation and extraction of unsupervised features, and extracts transformed low-dimensional data as unsupervised features.
与在标记数据上构建模型的有监督学习方法不同,无监督学习方法仅在未标记数据上构建模型。虽然由于缺乏标记数据,无监督学习到的特征不如有监督学习特征有效,但是,由于在实际应用中,获取标记数据可能比较困难,因此无监督特征的自动学习方法在这种场景下发挥了重要的作用。Unlike supervised learning methods that build models on labeled data, unsupervised learning methods only build models on unlabeled data. Although due to the lack of labeled data, unsupervised learning features are not as effective as supervised learning features, but in practical applications, it may be difficult to obtain labeled data, so automatic learning methods for unsupervised features play an important role in this scenario. The role of.
例如,可以利用主成分分析法、随机投影法或序列自动编码器训练得到无监督特征提取器。For example, the unsupervised feature extractor can be obtained by using principal component analysis method, random projection method or sequential autoencoder training.
例如,主成分分析(Principle Component Analysis,PCA)法使用正交变换将相关变量转换为线性不相关的主成分,这些主成分的数量远远少于原始变量,同时也提供了更多信息,然后将变换的低维主成分提取为无监督特征。For example, the principal component analysis (Principle Component Analysis, PCA) method uses orthogonal transformation to convert related variables into linear and uncorrelated principal components. The number of these principal components is far less than the original variables, and it also provides more information. Extract the transformed low-dimensional principal components as unsupervised features.
例如,随机投影(Random Projection,RP)法通过将原始数据乘以随机投影矩阵,将高维数据投影到较低维度以得到低维数据,然后将变换的低维数据提取为无监督特征。该RP算法简单且计算效率高。For example, the Random Projection (RP) method multiplies the original data by a random projection matrix, projects high-dimensional data to lower dimensions to obtain low-dimensional data, and then extracts the transformed low-dimensional data as unsupervised features. The RP algorithm is simple and computationally efficient.
例如,序列自动编码器(Sequence to Sequence Autoencoder,SeqAE)是自动编码器(AE)的变体,它将编码器和解码器中的全连接层替换为循环层。序列自动编码器首先将时序数据变换为隐藏层的表示,然后解码器将隐藏层的表示再次变换为时序数据,并尝试最小化原始序列与解码序列之间的距离。最后,将隐藏层的表示提取为无监督特征。For example, Sequence to Sequence Autoencoder (SeqAE) is a variant of Autoencoder (AE), which replaces the fully connected layer in the encoder and decoder with a cyclic layer. The sequence auto-encoder first transforms the time series data into the representation of the hidden layer, and then the decoder transforms the representation of the hidden layer into the time series data again, and tries to minimize the distance between the original sequence and the decoded sequence. Finally, the representation of the hidden layer is extracted as unsupervised features.
在该示例中,由于增加了无监督特征的获取,从而步骤S130可以相应地表示为:In this example, since the acquisition of unsupervised features is added, step S130 can be expressed as follows:
步骤S130:融合深度特征、专家特征以及无监督特征,以获得待识别图像的融合特征。Step S130: Fusion of depth features, expert features, and unsupervised features to obtain fusion features of the image to be recognized.
例如,在该示例中,如图5所示,将不同来源的特征向量进行融合,例如,可以通过拼接深度特征、专家特征以及无监督特征,以获得融合特征。For example, in this example, as shown in FIG. 5, feature vectors from different sources are fused, for example, depth features, expert features, and unsupervised features can be spliced to obtain fusion features.
例如,可以分别对深度特征、专家特征与无监督特征进行全局池化操作和均值池化操作,以分别获取深度特征的全局向量和均值向量、专家特征的全局向量和均值向量以及无监督特征的全局向量和均值向量;拼接深度特征的全局向量和均值向量的至少之一、专家特征的全局向量和均值向量的至少之一以及无监督特征的全局向量和均值向量的至少之 一,以获得融合特征。For example, you can perform global pooling operations and mean pooling operations on depth features, expert features, and unsupervised features, respectively, to obtain the global vector and mean vector of depth features, the global vector and mean vector of expert features, and unsupervised features. Global vector and mean vector; concatenate at least one of the global vector and the mean vector of the depth feature, at least one of the global vector and the mean vector of the expert feature, and at least one of the global vector and the mean vector of the unsupervised feature to obtain a fusion feature.
例如,具体的融合方法可参考图3B所示的融合过程的介绍,在此不再赘述。For example, the specific fusion method can refer to the introduction of the fusion process shown in FIG. 3B, which will not be repeated here.
需要注意的是,该步骤S130可以不限于上述特征的融合,还可以包括其余更多特征的融合,本公开的实施例对此不作限制。It should be noted that this step S130 may not be limited to the fusion of the above-mentioned features, and may also include the fusion of more other features, which is not limited in the embodiment of the present disclosure.
在本公开实施例中,融合特征融合了深度特征、专家特征以及无监督特征,实现数据降维,可以克服深度特征和专家特征中的噪声污染严重的问题,从而可以进一步提升房颤自动检测的精度。In the embodiments of the present disclosure, the fusion feature combines the depth feature, the expert feature, and the unsupervised feature to achieve data dimensionality reduction, which can overcome the serious noise pollution problem in the depth feature and the expert feature, thereby further improving the automatic detection of atrial fibrillation. Accuracy.
具体地,传统的基于领域知识的方法虽然可以解决深度神经网络对局有良好标注的样本量的依赖性强的问题,但是对噪声敏感,而无监督学习的方法虽然识别精确度较低,但具有自动降噪和不要标签的作用,本公开实施例提供的图形处理方法同时融合了多种不同技术(深度神经网络、基于领域知识以及无监督学习)的优势,在有良好标注的学习样本量不足的情况下,针对有噪声的心电图数据,依然能够提供较可靠的房颤预测精度。Specifically, although the traditional method based on domain knowledge can solve the problem of deep neural network's strong dependence on the sample size of the local well-labeled, but it is sensitive to noise, while the unsupervised learning method has low recognition accuracy, but With the functions of automatic noise reduction and no labeling, the graphics processing method provided by the embodiments of the present disclosure combines the advantages of multiple different technologies (deep neural network, domain-based knowledge, and unsupervised learning) at the same time. In the case of insufficient, for noisy ECG data, it can still provide more reliable atrial fibrillation prediction accuracy.
需要注意的是,还可融合其他的技术,本公开的实施例对此不作限制。It should be noted that other technologies can also be integrated, which is not limited in the embodiments of the present disclosure.
对于步骤S140,在一些示例中,根据待识别图像的融合特征对待识别图像进行分类,包括:根据待识别图像的融合特征判断待识别图像是否包括房颤特征。For step S140, in some examples, classifying the image to be recognized according to the fusion feature of the image to be recognized includes: judging whether the image to be recognized includes atrial fibrillation features according to the fusion feature of the image to be recognized.
例如,当将上述图像处理方法应于与其他领域,例如,机械领域时,还可以用于检测周期性波形的变化。本公开的实施例对此不作限制。For example, when the above-mentioned image processing method is applied to other fields, such as the mechanical field, it can also be used to detect changes in periodic waveforms. The embodiment of the present disclosure does not limit this.
例如,将上述实施例中获取的融合特征(例如,图3B中获取的融合特征或图5中获取的融合特征)输入至房颤检测器中,以实现房颤识别检测。例如,该房颤检测器可以是神经网络分类器或SVM分类器等,本公开的实施例对此不作限制。For example, the fusion feature acquired in the above-mentioned embodiment (for example, the fusion feature acquired in FIG. 3B or the fusion feature acquired in FIG. 5) is input into the atrial fibrillation detector to realize atrial fibrillation recognition and detection. For example, the atrial fibrillation detector may be a neural network classifier or an SVM classifier, etc., which is not limited in the embodiment of the present disclosure.
例如,可以提供分类处理器,并通过该分类处理器根据待识别图像的融合特征对待识别图像进行分类;例如,也可以通过中央处理单元(CPU)、图像处理器(GPU)、张量处理器(TPU)、现场可编程逻辑门阵列(FPGA)或者具有数据处理能力和/或指令执行能力的其它形式的处理单元以及相应计算机指令来实现该分类处理器。For example, a classification processor can be provided, and the classification processor can classify the image to be recognized according to the fusion characteristics of the image to be recognized; for example, the central processing unit (CPU), image processor (GPU), or tensor processor can also be used to classify the image to be recognized. (TPU), Field Programmable Logic Gate Array (FPGA), or other forms of processing units with data processing capabilities and/or instruction execution capabilities and corresponding computer instructions to implement the classification processor.
需要说明的是,在本公开的实施例中,该图像处理方法的流程可以包括更多或更少的操作,这些操作可以顺序执行或并行执行。虽然上文描述的图像处理方法的流程包括特定顺序出现的多个操作,但是应该清楚地了解,多个操作的顺序并不受限制。上文描述的图像处理方法可以执行一次,也可以按照预定条件执行多次。It should be noted that, in the embodiments of the present disclosure, the flow of the image processing method may include more or fewer operations, and these operations may be executed sequentially or in parallel. Although the flow of the image processing method described above includes multiple operations appearing in a specific order, it should be clearly understood that the order of the multiple operations is not limited. The image processing method described above may be executed once, or may be executed multiple times according to predetermined conditions.
本公开上述实施例提供的图像处理方法,通过基于领域知识的专家特征的表示与提取,以及基于深度神经网络的深度特征的表示与提取,并采用统一的框架对专家特征和深度特征进行表示与融合,以达到提升房颤自动检测精度的目的,从而为实时、动态地房颤识别与诊断提供高精度的智能决策支持方法,帮助医生及时诊断和准确发现患者的房颤的发生,帮助病人及时了解病情的变化,从而提高医疗质量、降低心脏性猝死等危及生命情况的发生率,最终减少给家庭和社会带来的健康和经济负担。The image processing method provided by the foregoing embodiments of the present disclosure uses the representation and extraction of expert features based on domain knowledge and the representation and extraction of deep features based on deep neural networks, and uses a unified framework to represent and extract expert features and depth features. Fusion, in order to achieve the purpose of improving the accuracy of automatic detection of atrial fibrillation, so as to provide high-precision intelligent decision support methods for real-time and dynamic atrial fibrillation recognition and diagnosis, helping doctors diagnose and accurately discover the occurrence of atrial fibrillation in patients in time, and help patients in time Understand the changes in the disease, so as to improve the quality of medical care, reduce the incidence of life-threatening conditions such as sudden cardiac death, and ultimately reduce the health and economic burden brought to the family and society.
例如,上述图形处理方法可以通过图1B所示的图像处理系统实现。如图1B所示, 该图像处理系统10可以包括用户终端11、网络12、服务器13以及数据库14。For example, the foregoing graphics processing method can be implemented by the image processing system shown in FIG. 1B. As shown in FIG. 1B, the image processing system 10 may include a user terminal 11, a network 12, a server 13, and a database 14.
用户终端11可以是例如图1B中示出的电脑11-1、手机11-2。可以理解的是,用户终端11可以是能够执行数据处理的任何其他类型的电子设备,其可以包括但不限于台式电脑、笔记本电脑、平板电脑、智能手机、智能家居设备、可穿戴设备、车载电子设备、监控设备等。用户终端也可以是设置有电子设备的任何装备,例如车辆、机器人等。The user terminal 11 may be, for example, the computer 11-1 and the mobile phone 11-2 shown in FIG. 1B. It is understandable that the user terminal 11 may be any other type of electronic device capable of performing data processing, which may include, but is not limited to, a desktop computer, a notebook computer, a tablet computer, a smart phone, a smart home device, a wearable device, and in-vehicle electronics. Equipment, monitoring equipment, etc. The user terminal may also be any equipment provided with electronic equipment, such as vehicles, robots, and so on.
根据本公开实施例提供的用户终端可以用于接收待识别图像,并利用本公开实施例提供的方法实现图像识别和分类。例如,用户终端11可以通过用户终端11上设置的图像采集设备(图中未示出,例如照相机、摄像机等)采集待识别图像。又例如,用户终端11也可以从独立设置的图像采集设备接收待识别图像。再例如,用户终端11也可以经由网络从服务器13接收待识别图像。这里所述的待识别图像可以是单独的图像,也可以是视频中的一帧。在待识别图像是医学图像的情况下,用户终端也可以同医学采集设备接收待识别图像。The user terminal provided according to the embodiment of the present disclosure may be used to receive the image to be recognized, and use the method provided by the embodiment of the present disclosure to realize image recognition and classification. For example, the user terminal 11 may collect an image to be recognized through an image acquisition device (not shown in the figure, such as a camera, a video camera, etc.) provided on the user terminal 11. For another example, the user terminal 11 may also receive the image to be recognized from an independently set image acquisition device. For another example, the user terminal 11 may also receive the image to be recognized from the server 13 via the network. The image to be recognized can be a single image or a frame in the video. In the case that the image to be recognized is a medical image, the user terminal may also receive the image to be recognized with the medical acquisition device.
在一些实施例中,可以利用用户终端11的处理单元执行本公开实施例提供的图像处理方法。在一些实现方式中,用户终端11可以利用用户终端11内置的应用程序执行图像处理方法。在另一些实现方式中,用户终端11可以通过调用用户终端11外部存储的应用程序执行本公开至少一实施例提供的图像处理方法。In some embodiments, the processing unit of the user terminal 11 may be used to execute the image processing method provided in the embodiments of the present disclosure. In some implementation manners, the user terminal 11 may use a built-in application program of the user terminal 11 to execute the image processing method. In other implementation manners, the user terminal 11 may execute the image processing method provided by at least one embodiment of the present disclosure by calling an application program stored externally of the user terminal 11.
在另一些实施例中,用户终端11将接收的待识别图像经由网络12发送至服务器13,并由服务器13执行图像处理方法。在一些实现方式中,服务器13可以利用服务器内置的应用程序执行图像处理方法。在另一些实现方式中,服务器13可以通过调用服务器13外部存储的应用程序执行图像处理方法。In other embodiments, the user terminal 11 sends the received image to be recognized to the server 13 via the network 12, and the server 13 executes the image processing method. In some implementation manners, the server 13 may execute the image processing method by using an application program built in the server. In other implementation manners, the server 13 may execute the image processing method by calling an application program stored externally of the server 13.
网络12可以是单个网络,或至少两个不同网络的组合。例如,网络12可以包括但不限于局域网、广域网、公用网络、专用网络等中的一种或几种的组合。The network 12 may be a single network, or a combination of at least two different networks. For example, the network 12 may include, but is not limited to, one or a combination of several of a local area network, a wide area network, a public network, and a private network.
服务器13可以是一个单独的服务器,或一个服务器群组,群组内的各个服务器通过有线的或无线的网络进行连接。一个服务器群组可以是集中式的,例如数据中心,也可以是分布式的。服务器13可以是本地的或远程的。The server 13 may be a single server or a server group, and each server in the group is connected through a wired or wireless network. A server group can be centralized, such as a data center, or distributed. The server 13 may be local or remote.
数据库14可以泛指具有存储功能的设备。数据库13主要用于存储从用户终端11和服务器13工作中所利用、产生和输出的各种数据。数据库14可以是本地的,或远程的。数据库14可以包括各种存储器、例如随机存取存储器(Random Access Memory(RAM))、只读存储器(Read Only Memory(ROM))等。以上提及的存储设备只是列举了一些例子,该系统可以使用的存储设备并不局限于此。The database 14 can generally refer to a device with a storage function. The database 13 is mainly used to store various data used, generated, and output from the work of the user terminal 11 and the server 13. The database 14 can be local or remote. The database 14 may include various memories, such as Random Access Memory (RAM), Read Only Memory (ROM), and so on. The storage devices mentioned above are just a few examples, and the storage devices that can be used in the system are not limited to these.
数据库14可以经由网络12与服务器13或其一部分相互连接或通信,或直接与服务器13相互连接或通信,或是上述两种方式的结合。The database 14 may be connected or communicated with the server 13 or a part thereof via the network 12, or directly connected or communicated with the server 13 or a combination of the above two methods.
在一些实施例中,数据库15可以是独立的设备。在另一些实施例中,数据库15也可以集成在用户终端11和服务器14中的至少一个中。例如,数据库15可以设置在用户终端11上,也可以设置在服务器14上。又例如,数据库15也可以是分布式的,其一部分 设置在用户终端11上,另一部分设置在服务器14上。In some embodiments, the database 15 may be a stand-alone device. In other embodiments, the database 15 may also be integrated in at least one of the user terminal 11 and the server 14. For example, the database 15 may be set on the user terminal 11 or on the server 14. For another example, the database 15 may also be distributed, a part of which is set on the user terminal 11, and the other part is set on the server 14.
图6为本公开至少一实施例提供的一种图像处理装置的示意框图。例如,在图6所示的示例中,该图像处理装置100包括深度特征提取器110、专家特征提取器120、融合处理器130和分类处理器140。例如,这些特征提取器和处理器可以通过硬件(例如电路)模块或软件模块等实现,以下是实施例与此相同,不再赘述。例如,可以通过中央处理单元(CPU)、图像处理器(GPU)、张量处理器(TPU)、现场可编程逻辑门阵列(FPGA)或者具有数据处理能力和/或指令执行能力的其它形式的处理单元以及相应计算机指令来实现这些处理器或提取器。FIG. 6 is a schematic block diagram of an image processing apparatus provided by at least one embodiment of the present disclosure. For example, in the example shown in FIG. 6, the image processing device 100 includes a depth feature extractor 110, an expert feature extractor 120, a fusion processor 130 and a classification processor 140. For example, these feature extractors and processors can be implemented by hardware (for example, circuit) modules or software modules, etc. The following embodiments are the same as this, and will not be repeated. For example, a central processing unit (CPU), an image processor (GPU), a tensor processor (TPU), a field programmable logic gate array (FPGA), or other forms of data processing capabilities and/or instruction execution capabilities can be used. Processing units and corresponding computer instructions implement these processors or extractors.
深度特征提取器110配置为获取待识别图像的深度特征。例如,待识别图像为医学图像。例如,该深度特征提取器110可以实现步骤S110,其具体实现方法可以参考步骤S110的相关描述,在此不再赘述。The depth feature extractor 110 is configured to obtain the depth feature of the image to be recognized. For example, the image to be recognized is a medical image. For example, the depth feature extractor 110 can implement step S110, and its specific implementation method can refer to the related description of step S110, which will not be repeated here.
专家特征提取器120配置为获取待识别图像的专家特征。例如,该专家特征提取器120可以实现步骤S120,其具体实现方法可以参考步骤S120的相关描述,在此不再赘述。The expert feature extractor 120 is configured to obtain expert features of the image to be recognized. For example, the expert feature extractor 120 can implement step S120, and its specific implementation method can refer to the related description of step S120, which will not be repeated here.
融合处理器130配置为融合深度特征以及专家特征,以获得待识别图像的融合特征。例如,该融合处理器130可以实现步骤S130,其具体实现方法可以参考步骤S130的相关描述,在此不再赘述。The fusion processor 130 is configured to fuse the depth feature and the expert feature to obtain the fusion feature of the image to be recognized. For example, the fusion processor 130 may implement step S130, and its specific implementation method can refer to the related description of step S130, which will not be repeated here.
分类处理器140配置为根据待识别图像的融合特征对待识别图像进行分类。例如,该分类处理器140可以实现步骤S140,其具体实现方法可以参考步骤S140的相关描述,在此不再赘述。The classification processor 140 is configured to classify the image to be recognized according to the fusion feature of the image to be recognized. For example, the classification processor 140 can implement step S140, and the specific implementation method can refer to the related description of step S140, which will not be repeated here.
例如,深度特征提取器110还配置为利用深度神经网络获取待识别图像的深度特征。For example, the depth feature extractor 110 is further configured to obtain the depth feature of the image to be recognized by using a deep neural network.
例如,在本公开至少一实施例提供的图像处理装置中,专家特征提取器120还配置为基于根据医学图像数据获得的经验公式、规则和特征值,提取待识别图像的专家特征。For example, in the image processing apparatus provided by at least one embodiment of the present disclosure, the expert feature extractor 120 is further configured to extract the expert features of the image to be recognized based on empirical formulas, rules, and feature values obtained from medical image data.
例如,在本公开至少一实施例提供的图像处理装置中,专家特征的类别包括统计、形态、时域和频域中的至少之一。For example, in the image processing device provided by at least one embodiment of the present disclosure, the category of expert features includes at least one of statistics, morphology, time domain, and frequency domain.
例如,在本公开至少一实施例提供的图像处理装置中,分类处理器140还配置为:根据待识别图像的融合特征判断待识别图像是否包含房颤特征。For example, in the image processing device provided by at least one embodiment of the present disclosure, the classification processor 140 is further configured to determine whether the image to be identified contains atrial fibrillation features according to the fusion features of the image to be identified.
图7为本公开至少一实施例提供的另一种图像处理装置的示意框图。例如,如图7所示,在图6所示的示例的基础上,该图像处理装置100还包括无监督特征提取器150。FIG. 7 is a schematic block diagram of another image processing apparatus provided by at least one embodiment of the present disclosure. For example, as shown in FIG. 7, based on the example shown in FIG. 6, the image processing apparatus 100 further includes an unsupervised feature extractor 150.
例如,无监督特征提取器150配置为获取待识别图像的无监督特征。例如,该无监督特征提取器150可以实现步骤S150,其具体实现方法可以参考步骤S150的相关描述,在此不再赘述。For example, the unsupervised feature extractor 150 is configured to obtain unsupervised features of the image to be recognized. For example, the unsupervised feature extractor 150 can implement step S150, and its specific implementation method can refer to the related description of step S150, which will not be repeated here.
例如,在本公开至少一实施例提供的图像处理装置中,无监督特征提取器150还配置为在基于无监督特征提取器获取待识别图像的无监督特征之前,利用主成分分析法、随机投影法和序列自动编码器中的至少之一训练得到无监督特征提取器。For example, in the image processing device provided by at least one embodiment of the present disclosure, the unsupervised feature extractor 150 is further configured to use principal component analysis and random projection before acquiring the unsupervised features of the image to be recognized based on the unsupervised feature extractor. At least one of the method and sequence autoencoder is trained to obtain an unsupervised feature extractor.
例如,在该示例中,融合处理器130还配置为融合深度特征、专家特征以及无监督特 征,以获得待识别图像的融合特征。For example, in this example, the fusion processor 130 is also configured to fuse depth features, expert features, and unsupervised features to obtain fusion features of the image to be recognized.
例如,在本公开至少一实施例提供的图像处理装置中,融合处理器130还配置为:拼接深度特征、专家特征以及无监督特征,以获得融合特征。For example, in the image processing device provided by at least one embodiment of the present disclosure, the fusion processor 130 is further configured to splice depth features, expert features, and unsupervised features to obtain fusion features.
例如,在本公开至少一实施例提供的图像处理装置中,融合处理器130还配置为:分别对深度特征、专家特征与无监督特征进行全局池化操作和均值池化操作,以分别获取深度特征的全局向量和均值向量、专家特征的全局向量和均值向量以及无监督特征的全局向量和均值向量;拼接深度特征的全局向量和均值向量的至少之一、专家特征的全局向量和均值向量的至少之一以及无监督特征的全局向量和均值向量的至少之一,以获得融合特征。For example, in the image processing apparatus provided by at least one embodiment of the present disclosure, the fusion processor 130 is further configured to perform global pooling operations and average pooling operations on depth features, expert features, and unsupervised features, respectively, to obtain depths respectively. The global vector and mean vector of the feature, the global vector and the mean vector of the expert feature, the global vector and the mean vector of the unsupervised feature; at least one of the global vector and the mean vector of the spliced depth feature, the global vector and the mean vector of the expert feature At least one of and at least one of the global vector and the mean vector of the unsupervised feature to obtain the fusion feature.
需要注意的是,在本公开的实施例中,可以包括更多或更少的电路或单元,并且各个电路或单元之间的连接关系不受限制,可以根据实际需求而定。各个电路的具体构成方式不受限制,可以根据电路原理由模拟器件构成,也可以由数字芯片构成,或者以其他适用的方式构成。It should be noted that in the embodiments of the present disclosure, more or fewer circuits or units may be included, and the connection relationship between the respective circuits or units is not limited, and may be determined according to actual requirements. The specific structure of each circuit is not limited, and may be composed of analog devices according to the circuit principle, or may be composed of digital chips, or be composed in other suitable manners.
图8为本公开至少一实施例提供的又一种图像处理装置的示意框图。例如,如图8所示,该图像处理装置200包括处理器210、存储器220以及一个或多个计算机程序模块221。FIG. 8 is a schematic block diagram of another image processing apparatus provided by at least one embodiment of the present disclosure. For example, as shown in FIG. 8, the image processing apparatus 200 includes a processor 210, a memory 220, and one or more computer program modules 221.
例如,处理器210与存储器220通过总线系统230连接。例如,一个或多个计算机程序模块221被存储在存储器220中。例如,一个或多个计算机程序模块221包括用于执行本公开任一实施例提供的图像处理方法的指令。例如,一个或多个计算机程序模块221中的指令可以由处理器210执行。例如,总线系统230可以是常用的串行、并行通信总线等,本公开的实施例对此不作限制。For example, the processor 210 and the memory 220 are connected through a bus system 230. For example, one or more computer program modules 221 are stored in the memory 220. For example, one or more computer program modules 221 include instructions for executing the image processing method provided by any embodiment of the present disclosure. For example, instructions in one or more computer program modules 221 may be executed by the processor 210. For example, the bus system 230 may be a commonly used serial or parallel communication bus, etc., which is not limited in the embodiments of the present disclosure.
例如,该处理器210可以是中央处理单元(CPU)、图像处理器(GPU)或者具有数据处理能力和/或指令执行能力的其它形式的处理单元,可以为通用处理器或专用处理器,并且可以控制图像处理装置200中的其它组件以执行期望的功能。For example, the processor 210 may be a central processing unit (CPU), an image processing unit (GPU), or another form of processing unit with data processing capabilities and/or instruction execution capabilities, and may be a general-purpose processor or a special-purpose processor, and Other components in the image processing apparatus 200 can be controlled to perform desired functions.
存储器220可以包括一个或多个计算机程序产品,该计算机程序产品可以包括各种形式的计算机可读存储介质,例如易失性存储器和/或非易失性存储器。该易失性存储器例如可以包括随机存取存储器(RAM)和/或高速缓冲存储器(cache)等。该非易失性存储器例如可以包括只读存储器(ROM)、硬盘、闪存等。在计算机可读存储介质上可以存储一个或多个计算机程序指令,处理器210可以运行该程序指令,以实现本公开实施例中(由处理器210实现)的功能以及/或者其它期望的功能,例如图像处理方法等。在该计算机可读存储介质中还可以存储各种应用程序和各种数据,例如深度特征、专家特征以及应用程序使用和/或产生的各种数据等。The memory 220 may include one or more computer program products, and the computer program products may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory. The volatile memory may include random access memory (RAM) and/or cache memory (cache), for example. The non-volatile memory may include read-only memory (ROM), hard disk, flash memory, etc., for example. One or more computer program instructions may be stored on a computer-readable storage medium, and the processor 210 may run the program instructions to implement the functions (implemented by the processor 210) and/or other desired functions in the embodiments of the present disclosure, For example, image processing methods. Various application programs and various data, such as depth features, expert features, and various data used and/or generated by the application programs, can also be stored in the computer-readable storage medium.
需要说明的是,为表示清楚、简洁,本公开实施例并没有给出该图像处理装置200的全部组成单元。为实现图像处理装置200的必要功能,本领域技术人员可以根据具体需要提供、设置其他未示出的组成单元,本公开的实施例对此不作限制。It should be noted that, for the sake of clarity and conciseness, the embodiment of the present disclosure does not provide all the components of the image processing apparatus 200. In order to realize the necessary functions of the image processing apparatus 200, those skilled in the art can provide and set other unshown component units according to specific needs, and the embodiments of the present disclosure do not limit this.
关于不同实施例中的图像处理装置100和图像处理装置200的技术效果可以参考本公开的实施例中提供的图像处理方法的技术效果,这里不再赘述。Regarding the technical effects of the image processing device 100 and the image processing device 200 in different embodiments, reference may be made to the technical effects of the image processing method provided in the embodiments of the present disclosure, which will not be repeated here.
图像处理装置100和图像处理装置200可以用于各种适当的电子设备。图9为本公开至少一实施例提供的一种电子设备的示意图。例如,如图9所示,在一些示例中,电子设备300包括中央处理单元(CPU)301,其可以根据存储在只读存储器(ROM)302中的程序或者从存储装置308加载到随机访问存储器(RAM)303中的程序而执行各种适当的动作和处理。在RAM 303中,还存储有计算机系统操作所需的各种程序和数据。CPU 301、ROM302以及RAM303通过总线304被此相连。输入/输出(I/O)接口305也连接至总线304。The image processing apparatus 100 and the image processing apparatus 200 can be used in various appropriate electronic devices. FIG. 9 is a schematic diagram of an electronic device provided by at least one embodiment of the present disclosure. For example, as shown in FIG. 9, in some examples, the electronic device 300 includes a central processing unit (CPU) 301, which can be loaded to a random access memory according to a program stored in a read-only memory (ROM) 302 or from a storage device 308 (RAM) The program in 303 executes various appropriate actions and processing. In the RAM 303, various programs and data required for the operation of the computer system are also stored. The CPU 301, the ROM 302, and the RAM 303 are connected by this through the bus 304. An input/output (I/O) interface 305 is also connected to the bus 304.
以下部件连接至I/O接口305:包括键盘、鼠标等的输入装置306;包括诸如液晶显示器(LCD)等以及扬声器等的输出装置307;包括硬盘等的存储装置308;以及包括诸如LAN卡、调制解调器等的网络接口卡的通信装置309。通信装置309经由诸如因特网的网络执行通信处理。驱动器310也根据需要连接至I/O接口305。可拆卸介质311,诸如磁盘、光盘、磁光盘、半导体存储器等等,根据需要安装在驱动器310上,以便于从其上读出的计算机程序根据需要被安装入存储装置309。The following components are connected to the I/O interface 305: an input device 306 including a keyboard, a mouse, etc.; an output device 307 such as a liquid crystal display (LCD) and a speaker; a storage device 308 including a hard disk; and a storage device 308, such as a LAN card, A communication device 309 of a network interface card such as a modem. The communication device 309 performs communication processing via a network such as the Internet. The driver 310 is also connected to the I/O interface 305 as needed. A removable medium 311, such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, etc., is installed on the drive 310 as needed, so that the computer program read from it can be installed into the storage device 309 as needed.
例如,该电子设备300还可以进一步包括图像采集装置(图中未示出)和外设接口(图中未示出)等。例如,图像采集装置可以包括成像传感器以及镜头,该图像传感器可以为CMOS型或CCD型,镜头包括一个或多个透镜(凸透镜或凹透镜等)。该外设接口可以为各种类型的接口,例如为USB接口、闪电(lighting)接口等。该通信装置309可以通过无线通信来与网络和其他设备进行通信,该网络例如为因特网、内部网和/或诸如蜂窝电话网络之类的无线网络、无线局域网(LAN)和/或城域网(MAN)。无线通信可以使用多种通信标准、协议和技术中的任何一种,包括但不局限于全球移动通信系统(GSM)、增强型数据GSM环境(EDGE)、宽带码分多址(W-CDMA)、码分多址(CDMA)、时分多址(TDMA)、蓝牙、Wi-Fi(例如基于IEEE 802.11a、IEEE 802.11b、IEEE 802.11g和/或IEEE 802.11n标准)、基于因特网协议的语音传输(VoIP)、Wi-MAX,用于电子邮件、即时消息传递和/或短消息服务(SMS)的协议,或任何其他合适的通信协议。For example, the electronic device 300 may further include an image acquisition device (not shown in the figure), a peripheral interface (not shown in the figure), and the like. For example, the image acquisition device may include an imaging sensor and a lens, the image sensor may be of a CMOS type or a CCD type, and the lens may include one or more lenses (convex lens or concave lens, etc.). The peripheral interface can be various types of interfaces, such as a USB interface, a lightning interface, and the like. The communication device 309 can communicate with a network and other devices through wireless communication, such as the Internet, an intranet, and/or a wireless network such as a cellular telephone network, a wireless local area network (LAN), and/or a metropolitan area network ( MAN). Wireless communication can use any of a variety of communication standards, protocols and technologies, including but not limited to Global System for Mobile Communications (GSM), Enhanced Data GSM Environment (EDGE), Wideband Code Division Multiple Access (W-CDMA) , Code Division Multiple Access (CDMA), Time Division Multiple Access (TDMA), Bluetooth, Wi-Fi (e.g. based on IEEE 802.11a, IEEE 802.11b, IEEE 802.11g and/or IEEE 802.11n standards), voice transmission based on Internet protocol (VoIP), Wi-MAX, protocols used for e-mail, instant messaging and/or short message service (SMS), or any other suitable communication protocol.
例如,电子设备可以为手机、平板电脑、笔记本电脑、电子书、游戏机、电视机、数码相框、导航仪等任何设备,也可以为任意的电子设备及硬件的组合,本公开的实施例对此不作限制。For example, the electronic device can be any device such as a mobile phone, a tablet computer, a notebook computer, an e-book, a game console, a television, a digital photo frame, a navigator, etc., or can be any combination of electronic devices and hardware. This is not limited.
例如,该电子设备可以是医疗电子设备。图像采集装置可以用于采集待识别图像,例如,医学图像。这里所说的医学图像可以是例如通过CT、MRI、超声、X光、核素显像(如SPECT、PET)等方法采集的医学图像,也可以是例如心电图、脑电图、光学摄影等显示人体生理信息的图像。For example, the electronic device may be a medical electronic device. The image acquisition device may be used to acquire an image to be recognized, for example, a medical image. The medical images mentioned here can be, for example, medical images collected by CT, MRI, ultrasound, X-ray, radionuclide imaging (such as SPECT, PET), etc., or can be displays such as electrocardiogram, electroencephalogram, optical photography, etc. Images of human body physiological information.
例如,在一些示例中,该医疗电子设备可以是CT、MRI、超声、X光仪器等任何医学成像设备。图像采集装置可以实现为上述医学成像设备的成像单元,图像处理装置 100/200可以通过医学成像设备的内部处理单元(例如处理器)实现。For example, in some examples, the medical electronic equipment may be any medical imaging equipment such as CT, MRI, ultrasound, X-ray equipment. The image acquisition device may be implemented as the imaging unit of the above-mentioned medical imaging device, and the image processing device 100/200 may be implemented by the internal processing unit (for example, a processor) of the medical imaging device.
本公开至少一实施例还提供一种存储介质。图10为本公开至少一实施例提供的一种存储介质的示意图。例如,如图10所示,该存储介质400非存储有计算机可读指令401,当计算机可读指令由计算机(包括处理器)执行时可以执行本公开任一实施例提供的图像处理方法。At least one embodiment of the present disclosure also provides a storage medium. FIG. 10 is a schematic diagram of a storage medium provided by at least one embodiment of the present disclosure. For example, as shown in FIG. 10, the storage medium 400 does not store computer-readable instructions 401. When the computer-readable instructions are executed by a computer (including a processor), the image processing method provided in any embodiment of the present disclosure can be executed.
例如,该存储介质可以是一个或多个计算机可读存储介质的任意组合,例如一个计算机可读存储介质包含提取待识别图像中的深度特征的计算机可读的程序代码,另一个计算机可读存储介质包含融合待识别图像的深度特征和专家特征以获取融合特征的计算机可读的程序代码。例如,当该程序代码由计算机读取时,计算机可以执行该计算机存储介质中存储的程序代码,执行例如本公开任一实施例提供的图像处理方法。For example, the storage medium may be any combination of one or more computer-readable storage media. For example, one computer-readable storage medium contains computer-readable program code for extracting depth features in the image to be recognized, and another computer-readable storage medium The medium contains computer-readable program codes that fuse the depth feature and expert feature of the image to be recognized to obtain the fusion feature. For example, when the program code is read by a computer, the computer can execute the program code stored in the computer storage medium, and execute, for example, the image processing method provided in any embodiment of the present disclosure.
例如,存储介质可以包括智能电话的存储卡、平板电脑的存储部件、个人计算机的硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM)、便携式紧致盘只读存储器(CD-ROM)、闪存、或者上述存储介质的任意组合,也可以为其他适用的存储介质。For example, the storage medium may include a memory card of a smart phone, a storage component of a tablet computer, a hard disk of a personal computer, random access memory (RAM), read-only memory (ROM), erasable programmable read-only memory (EPROM), Portable compact disk read-only memory (CD-ROM), flash memory, or any combination of the foregoing storage media may also be other suitable storage media.
有以下几点需要说明:The following points need to be explained:
(1)本公开实施例附图只涉及到与本公开实施例涉及到的结构,其他结构可参考通常设计。(1) The drawings of the embodiments of the present disclosure only refer to the structures related to the embodiments of the present disclosure, and other structures can refer to the usual design.
(2)在不冲突的情况下,本公开的实施例及实施例中的特征可以相互组合以得到新的实施例。(2) In the case of no conflict, the embodiments of the present disclosure and the features in the embodiments can be combined with each other to obtain new embodiments.
以上所述仅是本公开的示范性实施方式,而非用于限制本公开的保护范围,本公开的保护范围由所附的权利要求确定。The above are only exemplary implementations of the present disclosure, and are not used to limit the protection scope of the present disclosure, which is determined by the appended claims.