CN112215054B - Depth generation countermeasure method for denoising underwater sound signal - Google Patents
Depth generation countermeasure method for denoising underwater sound signal Download PDFInfo
- Publication number
- CN112215054B CN112215054B CN202010733300.6A CN202010733300A CN112215054B CN 112215054 B CN112215054 B CN 112215054B CN 202010733300 A CN202010733300 A CN 202010733300A CN 112215054 B CN112215054 B CN 112215054B
- Authority
- CN
- China
- Prior art keywords
- model
- sample
- training
- data
- countermeasure
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 230000005236 sound signal Effects 0.000 title claims abstract 6
- 238000000605 extraction Methods 0.000 claims abstract description 8
- 238000005070 sampling Methods 0.000 claims abstract description 5
- 238000013507 mapping Methods 0.000 claims description 13
- 230000006870 function Effects 0.000 claims description 11
- 239000011159 matrix material Substances 0.000 claims description 2
- 238000005192 partition Methods 0.000 claims description 2
- 230000008485 antagonism Effects 0.000 claims 1
- 238000013135 deep learning Methods 0.000 description 7
- 238000000354 decomposition reaction Methods 0.000 description 4
- 238000013527 convolutional neural network Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000005457 optimization Methods 0.000 description 3
- 230000003044 adaptive effect Effects 0.000 description 2
- 230000005534 acoustic noise Effects 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000000739 chaotic effect Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/02—Preprocessing
- G06F2218/04—Denoising
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/08—Feature extraction
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Signal Processing (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Complex Calculations (AREA)
Abstract
Description
技术领域technical field
本发明属于水声信号降噪技术领域,可以从信噪比较大的水声信号中,有效的复现出原始有用信号。The invention belongs to the technical field of underwater acoustic signal noise reduction, and can effectively reproduce the original useful signal from the underwater acoustic signal with a large signal-to-noise ratio.
背景技术Background technique
现有的水下声信号去噪中,有传统的去噪方法,基于时域的模态分解方法和基于频域的整体模态分解方法,需要提前设置一些经验参数,使得去噪过程依赖经验值,经典模态分解方法可以让去噪过程不再需要提前设置函数的情况下完成,但是在分解的过程中容易产生模式混合和边界效应。为了克服边界混合,基于CEEMDAN、精化复合多尺度色散熵和小波阈值去噪的水声信号降噪方法,在分析混沌信号复杂度方面取得很好的效果,但是没有很好的鲁棒性。In the existing underwater acoustic signal denoising, there are traditional denoising methods, modal decomposition method based on time domain and overall modal decomposition method based on frequency domain, which need to set some empirical parameters in advance, so that the denoising process depends on experience. The classical mode decomposition method can make the denoising process no longer need to set the function in advance, but it is easy to produce mode mixing and boundary effects in the decomposition process. In order to overcome boundary mixing, the underwater acoustic signal denoising method based on CEEMDAN, refined composite multi-scale dispersion entropy and wavelet threshold denoising has achieved good results in analyzing the complexity of chaotic signals, but has no good robustness.
近几年,水声去噪领域引入深度学习的算法来增加系统的鲁棒性,自编码网络就是典型深度学习算法,在去噪算法中对信号与噪声之间的独立性没有做任何假设,使得其去噪系统的鲁棒性优于传统的去噪方法,在自编码网络中加入玻尔兹曼机原理,进一步增强去噪自编码网络的稳健性,使得网络模型更加健壮,然而,深度自编码网络由于其全连接特性,在少数输入增加特征值的情况下,会导致网络参数数量的增加非常庞大,参数太多会带来梯度不稳定性,而且加大整个系统的计算量,所以需要在保证去噪结果的前提下,降低参数数量。在深度学习生成建模领域的一个新突破是生成对抗性网络,其在计算机视觉领域取得了很好的成功,能够生成逼真的图像,在生成对抗基础上改善模型参数与优化算法,均取得很好的效果。生成对抗模型利用生成模型和对抗模型相互优化的作用,可以在有限数量的训练集下取得很好的效果,有效解决水声去噪在深度学习应用方面的局限性。In recent years, deep learning algorithms have been introduced in the field of underwater acoustic denoising to increase the robustness of the system. The self-encoding network is a typical deep learning algorithm. In the denoising algorithm, no assumption is made about the independence between signal and noise. The robustness of its denoising system is better than that of traditional denoising methods. The Boltzmann machine principle is added to the auto-encoding network to further enhance the robustness of the de-noising auto-encoding network and make the network model more robust. However, the depth of Due to its fully connected feature, the auto-encoding network will increase the number of network parameters in the case of a small number of inputs with increased eigenvalues. Too many parameters will bring about gradient instability and increase the calculation amount of the entire system, so It is necessary to reduce the number of parameters on the premise of ensuring the denoising results. A new breakthrough in the field of deep learning generative modeling is the generative adversarial network, which has achieved great success in the field of computer vision, can generate realistic images, and improve model parameters and optimization algorithms on the basis of generative confrontation. Good results. The generative adversarial model utilizes the mutual optimization of the generative model and the adversarial model, which can achieve good results under a limited number of training sets, effectively solving the limitations of underwater acoustic denoising in deep learning applications.
在水声信号降噪过程中,当需要提前设置相应的参数,即对于经验要求很高的情况下,而且系统的鲁棒性较差,传统的降噪方法就不再适用,于是深度学习用于降噪就显示出来其优点,目前深度学习中的生成对抗网络,其生成模型和对抗模型均采用的卷积神经网络,由于水声信号的特点是训练样本较少,在小样本训练的过程中,卷积神经网络的自适应能力较弱,学习出来的网络的鲁棒性较差。In the process of noise reduction of underwater acoustic signals, when the corresponding parameters need to be set in advance, that is, in the case of high experience requirements, and the robustness of the system is poor, the traditional noise reduction method is no longer applicable, so deep learning uses The advantages of noise reduction are shown. At present, the generative adversarial network in deep learning, the convolutional neural network used in both the generative model and the adversarial model, because the underwater acoustic signal is characterized by fewer training samples, in the process of small sample training Among them, the adaptive ability of the convolutional neural network is weak, and the robustness of the learned network is poor.
发明内容SUMMARY OF THE INVENTION
为了解决水声信号特有的小样本训练问题,本文提出深度生成对抗方法,用于水下声信号去噪技术。本方法中的限制玻尔兹曼网络对样本的训练中加入了样本统计特性,所以对于小样本的训练效果很好。针对水声的特点以及水声背景噪声的特征,进一步优化生成对抗网络,使得训练模型适用于水声信道的去噪。In order to solve the problem of small sample training unique to underwater acoustic signals, this paper proposes a deep generative adversarial method for underwater acoustic signal denoising technology. The restricted Boltzmann network in this method adds sample statistical characteristics to the training of the samples, so the training effect for small samples is very good. According to the characteristics of underwater acoustics and the characteristics of underwater acoustic background noise, the generative adversarial network is further optimized, so that the training model is suitable for denoising of underwater acoustic channels.
本发明解决其技术问题所采用的技术方案:一种用于水声信号去噪的深度生成对抗方法,其特点包括下述步骤:The technical solution adopted by the present invention to solve the technical problem: a deep generation confrontation method for denoising underwater acoustic signals, and its characteristics include the following steps:
步骤1:对原始水声信号进行采样、特征提取,本方法采用的是MFCC特征提取方法,梅尔倒谱系数(简称MFCC)是在Mel标度频率域提取出来的倒谱参数,Mel标度描述了人耳频率的非线性特性,它与频率的关系可用下式近似表示:Step 1: Sampling and feature extraction of the original underwater acoustic signal. This method adopts the MFCC feature extraction method. The Mel cepstral coefficient (referred to as MFCC) is the cepstral parameter extracted in the Mel scale frequency domain, and the Mel scale Describes the nonlinear characteristics of the human ear frequency, and its relationship with frequency can be approximated by the following formula:
其中f为采样频率。where f is the sampling frequency.
步骤2:将提取的信号送入高斯-受限玻尔兹曼机中,进行概率生成模型的半监督预训练。Step 2: Send the extracted signal into a Gauss-Restricted Boltzmann Machine for semi-supervised pre-training of the probability generation model.
概率生成模型是由高斯受限玻尔兹曼机堆叠构成的,首先深度模型通过利用不带标签数据,用高斯受限玻尔兹曼机算法从底层开始向上逐层进行非监督预训练得到深度网络超参数的初值,在非监督预训练之后,网络通过利用带标签数据,进行监督训练来调整权值。高斯受限玻尔兹曼机是一种生成式随机网络,由可见层和隐层组成,网络的权值θ由可见层和隐层的连接权值矩阵ω和可见层的偏置向量c,隐层的偏置向量b组成,当给定一组可见层状态v,和隐层状态h,受限玻尔兹曼机的能量函数和似然函数分布表示为:The probabilistic generation model is composed of a stack of Gaussian Restricted Boltzmann Machines. First, the depth model uses unlabeled data and uses the Gauss Restricted Boltzmann Machine algorithm to perform unsupervised pre-training from the bottom layer to the top to obtain the depth. Initial values of network hyperparameters. After unsupervised pre-training, the network adjusts the weights by supervised training using labeled data. The Gauss-restricted Boltzmann machine is a generative random network consisting of a visible layer and a hidden layer. The weight θ of the network consists of the connection weight matrix ω of the visible layer and the hidden layer and the bias vector c of the visible layer, The bias vector b of the hidden layer is composed of, when a set of visible layer states v and hidden layer states h are given, the energy function and likelihood function distribution of the restricted Boltzmann machine are expressed as:
其中,vi∈{0,1};hj∈{0,1};是配分函数,当可见层和隐层其中之一固定状态时,受限玻尔兹曼机的条件概率分布可以表述为Among them, v i ∈{0,1}; h j ∈{0,1}; is the partition function. When one of the visible layer and the hidden layer is in a fixed state, the conditional probability distribution of the restricted Boltzmann machine can be expressed as
其中, in,
生成部分是用一个样本z在一组先验分布Z的映射下,生成x样本的分布,生成器是一种能够模拟真实数据分布以生成与训练集相关新样本的有效映射,生成模型学习的不是传统的输入到输出的映射而是输入的数据流到输出数据流的映射。The generation part is to use a sample z under the mapping of a set of prior distribution Z to generate the distribution of x samples. The generator is an effective mapping that can simulate the distribution of real data to generate new samples related to the training set. Not a traditional input to output mapping but a mapping of input data streams to output data streams.
步骤3:构建深度生成对抗模型,将概率生成模型中生成的数据与真实的标签数据流送入伯努利受限玻尔兹曼机对抗模型,进行有监督训练;Step 3: Build a deep generative adversarial model, and send the data generated in the probabilistic generative model and the real label data stream into the Bernoulli restricted Boltzmann machine adversarial model for supervised training;
生成对抗模型是由生成模型和对抗模型组成,在生成模型中,由于水下声信号特征的复杂性,使得深度神经网络在反向传播来优化网络权值的时候,层数太多,训练系统不稳定,所以采用的概率生成模型,该结构特点是训练过程为逐层训练,每一个玻尔兹曼机都单独训练,然后相叠加,有效避免由于层与层之间的复杂结构造成的反向传播参数更新慢,误差大等缺点。概率生成模型为层内无连接,层内全连接的网络,网络的输入层特征值与输出层特征值数量相等,中间为四层高斯受限玻尔兹曼网络相叠加。The generative adversarial model is composed of a generative model and an adversarial model. In the generative model, due to the complexity of the underwater acoustic signal features, when the deep neural network is back-propagated to optimize the network weights, there are too many layers, and the training system It is unstable, so the probability generation model is adopted. The structural feature is that the training process is layer-by-layer training. Each Boltzmann machine is trained separately and then superimposed to effectively avoid the inverse caused by the complex structure between layers. The propagation parameters are updated slowly and the error is large. The probability generation model is a network with no connection in the layer and full connection in the layer. The eigenvalues of the input layer of the network are equal to the eigenvalues of the output layer, and four layers of Gaussian restricted Boltzmann networks are superimposed in the middle.
伯努利受限玻尔兹曼机对抗模型是整个模型的优化部分,生成部分学习最优映射的方式是通过对抗训练完成的,对抗模型中的自编码模型采用的是伯努利受限玻尔兹曼机模型,很好的对数据训练和微调加入概率判断,最后加入判别层,是一个二值化的分类器,对抗模型的输入数据有两个,一个是来自真实的样本分布X,一个是来自生成器模拟出来的真实样本分布判别器的功能就是可以判断出真实的样本分布X为真实数据,而生成器生成的样本分布为假数据,生成器根据对抗器的输出结果不断的进行优化,直到对抗器无法判断输入数据是真实数据还是生成数据,说明输入数据的分布已经非常接近真实数据的分布了,这种训练成为生成对抗模型的最大最小博弈训练,其目的是如式6所示:The Bernoulli restricted Boltzmann machine adversarial model is the optimization part of the whole model. The way of generating the partial learning optimal mapping is done through adversarial training. The self-encoding model in the adversarial model uses Bernoulli restricted glass. The Ertzmann machine model is a good way to add probability judgment to data training and fine-tuning, and finally adds a discriminant layer, which is a binary classifier. There are two input data for the adversarial model, one is from the real sample distribution X, One is the real sample distribution simulated by the generator The function of the discriminator is to determine that the real sample distribution X is the real data, and the sample distribution generated by the generator For fake data, the generator continuously optimizes according to the output of the adversary, until the adversary cannot judge whether the input data is real data or generated data, indicating that the distribution of the input data is very close to the distribution of the real data, and this training becomes the generated data. The max-min game training of the adversarial model, whose purpose is as shown in Equation 6:
在此基础上,我们可以加入一些额外条件值来执行映射和分类,在去噪模型中,额外条件值为带噪样本xc,加入以后式6可以变为式7所示:On this basis, we can add some additional conditional values to perform mapping and classification. In the denoising model, the additional conditional value is the noisy sample x c . After adding, Equation 6 can be changed to Equation 7:
为了训练过程中的稳定性,将生成器和对抗器分开训练,分别优化两个函数的最小值,如式8、9所示:For the stability of the training process, the generator and the adversary are trained separately, and the minimum values of the two functions are optimized respectively, as shown in equations 8 and 9:
判别网络模型主要负责对抗作用,当判别系统在开始很容易判别出干净样本为1,生成样本为0,当判别生成器的生成样本为0时,生成器开始优化自身参数,使得自身从输入样本映射出的值接近真实样本的值,当生成样本很接近真实样本使得判别器无法判别时,判别器开始优化自身参数,使得可以很好的判别出生成样本与真实样本,生成模型中输入为一个二维数组,当输入生成样本与带噪样本时,判别器判别为0;当输入干净样本与带噪样本时,判别结果为1。The discriminant network model is mainly responsible for the confrontation. When the discriminant system can easily discriminate that the clean sample is 1 and the generated sample is 0 at the beginning, when the generated sample of the discriminant generator is 0, the generator starts to optimize its own parameters, so that it starts from the input sample. The mapped value is close to the value of the real sample. When the generated sample is so close to the real sample that the discriminator cannot discriminate, the discriminator starts to optimize its own parameters, so that the generated sample and the real sample can be well discriminated. The input in the generation model is a A two-dimensional array. When the generated samples and noisy samples are input, the discriminator judges as 0; when the input of clean samples and noisy samples, the discriminant result is 1.
本发明的有益效果是:针对水声信号的特征提取特点,在受限玻尔兹曼的概率模型中引入生成对抗模型,有效的消除由于水声携带的复杂信号引起的受限玻尔兹曼机在训练过程中的强依赖性和过拟合问题,从而使得训练模型自适用性更强。The beneficial effects of the present invention are: in view of the feature extraction characteristics of underwater acoustic signals, a generative confrontation model is introduced into the restricted Boltzmann probability model, which effectively eliminates the restricted Boltzmann caused by complex signals carried by underwater acoustics. The strong dependence and over-fitting problems of the machine in the training process make the training model more self-applicable.
本发明提出深度生成对抗算法,提高水声信号特征提取技术中的信噪比,传统的水声信号降噪方法,如小波去噪算法和集成经验模态分解算法,均在去噪之前有些对样本的假设条件,然而有些假设条件在实际的水下环境中是无法完全满足的,生成对抗网络的编码模型是典型的深度学习网络模型,它可以在不用假设独立性的情况下很好的对带噪信号进行去噪,但是普通的生成对抗去噪模型中生成网络和对抗网络都用卷积神经网络,对具有大量训练样本的数据有很好的鲁棒性,但是针对小样本的水声信号,所以在生成对抗模型的生成单元和对抗单元分别加入具有概率统计特性的高斯受限玻尔兹曼机和伯努利首先玻尔兹曼机,首先利用半监督学习来优化网络参数,形成一种深度生成对抗网络模型。The invention proposes a deep generation confrontation algorithm to improve the signal-to-noise ratio in the feature extraction technology of underwater acoustic signals. The assumptions of the sample, however, some assumptions cannot be fully satisfied in the actual underwater environment. The coding model of the generative adversarial network is a typical deep learning network model, which can be well matched without assuming independence. The noisy signal is denoised, but in the common generative adversarial denoising model, both the generative network and the adversarial network use convolutional neural networks, which have good robustness to data with a large number of training samples, but are not suitable for small samples of underwater sound. Therefore, a Gaussian restricted Boltzmann machine with probabilistic and statistical characteristics and a Bernoulli first Boltzmann machine are added to the generation unit and the confrontation unit of the generative adversarial model. First, semi-supervised learning is used to optimize the network parameters to form A deep generative adversarial network model.
本发明提出的深度生成对抗网络,希望对原有的带噪信号的噪声进行清除,得到干净水声信号,水声噪声主要来源为海洋背景噪声和螺旋桨噪声,包含线性噪声和非线性噪声,而且水声信号训练样本数量有限,深度生成对抗方法学习数据不是学习数据的输入到输出的简单映射关系,而是学习给定样本的数据统计特性与统计特性之间的映射关系,然后利用对抗的方法来进行监督学习,对输入水声信号的统计特性进行不断的更新,所以在线性和非线性混合的系统中以及少量训练集的情况下,也可以很好的重构水声信号,使得其接近于干净水声信号,并且得到的具有数据统计特性的深度生成对抗模型具有很强的鲁棒性和自适应能力。The deep generative adversarial network proposed by the present invention hopes to remove the noise of the original noisy signal to obtain a clean underwater acoustic signal. The main sources of the underwater acoustic noise are ocean background noise and propeller noise, including linear noise and nonlinear noise, and The number of training samples for underwater acoustic signals is limited, and the deep generative adversarial method learning data is not a simple mapping relationship between learning data input and output, but learning the mapping relationship between the statistical characteristics and statistical characteristics of a given sample, and then use the confrontation method. In order to carry out supervised learning and continuously update the statistical characteristics of the input underwater acoustic signal, in the case of a mixed linear and nonlinear system and a small number of training sets, the underwater acoustic signal can also be well reconstructed, making it close to The deep generative adversarial model with statistical characteristics obtained has strong robustness and adaptive ability.
下面结合具体实施方式对本发明作详细说明。The present invention will be described in detail below with reference to specific embodiments.
附图说明Description of drawings
图1是系统的判别模型Figure 1 is the discriminant model of the system
图2是生成对抗网络模型。Figure 2 is a generative adversarial network model.
图3红色为应用深度生成对抗方法去噪以后的水声时域图,蓝色为带噪信号的时域图,绿色为干净信号的时域图。Figure 3: Red is the time-domain map of underwater acoustics after denoising using the deep generative adversarial method, blue is the time-domain map of the noisy signal, and green is the time-domain map of the clean signal.
图4左图为干净水声信号的频谱图,右图为应用深度生成对抗方法去噪以后的信号的频谱图The left picture of Fig. 4 is the spectrogram of the clean underwater acoustic signal, and the right picture is the spectrogram of the signal after denoising by applying the deep generative adversarial method.
具体实施方式Detailed ways
步骤1首先对样本进行分帧、批量处理。
步骤2然后将处理好的数据送入生成模型中进行模型的训练,生成模型是一个半监督模型,所以训练出来的数据与干净数据之间有差别
步骤3将生成器生成的数据加入带噪数据,与原本的干净带噪声数据一起送入判别模型进行判别,开始判别器能很好的判别出来,生成模型的数据为假样本,输出为0,原本干净带噪声数据为真样本输出为1,根据判别器判别出来的结果,生成器开始模拟自己的生成数据,使得数据尽量的接近真实数据,这样直到判别器没有办法判别,将生成器的生成数据再一次送入判别器中,判别器开始判别不出来,输入为两个样本时,输出都为0.5,经过一段时间的训练,可以很好的判别出生成样本为0,重复步骤3、4,直到判别器的输出值稳定在0.5,此时生成器生成的模型为干净样本。Step 3: Add the data generated by the generator to the noisy data, and send it to the discriminant model together with the original clean and noisy data for discrimination. The discriminator can discriminate well at the beginning. The data of the generated model is a fake sample, and the output is 0. The original clean and noisy data is the real sample and the output is 1. According to the result discriminated by the discriminator, the generator starts to simulate its own generated data, so that the data is as close to the real data as possible, so that until the discriminator has no way to distinguish, the generator will generate The data is sent to the discriminator again, and the discriminator starts to fail to discriminate. When the input is two samples, the output is 0.5. After a period of training, it can be well discriminated that the generated sample is 0. Repeat steps 3 and 4 , until the output value of the discriminator stabilizes at 0.5, at which point the model generated by the generator is a clean sample.
从图3、4可以看出,本实施例在相同信噪比下,对于小样本的水声信号去噪效果由于传统去噪方法,经过训练好的去噪模型的带噪信号时域图,与干净样本的信号时域图达到很高的重合度,放在频域中发现,干净样本的能量很好的集中在一个点,经过去噪模型的能量有偏移,但是能量还是会集中在一个特定的点,根据多次实验发现频移是一个固定值,所以有了频移经验值,可以很好的对带噪语音进行去噪。It can be seen from Figures 3 and 4 that, under the same signal-to-noise ratio in this embodiment, the denoising effect of the underwater acoustic signal of small samples is due to the traditional denoising method, and the time domain diagram of the noisy signal of the trained denoising model, It has a high degree of coincidence with the signal time domain map of the clean sample. It is found in the frequency domain that the energy of the clean sample is well concentrated at one point, and the energy after the denoising model is offset, but the energy will still be concentrated in At a specific point, according to many experiments, it is found that the frequency shift is a fixed value, so with the empirical value of the frequency shift, it is possible to denoise the noisy speech very well.
Claims (1)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010733300.6A CN112215054B (en) | 2020-07-27 | 2020-07-27 | Depth generation countermeasure method for denoising underwater sound signal |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010733300.6A CN112215054B (en) | 2020-07-27 | 2020-07-27 | Depth generation countermeasure method for denoising underwater sound signal |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112215054A CN112215054A (en) | 2021-01-12 |
CN112215054B true CN112215054B (en) | 2022-06-28 |
Family
ID=74059288
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010733300.6A Active CN112215054B (en) | 2020-07-27 | 2020-07-27 | Depth generation countermeasure method for denoising underwater sound signal |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112215054B (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113541835B (en) * | 2021-06-17 | 2022-05-20 | 江苏科技大学 | A Time-varying Underwater Acoustic Channel Simulation Method Based on Conditional Generative Adversarial Networks |
CN113823308B (en) * | 2021-09-18 | 2023-11-28 | 东南大学 | Method for denoising voice by using single voice sample with noise |
CN113962381B (en) * | 2021-09-30 | 2023-02-03 | 西南电子技术研究所(中国电子科技集团公司第十研究所) | Weak signal data enhancement method |
CN114785423B (en) * | 2022-06-20 | 2022-11-01 | 中国海洋大学三亚海洋研究院 | Method for building underwater laser digital information transmission enhancement model |
CN116403590B (en) * | 2023-06-08 | 2023-08-18 | 青岛科技大学 | Bionic signal processing method based on wavelet transformation and generation countermeasure network |
CN118467943A (en) * | 2024-05-22 | 2024-08-09 | 中国地质大学(武汉) | A method and system for denoising power frequency harmonic noise based on semi-supervised learning |
CN118379982A (en) * | 2024-06-27 | 2024-07-23 | 武汉普惠海洋光电技术有限公司 | High-frequency array environment noise reduction method and device |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105488466A (en) * | 2015-11-26 | 2016-04-13 | 中国船舶工业系统工程研究院 | Deep neural network and underwater sound target vocal print feature extraction method |
CN107563428A (en) * | 2017-08-25 | 2018-01-09 | 西安电子科技大学 | Classification of Polarimetric SAR Image method based on generation confrontation network |
CN109147817A (en) * | 2018-08-29 | 2019-01-04 | 昆明理工大学 | A kind of denoising audio feature extraction algorithm based on the limited Boltzmann machine that makes a variation |
CN109978079A (en) * | 2019-04-10 | 2019-07-05 | 东北电力大学 | A kind of data cleaning method of improved storehouse noise reduction self-encoding encoder |
CN109993710A (en) * | 2019-03-20 | 2019-07-09 | 西北工业大学 | An underwater image denoising method based on generative adversarial network |
CN110490816A (en) * | 2019-07-15 | 2019-11-22 | 哈尔滨工程大学 | A kind of underwater Heterogeneous Information data noise reduction |
CN111401236A (en) * | 2020-03-16 | 2020-07-10 | 西北工业大学 | Denoising method of underwater acoustic signal based on autoencoder neural network |
CN111428853A (en) * | 2020-01-16 | 2020-07-17 | 东华大学 | A Negative Sample Adversarial Generation Method with Noise Learning |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11257259B2 (en) * | 2017-08-15 | 2022-02-22 | Siemens Healthcare Gmbh | Topogram prediction from surface data in medical imaging |
CN108765319B (en) * | 2018-05-09 | 2020-08-14 | 大连理工大学 | Image denoising method based on generation countermeasure network |
CN110020685A (en) * | 2019-04-09 | 2019-07-16 | 山东超越数控电子股份有限公司 | A kind of preprocess method based on adaptive-filtering and limited Boltzmann machine, terminal and readable storage medium storing program for executing |
-
2020
- 2020-07-27 CN CN202010733300.6A patent/CN112215054B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105488466A (en) * | 2015-11-26 | 2016-04-13 | 中国船舶工业系统工程研究院 | Deep neural network and underwater sound target vocal print feature extraction method |
CN107563428A (en) * | 2017-08-25 | 2018-01-09 | 西安电子科技大学 | Classification of Polarimetric SAR Image method based on generation confrontation network |
CN109147817A (en) * | 2018-08-29 | 2019-01-04 | 昆明理工大学 | A kind of denoising audio feature extraction algorithm based on the limited Boltzmann machine that makes a variation |
CN109993710A (en) * | 2019-03-20 | 2019-07-09 | 西北工业大学 | An underwater image denoising method based on generative adversarial network |
CN109978079A (en) * | 2019-04-10 | 2019-07-05 | 东北电力大学 | A kind of data cleaning method of improved storehouse noise reduction self-encoding encoder |
CN110490816A (en) * | 2019-07-15 | 2019-11-22 | 哈尔滨工程大学 | A kind of underwater Heterogeneous Information data noise reduction |
CN111428853A (en) * | 2020-01-16 | 2020-07-17 | 东华大学 | A Negative Sample Adversarial Generation Method with Noise Learning |
CN111401236A (en) * | 2020-03-16 | 2020-07-10 | 西北工业大学 | Denoising method of underwater acoustic signal based on autoencoder neural network |
Non-Patent Citations (5)
Title |
---|
Boosting Noise Robustness of Acoustic Model via Deep Adversarial Training;Bin Liu 等;《2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)》;20180913;第5034-5038页 * |
Spectra-GANs: A New Automated Denoising Method for Low-S/N Stellar Spectra;Minglei Wu 等;《IEEE Access》;20200605;第8卷;第107912-107926页 * |
The research of underwater target recognition method based on deep learning;Yuechao Chen 等;《2017 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC)》;20180101;第1-5页 * |
基于Adversarial的生成式算法研究;罗金铭;《中国硕士学位论文全文数据库 信息科技专辑》;20190215;第2019年卷(第2期);I138-1093 * |
生成对抗网络的改进方法与超分辨率图像重建研究;李艳春;《中国博士学位论文全文数据库 信息科技辑》;20200115;第2020年卷(第1期);I138-77 * |
Also Published As
Publication number | Publication date |
---|---|
CN112215054A (en) | 2021-01-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112215054B (en) | Depth generation countermeasure method for denoising underwater sound signal | |
CN109919204B (en) | A Deep Learning Clustering Method for Noisy Images | |
CN113723295A (en) | Face counterfeiting detection method based on image domain frequency domain double-flow network | |
CN110334580A (en) | The equipment fault classification method of changeable weight combination based on integrated increment | |
CN112307714A (en) | Character style migration method based on double-stage deep network | |
CN109800785B (en) | Data classification method and device based on self-expression correlation | |
CN112562698B (en) | A defect diagnosis method for power equipment based on fusion of sound source information and thermal imaging features | |
CN110490265B (en) | Image steganalysis method based on double-path convolution and feature fusion | |
CN114283287A (en) | A Robust Domain Adaptive Image Learning Method Based on Self-training Noise Label Correction | |
CN112464004A (en) | Multi-view depth generation image clustering method | |
CN111681188B (en) | Image Deblurring Method Based on Combining Image Pixel Prior and Image Gradient Prior | |
CN109859310B (en) | A model that can be used to generate MR images and its establishment method | |
CN114821204A (en) | A semi-supervised learning image classification method and system based on meta-learning embedding | |
CN110321777A (en) | A kind of face identification method based on the sparse denoising self-encoding encoder of stack convolution | |
CN113850013A (en) | Ship radiation noise classification method | |
CN111160327B (en) | An Expression Recognition Method Based on Lightweight Convolutional Neural Network | |
CN113628615B (en) | Speech recognition method, device, electronic device and storage medium | |
CN112163605A (en) | Multi-domain image translation method based on attention network generation | |
CN117036901A (en) | Small sample fine adjustment method based on visual self-attention model | |
CN110766083A (en) | A feature fusion based Alexnet mural image classification method | |
CN115861779A (en) | Unbiased scene graph generation method based on effective feature representation | |
CN116776093A (en) | An EEG signal data amplification algorithm based on improved generative adversarial network | |
CN116884090A (en) | A cross-environment human activity recognition method based on dual adversarial networks | |
CN116434758A (en) | Voiceprint recognition model training method and device, electronic equipment and storage medium | |
CN115811304A (en) | A Method for Eliminating Ocean Impulse Noise Based on Conditional Variational Autoencoder Network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |