[go: up one dir, main page]

CN112215054B - Depth generation countermeasure method for denoising underwater sound signal - Google Patents

Depth generation countermeasure method for denoising underwater sound signal Download PDF

Info

Publication number
CN112215054B
CN112215054B CN202010733300.6A CN202010733300A CN112215054B CN 112215054 B CN112215054 B CN 112215054B CN 202010733300 A CN202010733300 A CN 202010733300A CN 112215054 B CN112215054 B CN 112215054B
Authority
CN
China
Prior art keywords
model
sample
training
data
countermeasure
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010733300.6A
Other languages
Chinese (zh)
Other versions
CN112215054A (en
Inventor
曾向阳
薛灵芝
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN202010733300.6A priority Critical patent/CN112215054B/en
Publication of CN112215054A publication Critical patent/CN112215054A/en
Application granted granted Critical
Publication of CN112215054B publication Critical patent/CN112215054B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/02Preprocessing
    • G06F2218/04Denoising
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Signal Processing (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Complex Calculations (AREA)

Abstract

The invention discloses a deep generation countermeasure method for denoising underwater sound signals, and belongs to the technical field of underwater sound signal denoising. Firstly, sampling and feature extraction are carried out on an original underwater sound signal, then the extracted signal is sent into a Gauss-limited Boltzmann machine, and semi-supervised pre-training of a probability generation model is carried out; and finally, constructing a deep generation countermeasure model, sending the data generated in the probability generation model and the real label data stream into a Bernoulli limited Boltzmann machine countermeasure model, and performing supervised training. Aiming at the characteristic extraction characteristic of the underwater sound signal, the generation countermeasure model is introduced into the probability model of the limited Boltzmann machine, so that the problems of strong dependence and overfitting of the limited Boltzmann machine in the training process caused by the complex signal carried by the underwater sound are effectively eliminated, and the training model has stronger self-applicability.

Description

一种用于水声信号去噪的深度生成对抗方法A Deep Generative Adversarial Method for Underwater Acoustic Signal Denoising

技术领域technical field

本发明属于水声信号降噪技术领域,可以从信噪比较大的水声信号中,有效的复现出原始有用信号。The invention belongs to the technical field of underwater acoustic signal noise reduction, and can effectively reproduce the original useful signal from the underwater acoustic signal with a large signal-to-noise ratio.

背景技术Background technique

现有的水下声信号去噪中,有传统的去噪方法,基于时域的模态分解方法和基于频域的整体模态分解方法,需要提前设置一些经验参数,使得去噪过程依赖经验值,经典模态分解方法可以让去噪过程不再需要提前设置函数的情况下完成,但是在分解的过程中容易产生模式混合和边界效应。为了克服边界混合,基于CEEMDAN、精化复合多尺度色散熵和小波阈值去噪的水声信号降噪方法,在分析混沌信号复杂度方面取得很好的效果,但是没有很好的鲁棒性。In the existing underwater acoustic signal denoising, there are traditional denoising methods, modal decomposition method based on time domain and overall modal decomposition method based on frequency domain, which need to set some empirical parameters in advance, so that the denoising process depends on experience. The classical mode decomposition method can make the denoising process no longer need to set the function in advance, but it is easy to produce mode mixing and boundary effects in the decomposition process. In order to overcome boundary mixing, the underwater acoustic signal denoising method based on CEEMDAN, refined composite multi-scale dispersion entropy and wavelet threshold denoising has achieved good results in analyzing the complexity of chaotic signals, but has no good robustness.

近几年,水声去噪领域引入深度学习的算法来增加系统的鲁棒性,自编码网络就是典型深度学习算法,在去噪算法中对信号与噪声之间的独立性没有做任何假设,使得其去噪系统的鲁棒性优于传统的去噪方法,在自编码网络中加入玻尔兹曼机原理,进一步增强去噪自编码网络的稳健性,使得网络模型更加健壮,然而,深度自编码网络由于其全连接特性,在少数输入增加特征值的情况下,会导致网络参数数量的增加非常庞大,参数太多会带来梯度不稳定性,而且加大整个系统的计算量,所以需要在保证去噪结果的前提下,降低参数数量。在深度学习生成建模领域的一个新突破是生成对抗性网络,其在计算机视觉领域取得了很好的成功,能够生成逼真的图像,在生成对抗基础上改善模型参数与优化算法,均取得很好的效果。生成对抗模型利用生成模型和对抗模型相互优化的作用,可以在有限数量的训练集下取得很好的效果,有效解决水声去噪在深度学习应用方面的局限性。In recent years, deep learning algorithms have been introduced in the field of underwater acoustic denoising to increase the robustness of the system. The self-encoding network is a typical deep learning algorithm. In the denoising algorithm, no assumption is made about the independence between signal and noise. The robustness of its denoising system is better than that of traditional denoising methods. The Boltzmann machine principle is added to the auto-encoding network to further enhance the robustness of the de-noising auto-encoding network and make the network model more robust. However, the depth of Due to its fully connected feature, the auto-encoding network will increase the number of network parameters in the case of a small number of inputs with increased eigenvalues. Too many parameters will bring about gradient instability and increase the calculation amount of the entire system, so It is necessary to reduce the number of parameters on the premise of ensuring the denoising results. A new breakthrough in the field of deep learning generative modeling is the generative adversarial network, which has achieved great success in the field of computer vision, can generate realistic images, and improve model parameters and optimization algorithms on the basis of generative confrontation. Good results. The generative adversarial model utilizes the mutual optimization of the generative model and the adversarial model, which can achieve good results under a limited number of training sets, effectively solving the limitations of underwater acoustic denoising in deep learning applications.

在水声信号降噪过程中,当需要提前设置相应的参数,即对于经验要求很高的情况下,而且系统的鲁棒性较差,传统的降噪方法就不再适用,于是深度学习用于降噪就显示出来其优点,目前深度学习中的生成对抗网络,其生成模型和对抗模型均采用的卷积神经网络,由于水声信号的特点是训练样本较少,在小样本训练的过程中,卷积神经网络的自适应能力较弱,学习出来的网络的鲁棒性较差。In the process of noise reduction of underwater acoustic signals, when the corresponding parameters need to be set in advance, that is, in the case of high experience requirements, and the robustness of the system is poor, the traditional noise reduction method is no longer applicable, so deep learning uses The advantages of noise reduction are shown. At present, the generative adversarial network in deep learning, the convolutional neural network used in both the generative model and the adversarial model, because the underwater acoustic signal is characterized by fewer training samples, in the process of small sample training Among them, the adaptive ability of the convolutional neural network is weak, and the robustness of the learned network is poor.

发明内容SUMMARY OF THE INVENTION

为了解决水声信号特有的小样本训练问题,本文提出深度生成对抗方法,用于水下声信号去噪技术。本方法中的限制玻尔兹曼网络对样本的训练中加入了样本统计特性,所以对于小样本的训练效果很好。针对水声的特点以及水声背景噪声的特征,进一步优化生成对抗网络,使得训练模型适用于水声信道的去噪。In order to solve the problem of small sample training unique to underwater acoustic signals, this paper proposes a deep generative adversarial method for underwater acoustic signal denoising technology. The restricted Boltzmann network in this method adds sample statistical characteristics to the training of the samples, so the training effect for small samples is very good. According to the characteristics of underwater acoustics and the characteristics of underwater acoustic background noise, the generative adversarial network is further optimized, so that the training model is suitable for denoising of underwater acoustic channels.

本发明解决其技术问题所采用的技术方案:一种用于水声信号去噪的深度生成对抗方法,其特点包括下述步骤:The technical solution adopted by the present invention to solve the technical problem: a deep generation confrontation method for denoising underwater acoustic signals, and its characteristics include the following steps:

步骤1:对原始水声信号进行采样、特征提取,本方法采用的是MFCC特征提取方法,梅尔倒谱系数(简称MFCC)是在Mel标度频率域提取出来的倒谱参数,Mel标度描述了人耳频率的非线性特性,它与频率的关系可用下式近似表示:Step 1: Sampling and feature extraction of the original underwater acoustic signal. This method adopts the MFCC feature extraction method. The Mel cepstral coefficient (referred to as MFCC) is the cepstral parameter extracted in the Mel scale frequency domain, and the Mel scale Describes the nonlinear characteristics of the human ear frequency, and its relationship with frequency can be approximated by the following formula:

Figure BDA0002604098430000031
Figure BDA0002604098430000031

其中f为采样频率。where f is the sampling frequency.

步骤2:将提取的信号送入高斯-受限玻尔兹曼机中,进行概率生成模型的半监督预训练。Step 2: Send the extracted signal into a Gauss-Restricted Boltzmann Machine for semi-supervised pre-training of the probability generation model.

概率生成模型是由高斯受限玻尔兹曼机堆叠构成的,首先深度模型通过利用不带标签数据,用高斯受限玻尔兹曼机算法从底层开始向上逐层进行非监督预训练得到深度网络超参数的初值,在非监督预训练之后,网络通过利用带标签数据,进行监督训练来调整权值。高斯受限玻尔兹曼机是一种生成式随机网络,由可见层和隐层组成,网络的权值θ由可见层和隐层的连接权值矩阵ω和可见层的偏置向量c,隐层的偏置向量b组成,当给定一组可见层状态v,和隐层状态h,受限玻尔兹曼机的能量函数和似然函数分布表示为:The probabilistic generation model is composed of a stack of Gaussian Restricted Boltzmann Machines. First, the depth model uses unlabeled data and uses the Gauss Restricted Boltzmann Machine algorithm to perform unsupervised pre-training from the bottom layer to the top to obtain the depth. Initial values of network hyperparameters. After unsupervised pre-training, the network adjusts the weights by supervised training using labeled data. The Gauss-restricted Boltzmann machine is a generative random network consisting of a visible layer and a hidden layer. The weight θ of the network consists of the connection weight matrix ω of the visible layer and the hidden layer and the bias vector c of the visible layer, The bias vector b of the hidden layer is composed of, when a set of visible layer states v and hidden layer states h are given, the energy function and likelihood function distribution of the restricted Boltzmann machine are expressed as:

Figure BDA0002604098430000032
Figure BDA0002604098430000032

Figure BDA0002604098430000033
Figure BDA0002604098430000033

其中,vi∈{0,1};hj∈{0,1};

Figure BDA0002604098430000034
是配分函数,当可见层和隐层其中之一固定状态时,受限玻尔兹曼机的条件概率分布可以表述为Among them, v i ∈{0,1}; h j ∈{0,1};
Figure BDA0002604098430000034
is the partition function. When one of the visible layer and the hidden layer is in a fixed state, the conditional probability distribution of the restricted Boltzmann machine can be expressed as

Figure BDA0002604098430000035
Figure BDA0002604098430000035

Figure BDA0002604098430000036
Figure BDA0002604098430000036

其中,

Figure BDA0002604098430000037
in,
Figure BDA0002604098430000037

生成部分是用一个样本z在一组先验分布Z的映射下,生成x样本的分布,生成器是一种能够模拟真实数据分布以生成与训练集相关新样本的有效映射,生成模型学习的不是传统的输入到输出的映射而是输入的数据流到输出数据流的映射。The generation part is to use a sample z under the mapping of a set of prior distribution Z to generate the distribution of x samples. The generator is an effective mapping that can simulate the distribution of real data to generate new samples related to the training set. Not a traditional input to output mapping but a mapping of input data streams to output data streams.

步骤3:构建深度生成对抗模型,将概率生成模型中生成的数据与真实的标签数据流送入伯努利受限玻尔兹曼机对抗模型,进行有监督训练;Step 3: Build a deep generative adversarial model, and send the data generated in the probabilistic generative model and the real label data stream into the Bernoulli restricted Boltzmann machine adversarial model for supervised training;

生成对抗模型是由生成模型和对抗模型组成,在生成模型中,由于水下声信号特征的复杂性,使得深度神经网络在反向传播来优化网络权值的时候,层数太多,训练系统不稳定,所以采用的概率生成模型,该结构特点是训练过程为逐层训练,每一个玻尔兹曼机都单独训练,然后相叠加,有效避免由于层与层之间的复杂结构造成的反向传播参数更新慢,误差大等缺点。概率生成模型为层内无连接,层内全连接的网络,网络的输入层特征值与输出层特征值数量相等,中间为四层高斯受限玻尔兹曼网络相叠加。The generative adversarial model is composed of a generative model and an adversarial model. In the generative model, due to the complexity of the underwater acoustic signal features, when the deep neural network is back-propagated to optimize the network weights, there are too many layers, and the training system It is unstable, so the probability generation model is adopted. The structural feature is that the training process is layer-by-layer training. Each Boltzmann machine is trained separately and then superimposed to effectively avoid the inverse caused by the complex structure between layers. The propagation parameters are updated slowly and the error is large. The probability generation model is a network with no connection in the layer and full connection in the layer. The eigenvalues of the input layer of the network are equal to the eigenvalues of the output layer, and four layers of Gaussian restricted Boltzmann networks are superimposed in the middle.

伯努利受限玻尔兹曼机对抗模型是整个模型的优化部分,生成部分学习最优映射的方式是通过对抗训练完成的,对抗模型中的自编码模型采用的是伯努利受限玻尔兹曼机模型,很好的对数据训练和微调加入概率判断,最后加入判别层,是一个二值化的分类器,对抗模型的输入数据有两个,一个是来自真实的样本分布X,一个是来自生成器模拟出来的真实样本分布

Figure BDA0002604098430000041
判别器的功能就是可以判断出真实的样本分布X为真实数据,而生成器生成的样本分布
Figure BDA0002604098430000042
为假数据,生成器根据对抗器的输出结果不断的进行优化,直到对抗器无法判断输入数据是真实数据还是生成数据,说明输入数据的分布已经非常接近真实数据的分布了,这种训练成为生成对抗模型的最大最小博弈训练,其目的是如式6所示:The Bernoulli restricted Boltzmann machine adversarial model is the optimization part of the whole model. The way of generating the partial learning optimal mapping is done through adversarial training. The self-encoding model in the adversarial model uses Bernoulli restricted glass. The Ertzmann machine model is a good way to add probability judgment to data training and fine-tuning, and finally adds a discriminant layer, which is a binary classifier. There are two input data for the adversarial model, one is from the real sample distribution X, One is the real sample distribution simulated by the generator
Figure BDA0002604098430000041
The function of the discriminator is to determine that the real sample distribution X is the real data, and the sample distribution generated by the generator
Figure BDA0002604098430000042
For fake data, the generator continuously optimizes according to the output of the adversary, until the adversary cannot judge whether the input data is real data or generated data, indicating that the distribution of the input data is very close to the distribution of the real data, and this training becomes the generated data. The max-min game training of the adversarial model, whose purpose is as shown in Equation 6:

Figure BDA0002604098430000051
Figure BDA0002604098430000051

在此基础上,我们可以加入一些额外条件值来执行映射和分类,在去噪模型中,额外条件值为带噪样本xc,加入以后式6可以变为式7所示:On this basis, we can add some additional conditional values to perform mapping and classification. In the denoising model, the additional conditional value is the noisy sample x c . After adding, Equation 6 can be changed to Equation 7:

Figure BDA0002604098430000052
Figure BDA0002604098430000052

为了训练过程中的稳定性,将生成器和对抗器分开训练,分别优化两个函数的最小值,如式8、9所示:For the stability of the training process, the generator and the adversary are trained separately, and the minimum values of the two functions are optimized respectively, as shown in equations 8 and 9:

Figure BDA0002604098430000053
Figure BDA0002604098430000053

Figure BDA0002604098430000054
Figure BDA0002604098430000054

判别网络模型主要负责对抗作用,当判别系统在开始很容易判别出干净样本为1,生成样本为0,当判别生成器的生成样本为0时,生成器开始优化自身参数,使得自身从输入样本映射出的值接近真实样本的值,当生成样本很接近真实样本使得判别器无法判别时,判别器开始优化自身参数,使得可以很好的判别出生成样本与真实样本,生成模型中输入为一个二维数组,当输入生成样本与带噪样本时,判别器判别为0;当输入干净样本与带噪样本时,判别结果为1。The discriminant network model is mainly responsible for the confrontation. When the discriminant system can easily discriminate that the clean sample is 1 and the generated sample is 0 at the beginning, when the generated sample of the discriminant generator is 0, the generator starts to optimize its own parameters, so that it starts from the input sample. The mapped value is close to the value of the real sample. When the generated sample is so close to the real sample that the discriminator cannot discriminate, the discriminator starts to optimize its own parameters, so that the generated sample and the real sample can be well discriminated. The input in the generation model is a A two-dimensional array. When the generated samples and noisy samples are input, the discriminator judges as 0; when the input of clean samples and noisy samples, the discriminant result is 1.

本发明的有益效果是:针对水声信号的特征提取特点,在受限玻尔兹曼的概率模型中引入生成对抗模型,有效的消除由于水声携带的复杂信号引起的受限玻尔兹曼机在训练过程中的强依赖性和过拟合问题,从而使得训练模型自适用性更强。The beneficial effects of the present invention are: in view of the feature extraction characteristics of underwater acoustic signals, a generative confrontation model is introduced into the restricted Boltzmann probability model, which effectively eliminates the restricted Boltzmann caused by complex signals carried by underwater acoustics. The strong dependence and over-fitting problems of the machine in the training process make the training model more self-applicable.

本发明提出深度生成对抗算法,提高水声信号特征提取技术中的信噪比,传统的水声信号降噪方法,如小波去噪算法和集成经验模态分解算法,均在去噪之前有些对样本的假设条件,然而有些假设条件在实际的水下环境中是无法完全满足的,生成对抗网络的编码模型是典型的深度学习网络模型,它可以在不用假设独立性的情况下很好的对带噪信号进行去噪,但是普通的生成对抗去噪模型中生成网络和对抗网络都用卷积神经网络,对具有大量训练样本的数据有很好的鲁棒性,但是针对小样本的水声信号,所以在生成对抗模型的生成单元和对抗单元分别加入具有概率统计特性的高斯受限玻尔兹曼机和伯努利首先玻尔兹曼机,首先利用半监督学习来优化网络参数,形成一种深度生成对抗网络模型。The invention proposes a deep generation confrontation algorithm to improve the signal-to-noise ratio in the feature extraction technology of underwater acoustic signals. The assumptions of the sample, however, some assumptions cannot be fully satisfied in the actual underwater environment. The coding model of the generative adversarial network is a typical deep learning network model, which can be well matched without assuming independence. The noisy signal is denoised, but in the common generative adversarial denoising model, both the generative network and the adversarial network use convolutional neural networks, which have good robustness to data with a large number of training samples, but are not suitable for small samples of underwater sound. Therefore, a Gaussian restricted Boltzmann machine with probabilistic and statistical characteristics and a Bernoulli first Boltzmann machine are added to the generation unit and the confrontation unit of the generative adversarial model. First, semi-supervised learning is used to optimize the network parameters to form A deep generative adversarial network model.

本发明提出的深度生成对抗网络,希望对原有的带噪信号的噪声进行清除,得到干净水声信号,水声噪声主要来源为海洋背景噪声和螺旋桨噪声,包含线性噪声和非线性噪声,而且水声信号训练样本数量有限,深度生成对抗方法学习数据不是学习数据的输入到输出的简单映射关系,而是学习给定样本的数据统计特性与统计特性之间的映射关系,然后利用对抗的方法来进行监督学习,对输入水声信号的统计特性进行不断的更新,所以在线性和非线性混合的系统中以及少量训练集的情况下,也可以很好的重构水声信号,使得其接近于干净水声信号,并且得到的具有数据统计特性的深度生成对抗模型具有很强的鲁棒性和自适应能力。The deep generative adversarial network proposed by the present invention hopes to remove the noise of the original noisy signal to obtain a clean underwater acoustic signal. The main sources of the underwater acoustic noise are ocean background noise and propeller noise, including linear noise and nonlinear noise, and The number of training samples for underwater acoustic signals is limited, and the deep generative adversarial method learning data is not a simple mapping relationship between learning data input and output, but learning the mapping relationship between the statistical characteristics and statistical characteristics of a given sample, and then use the confrontation method. In order to carry out supervised learning and continuously update the statistical characteristics of the input underwater acoustic signal, in the case of a mixed linear and nonlinear system and a small number of training sets, the underwater acoustic signal can also be well reconstructed, making it close to The deep generative adversarial model with statistical characteristics obtained has strong robustness and adaptive ability.

下面结合具体实施方式对本发明作详细说明。The present invention will be described in detail below with reference to specific embodiments.

附图说明Description of drawings

图1是系统的判别模型Figure 1 is the discriminant model of the system

图2是生成对抗网络模型。Figure 2 is a generative adversarial network model.

图3红色为应用深度生成对抗方法去噪以后的水声时域图,蓝色为带噪信号的时域图,绿色为干净信号的时域图。Figure 3: Red is the time-domain map of underwater acoustics after denoising using the deep generative adversarial method, blue is the time-domain map of the noisy signal, and green is the time-domain map of the clean signal.

图4左图为干净水声信号的频谱图,右图为应用深度生成对抗方法去噪以后的信号的频谱图The left picture of Fig. 4 is the spectrogram of the clean underwater acoustic signal, and the right picture is the spectrogram of the signal after denoising by applying the deep generative adversarial method.

具体实施方式Detailed ways

步骤1首先对样本进行分帧、批量处理。Step 1 First, the samples are divided into frames and processed in batches.

步骤2然后将处理好的数据送入生成模型中进行模型的训练,生成模型是一个半监督模型,所以训练出来的数据与干净数据之间有差别Step 2 Then send the processed data into the generative model for model training. The generative model is a semi-supervised model, so there is a difference between the trained data and the clean data

步骤3将生成器生成的数据加入带噪数据,与原本的干净带噪声数据一起送入判别模型进行判别,开始判别器能很好的判别出来,生成模型的数据为假样本,输出为0,原本干净带噪声数据为真样本输出为1,根据判别器判别出来的结果,生成器开始模拟自己的生成数据,使得数据尽量的接近真实数据,这样直到判别器没有办法判别,将生成器的生成数据再一次送入判别器中,判别器开始判别不出来,输入为两个样本时,输出都为0.5,经过一段时间的训练,可以很好的判别出生成样本为0,重复步骤3、4,直到判别器的输出值稳定在0.5,此时生成器生成的模型为干净样本。Step 3: Add the data generated by the generator to the noisy data, and send it to the discriminant model together with the original clean and noisy data for discrimination. The discriminator can discriminate well at the beginning. The data of the generated model is a fake sample, and the output is 0. The original clean and noisy data is the real sample and the output is 1. According to the result discriminated by the discriminator, the generator starts to simulate its own generated data, so that the data is as close to the real data as possible, so that until the discriminator has no way to distinguish, the generator will generate The data is sent to the discriminator again, and the discriminator starts to fail to discriminate. When the input is two samples, the output is 0.5. After a period of training, it can be well discriminated that the generated sample is 0. Repeat steps 3 and 4 , until the output value of the discriminator stabilizes at 0.5, at which point the model generated by the generator is a clean sample.

从图3、4可以看出,本实施例在相同信噪比下,对于小样本的水声信号去噪效果由于传统去噪方法,经过训练好的去噪模型的带噪信号时域图,与干净样本的信号时域图达到很高的重合度,放在频域中发现,干净样本的能量很好的集中在一个点,经过去噪模型的能量有偏移,但是能量还是会集中在一个特定的点,根据多次实验发现频移是一个固定值,所以有了频移经验值,可以很好的对带噪语音进行去噪。It can be seen from Figures 3 and 4 that, under the same signal-to-noise ratio in this embodiment, the denoising effect of the underwater acoustic signal of small samples is due to the traditional denoising method, and the time domain diagram of the noisy signal of the trained denoising model, It has a high degree of coincidence with the signal time domain map of the clean sample. It is found in the frequency domain that the energy of the clean sample is well concentrated at one point, and the energy after the denoising model is offset, but the energy will still be concentrated in At a specific point, according to many experiments, it is found that the frequency shift is a fixed value, so with the empirical value of the frequency shift, it is possible to denoise the noisy speech very well.

Claims (1)

1. A depth-generated countermeasure method for denoising underwater acoustic signals is characterized by comprising the following steps:
step 1: sampling and feature extracting are carried out on the original underwater sound signal; the used feature extraction method is a Mel cepstrum coefficient feature extraction method, wherein Mel cepstrum coefficients are abbreviated as MFCC, MFCC is cepstrum parameters extracted in Mel scale frequency domain, Mel scale describes nonlinear characteristics of human ear frequency, and the relationship between the Mel cepstrum coefficients and the frequency is approximately expressed by the following formula:
Figure FDA0003578436170000011
wherein f is the sampling frequency;
step 2: sending the extracted signals into a Gauss limited Boltzmann machine for semi-supervised pre-training of a probability generation model;
Firstly, performing unsupervised pre-training on a depth model layer by layer from the bottom layer upwards by using data without labels by using a limited Gauss boltzmann machine algorithm to obtain an initial value of a hyper-parameter of a depth network, and after unsupervised pre-training, performing supervised training on the network by using data with labels to adjust a weight; the Gauss limited Boltzmann machine is a generating random network, which consists of a visible layer and a hidden layer, wherein a weight theta of the network consists of a connection weight matrix omega of the visible layer and the hidden layer, a bias vector c of the visible layer and a bias vector b of the hidden layer; given a set of visible layer states v, and hidden layer states h, the energy function and likelihood function distribution of a constrained boltzmann machine is represented as:
Figure FDA0003578436170000012
Figure FDA0003578436170000021
wherein v isi∈{0,1};hj∈{0,1};
Figure FDA0003578436170000022
Is a partition function, when one of the visible layer and the hidden layer is in a fixed state, the conditional probability distribution of the restricted Boltzmann machine can be expressed as
Figure FDA0003578436170000023
Figure FDA0003578436170000024
Wherein,
Figure FDA0003578436170000025
the generation part generates the distribution of x samples by using a sample S under the mapping of a group of prior distribution S, the generator is an effective mapping which can simulate real data distribution to generate new samples related to a training set, and the generation model learns the mapping from input data flow to output data flow instead of the traditional mapping from input to output;
And 3, step 3: constructing a deep generation countermeasure model, sending data generated in the probability generation model and a real label data stream into a Bernoulli limited Boltzmann machine countermeasure model, and carrying out supervised training;
the generation countermeasure model consists of a generation model and a countermeasure model, wherein the generation model adopts a probability generation model; the countermeasure model is a Bernoulli limited Boltzmann machine countermeasure model which is an optimized part of the whole model, the mode of generating part learning optimal mapping is completed by countermeasure training, a self-coding model in the countermeasure model adopts a Bernoulli limited Boltzmann machine model, the probability judgment is well added to data training and fine tuning,finally, a discrimination layer is added, namely a binary classifier is adopted, and input data of the countermeasure model are divided into two parts, namely a real sample distribution X from the generator and a real sample distribution simulated by the generator
Figure FDA0003578436170000035
The function of the discriminator is to determine the true sample distribution X as true data, and the generator generates the sample distribution
Figure FDA0003578436170000036
For false data, the generator continuously optimizes according to the output result of the countermeasure until the countermeasure cannot judge whether the input data is real data or generated data, which shows that the distribution of the input data is very close to that of the real data, and the training becomes the maximum and minimum game training for generating the countermeasure model, and the purpose is as shown in formula 6:
Figure FDA0003578436170000031
In the de-noising model, the additional condition value is a noisy sample x1Formula 6 may become represented by formula 7 after addition:
Figure FDA0003578436170000032
training the generator and the countermeasure device separately, and respectively optimizing the minimum value of two functions, as shown in formulas 8 and 9:
Figure FDA0003578436170000033
Figure FDA0003578436170000034
the judgment network model is mainly responsible for antagonism, the judgment system easily judges that a clean sample is 1 and a generated sample is 0 at the beginning, when the generated sample of the judgment generator is 0, the generator starts to optimize self parameters to enable a value mapped by the generator from an input sample to be close to a value of a real sample, when the generated sample is close to the real sample and the judgment device cannot accurately judge whether the generated sample is the clean sample or the generated sample, the judgment device starts to optimize the self parameters to enable the generated sample and the real sample to be well judged, a two-dimensional array is input into the generation model, and when the generated sample and a noisy sample are input, the judgment device judges that the generated sample is 0; when the clean sample and the noisy sample are input, the discrimination result is 1.
CN202010733300.6A 2020-07-27 2020-07-27 Depth generation countermeasure method for denoising underwater sound signal Active CN112215054B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010733300.6A CN112215054B (en) 2020-07-27 2020-07-27 Depth generation countermeasure method for denoising underwater sound signal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010733300.6A CN112215054B (en) 2020-07-27 2020-07-27 Depth generation countermeasure method for denoising underwater sound signal

Publications (2)

Publication Number Publication Date
CN112215054A CN112215054A (en) 2021-01-12
CN112215054B true CN112215054B (en) 2022-06-28

Family

ID=74059288

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010733300.6A Active CN112215054B (en) 2020-07-27 2020-07-27 Depth generation countermeasure method for denoising underwater sound signal

Country Status (1)

Country Link
CN (1) CN112215054B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113541835B (en) * 2021-06-17 2022-05-20 江苏科技大学 A Time-varying Underwater Acoustic Channel Simulation Method Based on Conditional Generative Adversarial Networks
CN113823308B (en) * 2021-09-18 2023-11-28 东南大学 Method for denoising voice by using single voice sample with noise
CN113962381B (en) * 2021-09-30 2023-02-03 西南电子技术研究所(中国电子科技集团公司第十研究所) Weak signal data enhancement method
CN114785423B (en) * 2022-06-20 2022-11-01 中国海洋大学三亚海洋研究院 Method for building underwater laser digital information transmission enhancement model
CN116403590B (en) * 2023-06-08 2023-08-18 青岛科技大学 Bionic signal processing method based on wavelet transformation and generation countermeasure network
CN118467943A (en) * 2024-05-22 2024-08-09 中国地质大学(武汉) A method and system for denoising power frequency harmonic noise based on semi-supervised learning
CN118379982A (en) * 2024-06-27 2024-07-23 武汉普惠海洋光电技术有限公司 High-frequency array environment noise reduction method and device

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105488466A (en) * 2015-11-26 2016-04-13 中国船舶工业系统工程研究院 Deep neural network and underwater sound target vocal print feature extraction method
CN107563428A (en) * 2017-08-25 2018-01-09 西安电子科技大学 Classification of Polarimetric SAR Image method based on generation confrontation network
CN109147817A (en) * 2018-08-29 2019-01-04 昆明理工大学 A kind of denoising audio feature extraction algorithm based on the limited Boltzmann machine that makes a variation
CN109978079A (en) * 2019-04-10 2019-07-05 东北电力大学 A kind of data cleaning method of improved storehouse noise reduction self-encoding encoder
CN109993710A (en) * 2019-03-20 2019-07-09 西北工业大学 An underwater image denoising method based on generative adversarial network
CN110490816A (en) * 2019-07-15 2019-11-22 哈尔滨工程大学 A kind of underwater Heterogeneous Information data noise reduction
CN111401236A (en) * 2020-03-16 2020-07-10 西北工业大学 Denoising method of underwater acoustic signal based on autoencoder neural network
CN111428853A (en) * 2020-01-16 2020-07-17 东华大学 A Negative Sample Adversarial Generation Method with Noise Learning

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11257259B2 (en) * 2017-08-15 2022-02-22 Siemens Healthcare Gmbh Topogram prediction from surface data in medical imaging
CN108765319B (en) * 2018-05-09 2020-08-14 大连理工大学 Image denoising method based on generation countermeasure network
CN110020685A (en) * 2019-04-09 2019-07-16 山东超越数控电子股份有限公司 A kind of preprocess method based on adaptive-filtering and limited Boltzmann machine, terminal and readable storage medium storing program for executing

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105488466A (en) * 2015-11-26 2016-04-13 中国船舶工业系统工程研究院 Deep neural network and underwater sound target vocal print feature extraction method
CN107563428A (en) * 2017-08-25 2018-01-09 西安电子科技大学 Classification of Polarimetric SAR Image method based on generation confrontation network
CN109147817A (en) * 2018-08-29 2019-01-04 昆明理工大学 A kind of denoising audio feature extraction algorithm based on the limited Boltzmann machine that makes a variation
CN109993710A (en) * 2019-03-20 2019-07-09 西北工业大学 An underwater image denoising method based on generative adversarial network
CN109978079A (en) * 2019-04-10 2019-07-05 东北电力大学 A kind of data cleaning method of improved storehouse noise reduction self-encoding encoder
CN110490816A (en) * 2019-07-15 2019-11-22 哈尔滨工程大学 A kind of underwater Heterogeneous Information data noise reduction
CN111428853A (en) * 2020-01-16 2020-07-17 东华大学 A Negative Sample Adversarial Generation Method with Noise Learning
CN111401236A (en) * 2020-03-16 2020-07-10 西北工业大学 Denoising method of underwater acoustic signal based on autoencoder neural network

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
Boosting Noise Robustness of Acoustic Model via Deep Adversarial Training;Bin Liu 等;《2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)》;20180913;第5034-5038页 *
Spectra-GANs: A New Automated Denoising Method for Low-S/N Stellar Spectra;Minglei Wu 等;《IEEE Access》;20200605;第8卷;第107912-107926页 *
The research of underwater target recognition method based on deep learning;Yuechao Chen 等;《2017 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC)》;20180101;第1-5页 *
基于Adversarial的生成式算法研究;罗金铭;《中国硕士学位论文全文数据库 信息科技专辑》;20190215;第2019年卷(第2期);I138-1093 *
生成对抗网络的改进方法与超分辨率图像重建研究;李艳春;《中国博士学位论文全文数据库 信息科技辑》;20200115;第2020年卷(第1期);I138-77 *

Also Published As

Publication number Publication date
CN112215054A (en) 2021-01-12

Similar Documents

Publication Publication Date Title
CN112215054B (en) Depth generation countermeasure method for denoising underwater sound signal
CN109919204B (en) A Deep Learning Clustering Method for Noisy Images
CN113723295A (en) Face counterfeiting detection method based on image domain frequency domain double-flow network
CN110334580A (en) The equipment fault classification method of changeable weight combination based on integrated increment
CN112307714A (en) Character style migration method based on double-stage deep network
CN109800785B (en) Data classification method and device based on self-expression correlation
CN112562698B (en) A defect diagnosis method for power equipment based on fusion of sound source information and thermal imaging features
CN110490265B (en) Image steganalysis method based on double-path convolution and feature fusion
CN114283287A (en) A Robust Domain Adaptive Image Learning Method Based on Self-training Noise Label Correction
CN112464004A (en) Multi-view depth generation image clustering method
CN111681188B (en) Image Deblurring Method Based on Combining Image Pixel Prior and Image Gradient Prior
CN109859310B (en) A model that can be used to generate MR images and its establishment method
CN114821204A (en) A semi-supervised learning image classification method and system based on meta-learning embedding
CN110321777A (en) A kind of face identification method based on the sparse denoising self-encoding encoder of stack convolution
CN113850013A (en) Ship radiation noise classification method
CN111160327B (en) An Expression Recognition Method Based on Lightweight Convolutional Neural Network
CN113628615B (en) Speech recognition method, device, electronic device and storage medium
CN112163605A (en) Multi-domain image translation method based on attention network generation
CN117036901A (en) Small sample fine adjustment method based on visual self-attention model
CN110766083A (en) A feature fusion based Alexnet mural image classification method
CN115861779A (en) Unbiased scene graph generation method based on effective feature representation
CN116776093A (en) An EEG signal data amplification algorithm based on improved generative adversarial network
CN116884090A (en) A cross-environment human activity recognition method based on dual adversarial networks
CN116434758A (en) Voiceprint recognition model training method and device, electronic equipment and storage medium
CN115811304A (en) A Method for Eliminating Ocean Impulse Noise Based on Conditional Variational Autoencoder Network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant