CN108597531B - A method to improve dual-channel blind signal separation by multi-source activity detection - Google Patents
A method to improve dual-channel blind signal separation by multi-source activity detection Download PDFInfo
- Publication number
- CN108597531B CN108597531B CN201810265485.5A CN201810265485A CN108597531B CN 108597531 B CN108597531 B CN 108597531B CN 201810265485 A CN201810265485 A CN 201810265485A CN 108597531 B CN108597531 B CN 108597531B
- Authority
- CN
- China
- Prior art keywords
- channel
- signal
- sound source
- power
- output signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000000694 effects Effects 0.000 title claims abstract description 43
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000000926 separation method Methods 0.000 title claims abstract description 30
- 238000001514 detection method Methods 0.000 title claims abstract description 20
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 19
- 239000011159 matrix material Substances 0.000 claims description 18
- 230000004044 response Effects 0.000 claims description 8
- 230000003595 spectral effect Effects 0.000 claims description 7
- 238000006073 displacement reaction Methods 0.000 claims description 5
- 238000011478 gradient descent method Methods 0.000 claims description 4
- 238000004321 preservation Methods 0.000 claims description 4
- 230000008569 process Effects 0.000 claims description 4
- 238000012937 correction Methods 0.000 claims description 2
- 230000001131 transforming effect Effects 0.000 claims 2
- 230000009977 dual effect Effects 0.000 claims 1
- 238000000638 solvent extraction Methods 0.000 claims 1
- 238000012545 processing Methods 0.000 abstract description 8
- 238000004364 calculation method Methods 0.000 description 5
- 230000001364 causal effect Effects 0.000 description 4
- 238000012880 independent component analysis Methods 0.000 description 4
- 230000006872 improvement Effects 0.000 description 3
- 239000000203 mixture Substances 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 230000037433 frameshift Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 description 1
- 101001120757 Streptococcus pyogenes serotype M49 (strain NZ131) Oleate hydratase Proteins 0.000 description 1
- 229940083712 aldosterone antagonist Drugs 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000011423 initialization method Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 230000014759 maintenance of location Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000008263 repair mechanism Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0272—Voice signal separating
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L2021/02087—Noise filtering the noise being separate speech, e.g. cocktail party
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Circuit For Audible Band Transducer (AREA)
Abstract
本发明公开了一种通过多声源活动检测来改进双通道盲信号分离的方法。该方法基于双通道TRINICON算法进行盲信号分离,将初步的处理前后的功率进行对比,若一个输出通道的信号功率明显低于另一个输出通道,则可判断这段信号中要抑制的目标声源处于活动状态,依此可对各段数据中各个目标声源是否处于活动状态进行判断。利用多声源活动判断的结果对TRINICON算法进行修正,使用目标声源活动的数据来更新滤波器系数,从而达到去除干扰、提高语音分离的性能的目的。本发明的方法能有效地改善在间断交错混合及稀疏混合的场景中TRINICON方法的分离性能。
The invention discloses a method for improving dual-channel blind signal separation through multi-sound source activity detection. This method performs blind signal separation based on the dual-channel TRINICON algorithm, and compares the power before and after preliminary processing. If the signal power of one output channel is significantly lower than that of the other output channel, the target sound source to be suppressed in this signal can be determined. In the active state, according to this, it can be judged whether each target sound source in each segment of data is in the active state. The TRINICON algorithm is modified by using the results of multi-sound source activity judgment, and the filter coefficients are updated with the data of the target sound source activity, so as to achieve the purpose of removing interference and improving the performance of speech separation. The method of the invention can effectively improve the separation performance of the TRINICON method in the scene of discontinuous interlaced mixing and sparse mixing.
Description
技术领域technical field
本发明涉及盲信号分离的技术领域,具体涉及一种基于双通道系统的通过多声源活动检测来改进盲信号分离性能的方法,在基于TRINICON(Triple-N ICA forconvolutive mixtures)结构的频域盲信号分离过程中加入多声源活动检测的算法。The invention relates to the technical field of blind signal separation, in particular to a method for improving blind signal separation performance through multi-sound source activity detection based on a dual-channel system. An algorithm for multi-source activity detection is added to the signal separation process.
背景技术Background technique
分离未知时间序列的卷积混合的问题在多个领域都有重要应用。一个重要的例子是所谓的鸡尾酒会问题,即将单个语音信号在混响的声学环境中从多个说话者的混合声中提取出来。由于混响的存在,这个分离问题的源信号在被传声器阵列采集前,先通过一个线性多输入多输出(MIMO)系统进行滤波。盲信号分离(BSS)是利用传声器阵列,不需要先验信息,基于不同声源信号相互统计独立的基本假设,即可将多个声源的信号进行分离的算法(S.Makino,H.Sawada,and T.W.Lee,Blind Speech Separation.Springer Netherlands,2007,pp.169-192.)。The problem of separating convolutional mixtures of unknown time series has important applications in several fields. An important example is the so-called cocktail party problem, where a single speech signal is extracted from the mixture of multiple speakers in a reverberant acoustic environment. Due to the presence of reverberation, the source signal of this separation problem is filtered through a linear multiple-input multiple-output (MIMO) system before being collected by the microphone array. Blind Signal Separation (BSS) is an algorithm that uses a microphone array to separate the signals of multiple sound sources without prior information and based on the basic assumption that the signals of different sound sources are statistically independent of each other (S. Makino, H. Sawada). , and T.W. Lee, Blind Speech Separation. Springer Netherlands, 2007, pp. 169-192.).
基于独立分量分析(ICA)的盲信号分离是从干扰语音信号中提取所需语音信号的有效方法。但是频域ICA方法面临各频点排序模糊的问题,这需要基于频率间相关性信息的适当补充修复机制来修正。目前,该问题可以通过将ICA扩展到多变量情况,进行独立矢量分析(IVA)算法来解决,或者利用基于宽带标准的TRINICON方案来解决。TRINICON方法可以使用二阶统计量(SOS)在频域中有效地实现。Blind signal separation based on independent component analysis (ICA) is an effective method to extract desired speech signals from interfering speech signals. However, the frequency domain ICA method faces the problem of ambiguous ordering of each frequency point, which needs to be corrected by an appropriate supplementary repair mechanism based on the correlation information between frequencies. Currently, this problem can be solved by extending ICA to the multivariate case, performing the Independent Vector Analysis (IVA) algorithm, or by utilizing the TRINICON scheme based on the broadband standard. The TRINICON method can be efficiently implemented in the frequency domain using second order statistics (SOS).
频域的离线TRINICON方法通常在声源进行连续混合时具有最佳性能。在实际应用中,语音信号有时会间断检错或者稀疏地进行混合,而离线的算法不考虑语音的活动状态,将每一段语音都平等地计入计算中,这样会导致性能恶化,尤其是当各个声源处于传声器阵列中线的一侧时,需要生成非因果滤波器。Offline TRINICON methods in the frequency domain generally have the best performance when the sound sources are continuously mixed. In practical applications, the speech signal sometimes detects errors intermittently or is sparsely mixed, and the offline algorithm does not consider the activity state of the speech and counts each segment of speech equally into the calculation, which will lead to performance degradation, especially when Acausal filters need to be generated when each sound source is on one side of the midline of the microphone array.
发明内容SUMMARY OF THE INVENTION
本发明的目的是为了提高频域TRINICON算法在间断交错混合及稀疏混合的情景下的性能,提供一种通过多声源活动检测来改进双通道盲信号分离的方法。The purpose of the present invention is to improve the performance of the frequency-domain TRINICON algorithm in the scenarios of discontinuous interlaced mixing and sparse mixing, and to provide a method for improving dual-channel blind signal separation through multi-sound source activity detection.
本发明为解决上述技术问题采取的技术方案是:The technical scheme that the present invention takes for solving the above-mentioned technical problems is:
一种通过多声源活动检测来改进双通道盲信号分离的方法,包括如下步骤:A method for improving dual-channel blind signal separation through multi-sound source activity detection, comprising the following steps:
(1)对频域TRINICON算法中的滤波器矩阵参数进行初始化;(1) Initialize the filter matrix parameters in the frequency domain TRINICON algorithm;
(2)传声器接收到的声源信号输入双通道系统,将每个通道的输入信号分块,再用短时傅立叶变换将每一块的输入信号变换到频域;(2) The sound source signal received by the microphone is input into the two-channel system, the input signal of each channel is divided into blocks, and the short-time Fourier transform is used to transform the input signal of each block into the frequency domain;
(3)由Y (k)=X (k) W (k)计算每一块的输出信号,其中Y (k)、X (k)和W (k)分别为频域的输出信号、频域的输入信号及频域的滤波器系数矩阵,上标(k)为短时离散傅立叶变换中频点的序号;然后计算各输出信号的功率谱密度矩阵Φyy;(3) Calculate the output signal of each block by Y (k) = X (k) W (k) , where Y (k) , X (k) and W (k) are the output signal in the frequency domain, the The filter coefficient matrix of the input signal and the frequency domain, the superscript (k) is the sequence number of the intermediate frequency point of the short-time discrete Fourier transform; then calculate the power spectral density matrix Φ yy of each output signal;
(4)按照自然梯度下降法更新滤波器系数,使用逆短时傅立叶变换将滤波器系数矩阵变换回时域并将时域的滤波器系数大于滤波器长度减1的部分置零;(4) update the filter coefficients according to the natural gradient descent method, use the inverse short-time Fourier transform to transform the filter coefficient matrix back to the time domain and set the filter coefficients in the time domain to zero if the filter coefficients are greater than the filter length minus 1;
(5)使用步骤(4)得到的滤波器系数,重复第(3)和第(4)步骤,直至达到最大迭代次数;(5) using the filter coefficients obtained in step (4), repeat steps (3) and (4) until the maximum number of iterations is reached;
(6)得到初步的滤波器系数收敛结果,利用重叠保留法对输入信号进行卷积得到第一通道的初步输出信号和第二通道的初步输出信号;(6) obtain preliminary filter coefficient convergence result, utilize overlapping retention method to convolve the input signal to obtain the preliminary output signal of the first channel and the preliminary output signal of the second channel;
(7)先根据以下两个公式分别计算两个通道中各块的输出和输入功率:(7) Calculate the output and input power of each block in the two channels according to the following two formulas:
其中Ex(m)为输入功率,Eyp(m)为初步输出信号的功率,m为信号分块的序号,ku和kl分别表示计入计算的频点的上限和下限,下标p是输出通道的序号,下标1和2分别代表第一通道和第二通道;where E x (m) is the input power, E yp (m) is the power of the preliminary output signal, m is the sequence number of the signal block, ku and k l represent the upper and lower limits of the frequency points included in the calculation, respectively, and the subscript p is the serial number of the output channel, the
然后对每一块信号的活动状态进行判断,并得到修正的权重函数;Then judge the active state of each signal, and get the revised weight function;
(8)根据步骤(1)重新对权重函数进行初始化,修正后的梯度表达式为:(8) Re-initialize the weight function according to step (1), and the corrected gradient expression is:
其中in
Nsig是整个信号的块数,εp(m)是修正权重,当第m块中第p个声源的状态为活动时,εp(m)等于1,否则,εp(m)等于0;N sig is the block number of the whole signal, ε p (m) is the modification weight, when the state of the p-th sound source in the m-th block is active, ε p (m) is equal to 1, otherwise, ε p (m) is equal to 0;
根据以上修正后的梯度表达式,重复第(3)和第(4)步骤,即重新进行一次迭代,这次迭代中只用第一通道要抑制的目标声源活动的数据更新第一通道的滤波器系数,只用第二通道要抑制的目标声源活动的数据更新第二通道的滤波器系数,直到达到最大迭代次数;According to the above modified gradient expression, repeat steps (3) and (4), that is, perform an iteration again. In this iteration, only the data of the target sound source activity to be suppressed by the first channel is used to update the data of the first channel. filter coefficient, update the filter coefficient of the second channel with only the data of the target sound source activity to be suppressed by the second channel until the maximum number of iterations is reached;
(9)得到最终的滤波器系数收敛结果,利用重叠保留法对输入信号进行卷积得到两个通道的最终输出信号。(9) The final filter coefficient convergence result is obtained, and the input signal is convolved by the overlap preservation method to obtain the final output signal of the two channels.
本发明的检测方法可近似识别每个源的活动,提取出声源各自的有效活动帧并根据声源活动检测的结果对滤波器更新过程进行调整,能有效地改善在间断交错混合及稀疏混合的场景中TRINICON方法的分离性能。The detection method of the present invention can approximately identify the activity of each source, extract the respective effective activity frames of the sound source, and adjust the filter update process according to the result of the sound source activity detection, which can effectively improve the performance of intermittent interlaced mixing and sparse mixing. The separation performance of the TRINICON method in the scene.
附图说明Description of drawings
图1为本发明实施例中的信号模型图:(a)仅需要因果滤波器,(b)需要因果和非因果滤波器。FIG. 1 is a signal model diagram in an embodiment of the present invention: (a) only causal filters are required, (b) causal and non-causal filters are required.
图2为本发明实施例中多声源活动检测的流程图。FIG. 2 is a flowchart of multi-sound source activity detection in an embodiment of the present invention.
图3为本发明加入多声源活动检测的频域盲信号分离算法的流程图。FIG. 3 is a flow chart of the frequency domain blind signal separation algorithm with multi-sound source activity detection added in the present invention.
图4为本发明实施例中最终的优化滤波器的空间响应图:(a)普通频域TRINICON输出通道1,(b)普通频域TRINICON输出通道2,(c)加入多声源活动检测的频域TRINICON输出通道1,(d)加入多声源活动检测的频域TRINICON输出通道2。Fig. 4 is the spatial response diagram of the final optimized filter in the embodiment of the present invention: (a) common frequency domain
具体实施方式Detailed ways
下面结合附图和实施例对本发明进一步说明。The present invention will be further described below in conjunction with the accompanying drawings and embodiments.
1、双通道频域TRINICON算法1. Dual-channel frequency domain TRINICON algorithm
本发明适用于双通道的传声器阵列系统,假设声源个数为一个或两个,算法的信号模型如图1(只有一个信号时则模型中两个声源只有一个发声)。The present invention is suitable for a two-channel microphone array system, assuming that the number of sound sources is one or two, and the signal model of the algorithm is shown in Figure 1 (when there is only one signal, only one of the two sound sources in the model emits sound).
TRINICON算法的代价函数为:The cost function of the TRINICON algorithm is:
其中,是第q通道的D维的多元概率密函数的估计,是所有输出通道(这里Q=2)的联合概率密度的估计,m是块序号,j=0,…,N-1长度为N的块中时移的序号,β(i,m)是一个权重函数,根据归一化。in, is the estimate of the D-dimensional multivariate probability density function of the qth channel, is the estimate of the joint probability density of all output channels (here Q=2), m is the block sequence number, j=0,...,N-1 The sequence number of the time shift in blocks of length N, β(i,m) is a weight function, according to Normalized.
使用SOS的TRINICON的代价函数表示为(H.Buchner,R.Aichner,andW.Kellermann,"A generalization of blind source separation algorithms forconvolutive mixtures based on second-order statistics,"IEEE Transactions onSpeech&Audio Processing,vol.13,no.1,pp.120-134,2004):The cost function of TRINICON using SOS is expressed as (H. Buchner, R. Aichner, and W. Kellermann, "A generalization of blind source separation algorithms for convolutive mixtures based on second-order statistics," IEEE Transactions on Speech&Audio Processing, vol. 13, no .1, pp. 120-134, 2004):
其中bdiag操作是将一个分块矩阵的非对角子矩阵都置零。The bdiag operation is to zero out the off-diagonal submatrices of a block matrix.
滤波器系数W根据自然梯度下降的方法来进行更新:The filter coefficients W are updated according to the natural gradient descent method:
其中in
jiter是迭代的次数,μ为步长。j iter is the number of iterations, μ is the step size.
在频域的方法中,滤波器矩阵和信号分别表示为:In the frequency domain method, the filter matrix and the signal are expressed as:
和and
其中L是滤波器的长度,F4L×4L是傅立叶变换矩阵。带下划线的变量为频域的变量。where L is the length of the filter and F 4L×4L is the Fourier transform matrix. Variables underlined are variables in the frequency domain.
输出信号的功率谱密度矩阵Φyy和输入信号的功率谱密度矩阵Φxx分别表示为The power spectral density matrix Φ yy of the output signal and the power spectral density matrix Φ xx of the input signal are expressed as
和and
其中不同维度的约束矩阵G的作用是防止各个频率点之间的去耦以避免频域盲信号分离中的排序模糊问题和圆周卷积效应。应用到频域计算时,将G近似为单位阵,则可以在每个频点上单独进行迭代(式中省略了块序号m):Among them, the role of the constraint matrix G of different dimensions is to prevent the decoupling between each frequency point to avoid the ordering ambiguity problem and the circular convolution effect in the frequency domain blind signal separation. When applied to the frequency domain calculation, if G is approximated as a unit matrix, it can be iterated independently at each frequency point (the block number m is omitted in the formula):
其中上标(k)是STFT中频率点的序号。每步迭代完成后,分界矩阵必须转换回时域,并将l>L-1的wqq,l的值置零,以避免圆周卷积效应。The superscript (k) is the serial number of the frequency point in the STFT. After each iteration, the demarcation matrix must be converted back to the time domain and the value of w qq,l for l>L-1 is zeroed to avoid the circular convolution effect.
2、加入多声源活动检测的双通道频域TRINICON方法2. Add dual-channel frequency domain TRINICON method for multi-sound source activity detection
在离线的TRINICON方法中,当信号不是连续混合,尤其是当无语音段或者只有一个声源活动的时间较长时,对功率谱密度矩阵Φ的估计会有偏差,从而错误地引导了分离系统收敛的方向,导致语音分离的效果不佳。这种情况在图1(b)所示的需要非因果滤波器的状况下更加明显。In the offline TRINICON method, when the signals are not continuously mixed, especially when there is no speech segment or only one sound source is active for a long time, the estimation of the power spectral density matrix Φ is biased, thus misleading the separation system Convergence direction, resulting in poor speech separation effect. This situation is more pronounced in the situation shown in Figure 1(b) where acausal filters are required.
本发明将离线TRINICONO执行两次,并使用第一次的收敛的初步结果来进行一个粗糙的多声源活动估计,并根据估计的结果调整第二次的迭代过程,从而得到改进的语音分离效果。对于需要非因果滤波器的情况,一般的算法采用一个位移L/2的单位冲激响应来进行初始化,但是在频域算法中这种初始化的方法收敛特性较差,本实施例采用的方法是采用一个位移点数略微大于两个传声器之间的最大声程差的单位冲激响应来对滤波器进行初始化。The present invention executes off-line TRINICONO twice, uses the first convergence preliminary result to perform a rough multi-sound source activity estimation, and adjusts the second iterative process according to the estimated result, thereby obtaining an improved speech separation effect . For the case where a non-causal filter is required, the general algorithm uses a unit impulse response with a displacement of L/2 for initialization, but in the frequency domain algorithm, this initialization method has poor convergence characteristics. The method used in this embodiment is: The filter is initialized with a unit impulse response with a displacement point slightly larger than the maximum sound path difference between the two microphones.
首先要计算各块中输入和输出的功率:First calculate the power input and output in each block:
和and
其中ku和kl是计入计算的频点的上下限(下标u是upper bound的缩写,表示上限,下标l是lower bound的缩写,表示下限),可以根据信号的频谱特性自行调整以更好地估计信号功率和节省计算量。需要注意的是在低频段盲信号分离的效果一般很差,所以低频段一般不计入计算。Among them, k u and k l are the upper and lower limits of the frequency points included in the calculation (the subscript u is the abbreviation of upper bound, indicating the upper limit, and the subscript l is the abbreviation of lower bound, indicating the lower limit), which can be adjusted according to the spectral characteristics of the signal. To better estimate signal power and save computation. It should be noted that the effect of blind signal separation in the low frequency band is generally poor, so the low frequency band is generally not included in the calculation.
计算了各个通道的输入和输出功率后,根据第一次处理前后的功率对比,可以对各块中两个声源的活动情况进行一个大致的判断。具体的判断步骤见图2。假设在输出通道y1中声源s1被抑制。首先如果输入信号的功率小于设定的最低功率阈值Emin(该阈值可以通过MCRA等方法估计的平稳噪声功率来确定),则判断为两个声源都不发声,并设定ε1(m)=0,ε2(m)=0。如果功率大于阈值,则判断为至少有一个声源发声,进入下一步判断:若通道y1处理前后的功率比大于通道y2处理前后的功率比乘上系数λ,则认为声源s1不活动,声源s2处于活动状态,并设定ε1(m)=0,ε2(m)=1;若通道y2处理前后的功率比大于通道y1处理前后的功率比乘上系数λ,则认为声源s2不活动,声源s1处于活动状态,并设定ε1(m)=1,ε2(m)=0;如果以上情况都不符合,则认为两个声源都活动,并设定ε1(m)=1,ε2(m)=1。其中系数λ可以根据具体声源和使用场景进行调整,一般是大于1的实数。After calculating the input and output power of each channel, according to the power comparison before and after the first processing, a rough judgment can be made on the activity of the two sound sources in each block. The specific judgment steps are shown in Figure 2. Assume that the sound source s1 is suppressed in the output channel y1. First, if the power of the input signal is less than the set minimum power threshold E min (the threshold can be determined by the stationary noise power estimated by methods such as MCRA), then it is judged that the two sound sources do not emit sound, and ε 1 (m )=0, ε 2 (m)=0. If the power is greater than the threshold, it is judged that there is at least one sound source emitting sound, and the next step is to judge: if the power ratio before and after processing of channel y1 is greater than the power ratio before and after processing of channel y2 multiplied by the coefficient λ, then it is considered that the sound source s1 is inactive and the sound Source s2 is active, and set ε 1 (m)=0, ε 2 (m)=1; if the power ratio before and after processing of channel y2 is greater than the power ratio before and after processing of channel y1 multiplied by the coefficient λ, it is considered that the sound source s2 is inactive, sound source s1 is active, and set ε 1 (m)=1, ε 2 (m)=0; if none of the above conditions are met, consider both sound sources active, and set ε 1 (m)=1, ε 2 (m)=1. The coefficient λ can be adjusted according to the specific sound source and usage scenario, and is generally a real number greater than 1.
对声源的活动状态进行判断后,式(9)的梯度被修正为After judging the active state of the sound source, the gradient of equation (9) is modified as
其中in
式中Nsig是整个信号的块数,εp(m)是修正权重,当第m块中第p个声源的状态为活动时,εp(m)等于1,否则,εp(m)等于0。where N sig is the block number of the entire signal, ε p (m) is the correction weight, when the state of the p-th sound source in the m-th block is active, ε p (m) is equal to 1, otherwise, ε p (m ) is equal to 0.
使用修正后的梯度表达式(12)来更新系数,即只使用声源s1活动的数据来更新输出通道y1的滤波器,只使用使用声源s2活动的数据来更新输出通道y2的滤波器。这样可选取有效的帧分别对两个滤波器进行更新,改善收敛性能,有效地提高语音分离的效果。需要注意的是上述的多声源活动检测在初始的信扰比为10dB以下时会比较准确,但也并不是完全精确的,在本实施例中只需要大致地判断各个声源活动的时间段即可。The coefficients are updated using the modified gradient expression (12), that is, the filter of output channel y1 is updated using only the data of sound source s1 activity, and the filter of output channel y2 is updated using only the data of sound source s2 activity. In this way, an effective frame can be selected to update the two filters respectively, so as to improve the convergence performance and effectively improve the effect of speech separation. It should be noted that the above-mentioned multi-sound source activity detection is relatively accurate when the initial signal-to-interference ratio is below 10dB, but it is not completely accurate. In this embodiment, it is only necessary to roughly determine the time period of each sound source activity. That's it.
调整后整个算法的流程如图3。使用传声器接收到的输入信号x1,x2,先使用一般的频域TRINICON方法进行一次语音分离,然后使用第一次频域TRINICON的分离结果,按前文所述的多声源活动检测方式,确定两个声源各自的活动时间段;接着以两个声源各自的活动时间段分别更新两个滤波器系数,进行调整后的第二次频域TRINICON的语音分离,得到最终的输出信号y1和y2。The flow of the whole algorithm after adjustment is shown in Figure 3. Using the input signals x1 and x2 received by the microphone, first use the general frequency domain TRINICON method to perform a speech separation, and then use the separation result of the first frequency domain TRINICON, according to the multi-sound source activity detection method described above, determine the two. The respective active time periods of the two sound sources; then update the two filter coefficients with the respective active time periods of the two sound sources, and perform the adjusted second frequency-domain TRINICON speech separation to obtain the final output signals y1 and y2 .
图4展示了在一组仿真数据中使用普通频域TRINICON方法和采用加入了多声源活动检测的频域TRINICON方法各自收敛得到的滤波器的空间响应。声源位置为-70度和-15度,房间混响时间0.2秒,传声器阵列单元间距10厘米。Figure 4 shows the spatial responses of the filters converged using the normal frequency-domain TRINICON method and the frequency-domain TRINICON method with multi-source activity detection added to a set of simulated data. The sound source positions are -70 degrees and -15 degrees, the room reverberation time is 0.2 seconds, and the microphone array unit spacing is 10 cm.
可以看到加入了多声源活动检测后,滤波器优化得到的谷点更加准确,且深度更深,对声源的抑制更加明显。It can be seen that after adding multi-sound source activity detection, the valley points obtained by filter optimization are more accurate and deeper, and the suppression of sound sources is more obvious.
表1给出了多组数据的仿真结果,对比了不同情景下本发明对分离效果的提升。仿真的输入信号由纯净的语音信号经过Image Model工具声生成的房间冲激响应卷积得到。测试的指标有信号干扰比(signal to interference ratio,SIR)和信号失真比(signalto distortion ratio,SDR),测试工具为BSSEVAL。表中给出的评分为4次实验下的两个通道的评分的平均。SIR越高,说明干扰声被抑制得越好,目标语音更纯净;SDR越高,说明输出的语音对比纯净语音的失真越小。Table 1 presents the simulation results of multiple sets of data, and compares the improvement of the separation effect of the present invention in different scenarios. The simulated input signal is obtained by convolving the room impulse response generated by the sound of the Image Model tool with the pure speech signal. The indicators tested are signal to interference ratio (SIR) and signal to distortion ratio (SDR), and the test tool is BSSEVAL. The scores given in the table are the average of the scores of the two channels under 4 experiments. The higher the SIR, the better the interference sound is suppressed and the purer the target speech; the higher the SDR, the smaller the distortion of the output speech compared to the pure speech.
可以看到在大部分情况下本发明对分离效果的提升是非常明显的,SIR的提升可以达到5dB至10dB以上,同时可以保持音质不严重失真。It can be seen that the improvement of the separation effect of the present invention is very obvious in most cases, the improvement of SIR can reach more than 5dB to 10dB, and at the same time, the sound quality can be kept from serious distortion.
表1不同情景下普通方法和本发明处理结果的SIR和SDR评分Table 1 SIR and SDR scores of the common method and the processing results of the present invention under different scenarios
实施例Example
1、测试条件:1. Test conditions:
设定房间大小为6m×6m×4m,混响时间为250ms,传声器阵列单元间距为10cm,放置于靠近房间中心的位置。声源离阵列中心1.5m,分别位于-70度和-15度方向。两个声源的纯净信号来自TIMIT数据库,各自发声的时长占录制信号总时长的60%,其中有约总时长三分之一的时长两个声源同时发声。两个声源的信号的信扰比(SIR)设置为0dB,即发声功率基本相同。另外,整个信号中加入了比声源信号低30dB的白噪声。信号的采样频率为16000Hz。使用Image Model根据上述条件生成房间冲激响应并用其与纯净语音卷积得到输入信号。The size of the room is set to 6m×6m×4m, the reverberation time is 250ms, the spacing of the microphone array units is 10cm, and it is placed close to the center of the room. The sound source is 1.5m away from the center of the array, at -70 degrees and -15 degrees, respectively. The pure signals of the two sound sources come from the TIMIT database, and the sounding time of each of them accounts for 60% of the total duration of the recorded signal, of which about one-third of the total duration is the two sound sources sounding simultaneously. The signal-to-interference ratio (SIR) of the signals of the two sound sources is set to 0 dB, that is, the sound power is basically the same. In addition, white noise 30dB lower than the sound source signal is added to the whole signal. The sampling frequency of the signal is 16000Hz. Use the Image Model to generate the room impulse response according to the above conditions and use it to convolve with the pure speech to obtain the input signal.
2、算法流程:2. Algorithm process:
1)算法初始化:将(5)式中的w11、w22设为位移点数为10个点的移位单位冲激响应(即w11,10=1,其余值为0),w12、w21设为零矢量;步长μ设为0.02,最大迭代次数设置为200,滤波器长度L设置为1024。帧移设为L的一半。权重函数设置为1/Nsig。1) Algorithm initialization: set w 11 and w 22 in formula (5) as the displacement unit impulse response with 10 displacement points (that is, w 11,10 =1, and the remaining values are 0), w 12 , w 21 is set to the zero vector; the step size μ is set to 0.02, the maximum number of iterations is set to 200, and the filter length L is set to 1024. Frame shift is set to half of L. The weight function is set to 1/N sig .
2)输入传声器接收到的信号x1、x2,按照(6)式将输入信号分成长度为4L,帧移为L/2的块,并用短时傅立叶变换变换到频域。2) Input the signals x 1 and x 2 received by the microphone, divide the input signal into blocks with a length of 4L and a frame shift of L/2 according to formula (6), and transform them into the frequency domain with short-time Fourier transform.
3)由Y (k)=X (k) W (k)计算每一块的输出信号,并根据(7)式计算输出信号的功率谱密度矩阵Φyy。3) Calculate the output signal of each block by Y (k) = X (k) W (k) , and calculate the power spectral density matrix Φ yy of the output signal according to formula (7).
4)根据(3)式和(9)式,按照自然梯度下降法更新滤波器系数W,使用逆短时傅立叶变换变换回时域并将w大于L-1的部分置零。4) According to equations (3) and (9), update the filter coefficient W according to the natural gradient descent method, use the inverse short-time Fourier transform to transform back to the time domain and set the part of w greater than L-1 to zero.
5)使用更新得到的滤波器系数W,重复第3)步和第4)步,直至达到最大迭代次数。5) Using the updated filter coefficients W , repeat steps 3) and 4) until the maximum number of iterations is reached.
6)得到初步的滤波器系数收敛结果W,利用重叠保留法对输入信号进行卷积得到两个通道的初步输出信号y10和y20。6) Obtain a preliminary filter coefficient convergence result W , and convolve the input signal with the overlap-preserving method to obtain preliminary output signals y 10 and y 20 of two channels.
7)根据(10)式及(11)式分别计算两个通道各块的输出和输入功率,并根据图2的流程,对每一块信号的活动状态进行判断,并得到修正的权重函数B(m)。7) Calculate the output and input power of each block of the two channels according to formula (10) and formula (11) respectively, and according to the process of Figure 2, judge the active state of each block signal, and obtain the revised weight function B ( m).
8)根据步骤1)重新对权重函数进行初始化,将步骤4)中的(9)式改为(12)式,并重复第3)步和第4)步,即重新进行一次迭代,这次迭代中只用声源s1活动的数据更新通道y1的滤波器系数,只用声源s2活动的数据更新通道y2的滤波器系数,直到达到最大迭代次数。8) Re-initialize the weight function according to step 1), change formula (9) in step 4) to formula (12), and repeat steps 3) and 4), that is, perform an iteration again, this time In the iteration, only the active data of the sound source s1 is used to update the filter coefficient of the channel y1, and only the active data of the sound source s2 is used to update the filter coefficient of the channel y2 until the maximum number of iterations is reached.
9)得到最终的滤波器系数收敛结果W,利用重叠保留法对输入信号进行卷积得到最终的两个通道的输出信号y1和y2。9) The final filter coefficient convergence result W is obtained, and the input signal is convolved by the overlap preservation method to obtain the final output signals y 1 and y 2 of the two channels.
Claims (4)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810265485.5A CN108597531B (en) | 2018-03-28 | 2018-03-28 | A method to improve dual-channel blind signal separation by multi-source activity detection |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810265485.5A CN108597531B (en) | 2018-03-28 | 2018-03-28 | A method to improve dual-channel blind signal separation by multi-source activity detection |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108597531A CN108597531A (en) | 2018-09-28 |
CN108597531B true CN108597531B (en) | 2021-05-28 |
Family
ID=63623822
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810265485.5A Active CN108597531B (en) | 2018-03-28 | 2018-03-28 | A method to improve dual-channel blind signal separation by multi-source activity detection |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108597531B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111178398B (en) * | 2019-12-13 | 2023-08-22 | 天翼商业保理有限公司 | Method, system, storage medium and device for detecting tampering of identity card image information |
CN111696573B (en) * | 2020-05-20 | 2023-04-28 | 北京地平线机器人技术研发有限公司 | Sound source signal processing method and device, electronic equipment and storage medium |
CN113823316B (en) * | 2021-09-26 | 2023-09-12 | 南京大学 | A speech signal separation method for positions close to the sound source |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1017253B1 (en) * | 1998-12-30 | 2012-10-31 | Siemens Corporation | Blind source separation for hearing aids |
US7222961B2 (en) * | 2002-08-05 | 2007-05-29 | Kestrel Corporation | Method for detecting a functional signal in retinal images |
CN100392723C (en) * | 2002-12-11 | 2008-06-04 | 索夫塔马克斯公司 | Speech processing system and method using independent component analysis under stability constraints |
US7464029B2 (en) * | 2005-07-22 | 2008-12-09 | Qualcomm Incorporated | Robust separation of speech signals in a noisy environment |
DE102006027673A1 (en) * | 2006-06-14 | 2007-12-20 | Friedrich-Alexander-Universität Erlangen-Nürnberg | Signal isolator, method for determining output signals based on microphone signals and computer program |
CN106023984A (en) * | 2016-04-28 | 2016-10-12 | 成都之达科技有限公司 | Speech recognition method based on car networking |
CN106887238B (en) * | 2017-03-01 | 2020-05-15 | 中国科学院上海微系统与信息技术研究所 | Sound signal blind separation method based on improved independent vector analysis algorithm |
-
2018
- 2018-03-28 CN CN201810265485.5A patent/CN108597531B/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN108597531A (en) | 2018-09-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Aichner et al. | A real-time blind source separation scheme and its application to reverberant and noisy acoustic environments | |
CN110010148B (en) | Low-complexity frequency domain blind separation method and system | |
CN108141691B (en) | Adaptive Reverberation Cancellation System | |
US20080228470A1 (en) | Signal separating device, signal separating method, and computer program | |
Lim et al. | Robust multichannel dereverberation using relaxed multichannel least squares | |
Wang et al. | A region-growing permutation alignment approach in frequency-domain blind source separation of speech mixtures | |
CN108597531B (en) | A method to improve dual-channel blind signal separation by multi-source activity detection | |
Crocco et al. | Room impulse response estimation by iterative weighted l 1-norm | |
US10679642B2 (en) | Signal processing apparatus and method | |
CN114220453A (en) | Multi-channel non-negative matrix decomposition method and system based on frequency domain convolution transfer function | |
Bourgeois et al. | Time-domain beamforming and blind source separation: speech input in the car environment | |
Crocco et al. | Estimation of TDOA for room reflections by iterative weighted l 1 constraint | |
CN112201276A (en) | Microphone array speech separation method based on TC-ResNet network | |
CN106024001A (en) | Method used for improving speech enhancement performance of microphone array | |
Málek et al. | Sparse target cancellation filters with application to semi-blind noise extraction | |
Dam et al. | Blind signal separation using steepest descent method | |
Mazur et al. | Robust room equalization using sparse sound-field reconstruction | |
CN115132162B (en) | An active noise control method based on sound source separation | |
Bertrand et al. | Distributed LCMV beamforming in wireless sensor networks with node-specific desired signals | |
Zeng et al. | Distributed delay and sum beamformer in regular networks based on synchronous randomized gossip | |
Wang et al. | Multichannel Linear Prediction-Based Speech Dereverberation Considering Sparse and Low-Rank Priors | |
Bertrand et al. | Distributed distortionless signal estimation in wireless acoustic sensor networks | |
Osterwise et al. | A comparison of BSS algorithms in harsh environments | |
Kinoshita et al. | Blind source separation using spatially distributed microphones based on microphone-location dependent source activities. | |
TWI356398B (en) |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |