[go: up one dir, main page]

CN110111807B - Microphone array-based indoor sound source following and enhancing method - Google Patents

Microphone array-based indoor sound source following and enhancing method Download PDF

Info

Publication number
CN110111807B
CN110111807B CN201910346841.0A CN201910346841A CN110111807B CN 110111807 B CN110111807 B CN 110111807B CN 201910346841 A CN201910346841 A CN 201910346841A CN 110111807 B CN110111807 B CN 110111807B
Authority
CN
China
Prior art keywords
frame
sound
signal
array
noise
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910346841.0A
Other languages
Chinese (zh)
Other versions
CN110111807A (en
Inventor
沈伟
何天豪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Science and Technology
Original Assignee
Nanjing University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Science and Technology filed Critical Nanjing University of Science and Technology
Priority to CN201910346841.0A priority Critical patent/CN110111807B/en
Publication of CN110111807A publication Critical patent/CN110111807A/en
Application granted granted Critical
Publication of CN110111807B publication Critical patent/CN110111807B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S5/00Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
    • G01S5/18Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using ultrasonic, sonic, or infrasonic waves
    • G01S5/20Position of source determined by a plurality of spaced direction-finders
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0216Noise filtering characterised by the method used for estimating noise
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0264Noise filtering characterised by the type of parameter measurement, e.g. correlation techniques, zero crossing techniques or predictive techniques
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0216Noise filtering characterised by the method used for estimating noise
    • G10L2021/02161Number of inputs available containing the signal or the noise to be suppressed
    • G10L2021/02166Microphone arrays; Beamforming

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Quality & Reliability (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • General Physics & Mathematics (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

本发明公开了一种基于麦克风阵列的室内声源跟随与增强方法,对每一路麦克风拾取的模拟声音信号进行抗混叠低通滤波,并对滤波后的模拟信号进行A/D采样和帧划分;对各路声音信号进行帧电平检测,将帧划分为声音帧信号和噪声帧信号,若为噪声帧,则进行存储,若为声音帧,则进行谱对消;选取参考路信号,根据参考路声音帧估计波长,并将各路声音帧与当前参考路声音帧进行互相关操作,进而估计波达角;根据估计的波长、波达角,结合麦克风阵列的几何形状,确定各路声音帧对应的权系数;对各路声音帧进行加权求和确定增强的语音数字信号,并对得到的语音数字信号进行D/A转化,生成模拟信号后传递给扬声器。本发明提升了系统的噪声抑制能力与声音保真性。

Figure 201910346841

The invention discloses an indoor sound source following and enhancement method based on a microphone array, which performs anti-aliasing low-pass filtering on the analog sound signal picked up by each microphone, and performs A/D sampling and frame division on the filtered analog signal. ; Perform frame level detection on each sound signal, divide the frame into sound frame signal and noise frame signal, if it is a noise frame, then store it, if it is a sound frame, then perform spectrum cancellation; select the reference road signal, according to Estimate the wavelength of the sound frame of the reference road, and perform the cross-correlation operation between the sound frame of each channel and the sound frame of the current reference road, and then estimate the angle of arrival; according to the estimated wavelength and angle of arrival, combined with the geometry of the microphone array, determine the sound of each channel The weight coefficient corresponding to the frame; the weighted summation of each sound frame to determine the enhanced voice digital signal, and the D/A conversion of the obtained voice digital signal to generate an analog signal and transmit it to the speaker. The present invention improves the noise suppression capability and sound fidelity of the system.

Figure 201910346841

Description

Microphone array-based indoor sound source following and enhancing method
Technical Field
The invention relates to a voice signal processing technology, in particular to an indoor sound source following and enhancing method based on a microphone array.
Background
In an indoor meeting or teaching, the personal speaking always is limited by the existing voice system, and the learning and communication process is very inconvenient. For example, in a voice system using a wireless microphone or a fixed-point microphone, although the voice of a target sound source can be accurately collected, the transmission distance limits the position of a speaker, and the maintenance cost of a battery and the microphone is high. The voice system based on the microphone array enlarges the reception range by installing the microphones at different indoor positions, removes the position limitation of a speaker, and is more convenient and safer to manage due to the adoption of stable power supply sources such as a socket and the like. But because the signal processing mode is simple, the pickup sound is easily interfered by interference sources in all directions, the speaker can not be focused on the most main speaker, the reality degree of the sound is ensured, and the self-healing capability is not provided, and if any one microphone is damaged, the whole system can also be paralyzed.
Disclosure of Invention
The invention provides an indoor sound source following and enhancing method based on a microphone array.
The technical solution for realizing the purpose of the invention is as follows: an indoor sound source following and enhancing method based on a microphone array comprises the following steps:
step 1, performing anti-aliasing low-pass filtering on analog sound signals picked up by each microphone, and performing A/D sampling and frame division on the filtered analog signals;
step 2, carrying out frame level detection on each path of sound signal, dividing a frame into a sound frame signal and a noise frame signal, if the sound frame is a noise frame, storing the sound frame signal, and if the sound frame is a sound frame, carrying out spectrum cancellation with the previous noise frame;
step 3, selecting a reference path signal, estimating a wavelength according to a reference path sound frame, and performing cross-correlation operation on each path of sound frame and a current reference path sound frame to further estimate a wave arrival angle;
step 4, determining a weight coefficient corresponding to each sound frame according to the estimated wave wavelength and arrival angle and by combining the geometric shape of the microphone array;
step 5, carrying out weighted summation on each path of sound frame to determine an enhanced voice digital signal, carrying out D/A conversion on the obtained voice digital signal, generating an analog signal and transmitting the analog signal to a loudspeaker;
and 6, repeating the steps 1-5, and processing the sound signal of the next frame until the sound pickup is finished.
Compared with the prior art, the invention has the remarkable advantages that: 1) according to the invention, after a sound source is positioned, airspace filtering and noise spectrum cancellation processing are carried out, so that the noise suppression capability and the sound fidelity of the system are improved; 2) the invention has strong self-healing capability, can still continue to work normally when any microphone is damaged, and ensures the overall performance of the system.
Drawings
Fig. 1 is a flow chart of an indoor sound source following and enhancing method based on a microphone array according to the present invention.
Fig. 2 is a diagram of a speech system for an indoor quaternary linear microphone array.
Fig. 3 is a diagram of a geometric mathematical model of fig. 2.
Fig. 4 is a schematic diagram of framing a digital sound signal.
Fig. 5 is a diagram of the result of cross-correlating the same sound frame picked up by two microphone units.
Fig. 6 is a diagram illustrating the relationship between the sound signal time delay Δ t and the angle of arrival θ between the microphones.
Fig. 7 is a comparison graph between the calculated value and the actual value of the angle of arrival for one example.
Fig. 8 is an array gain pattern, where (a) is the array gain pattern with a weight of 0 degrees and (b) is the array gain pattern with a weight of 30 degrees.
Fig. 9 is a graph comparing the processing effect of the indoor sound source following and enhancing method based on the microphone array on a certain sound in a room with the processing effect of the common method.
Detailed Description
The principles and aspects of the present invention are further described below in conjunction with the following figures and the detailed description.
Fig. 1 shows a flow of a method for indoor sound source following and enhancement based on a microphone array, and for convenience of explanation, a speech system composed of 4 element linear microphone arrays shown in fig. 2 is taken as an example, wherein a speaker is located at the left front of the array, and a speaker is directed at a student seat in front of the microphone, the method of the invention includes the following steps:
step 1, anti-aliasing filtering
Generally, the frequency of sound which can be perceived by human ears is 20KHz at most, and the sound above the frequency band is filtered before sampling, so that the sampling is prevented from causing spectrum aliasing. Since the sampling rate can be satisfied only by satisfying the nyquist sampling law, the sampling rate of 40.96KHz >2 × 20KHz can be adopted in the exemplary procedure of the present invention.
Step 2, A/D sampling and frame division
By utilizing the forward masking effect of human ears, as long as the system completes data processing and plays the enhanced sound signal within the specified time, the human ears cannot perceive the sound directly from a speaker, but can only perceive the sound played by a loudspeaker, and the experience of a meeting place cannot be influenced.
It is also important to note that the time for the system to process data is not too long, and the frame must be the unit of signal processing, and the divided frame duration should not exceed the masking effect time. For example, the typical human ear forward masking effect is as short as 5ms, and if the sampling rate is fsIf the energy efficiency response duration is within f s200 points, considering the requirement of the number of the fast Fourier transform points, can be taken
Figure BDA0002042585390000031
The point is an optimal frame, wherein
Figure BDA0002042585390000032
Representing an integer less than and closest to x. In addition, in practical situations, since the speaker sound is much larger than the sound source sound, the sound intensity masking effect can significantly prolong the time domain forward masking effect, so that the time for subsequent signal processing is completely sufficient for such a framing manner.
In addition, the frame interval should be smaller than the frame length to avoid the serious loss of the useful signal caused by the asynchronization of the sound frames of each path. An example of frame division with a frame length of 200 and a frame interval of 100 is shown in fig. 4, where the data points in the solid frame are the previous data frame and the data points in the dashed frame are the current data frame.
Step 3, frame type judgment
The level amplitude of the noise frame signal is small, and the level amplitude of the sound frame signal is large, so that the division is performed by level values, and the method is intuitive, simple and accurate. The specific judgment method is that each frame signal is calculated according to a certain rule to generate a judgment level, and the judgment level is compared with a threshold level so as to divide the frame into a sound frame or a noise frame. The noise frame is only FFT stored and the voice frame proceeds to step 4.
The operation rule and the threshold level are chosen freely and variously, and the operation rule adopted by the exemplary program of the present invention is to average absolute values of all points in a frame, and the threshold level is 1/10 of the full-scale level of the sound collection system.
Step 4, spectrum cancellation
For random noise, due to the randomness of the random noise, the superposition operation among all paths of frame signals cannot obviously enhance the noise; for coherent noise, due to its coherence, the addition enhances the sound signal and also enhances the coherent noise to an equal extent, so measures should be taken to suppress the coherent noise.
Because coherent noise has a relatively fixed frequency spectrum, the coherent noise in the array beam pointing space can be suppressed by subtracting the FFT of the sound frame signal and the FFT of the noise frame signal of the nearest section.
Step 5, wavelength estimation
The human voice signal is a short-time stationary signal, and in a conference or teaching environment, the voice signal when a human normally speaks can be regarded as a narrow-band signal. Therefore, the FFT result of the sound signal frame is analyzed, and the central point of the frequency band range exceeding the set threshold value in the FFT result is selected as the main frequency festBackward deducing the wavelength lambda from the speed of sound cest
λest=c/fest
Step 6, cross-correlation operation
The cross-correlation function describes the degree of correlation of two time series. Defining the cross-correlation function of two sequences f [ n ] and g [ n ] of digital sound signals as:
Figure BDA0002042585390000041
wherein R isf,g[m]Represents the sequence g [ n ]]Sliding m points to left and then reacting with f [ n ]]The results are multiplied point by point and summed. If g [ n ]]Is f [ n ]]Time delay sequence of g [ n ]]Sliding J point back and f [ n ]]Aligned when the two sequences are most similar, the cross-correlation function also peaks at this point.
In general, cross-correlation and convolution are not directly calculated in a signal processing system due to large calculation amount, and are converted into a frequency domain to ensure processing speed. Therefore, the present invention performs inverse fourier transform after fourier transform on the cross-correlation function, so that the cross-correlation operation is actually calculated in the signal processing system by the following formula:
Rf,g[n]=IFFT{F[n]G*[n]}
where F [ n ] and G [ n ] are discrete Fourier transforms of F [ n ] and G [ n ], G [ n ] denotes the conjugation to the sequence G [ n ], IFFT { } denotes the inverse Fourier transform operation.
Because the function reaches the peak value when the sequences are similar, the time delay difference between the sequences can be obtained by inverse solution after the peak value position of the cross-correlation result is obtained. For example, for a two-voice frame signal f [ N ] with a frame length N]And g [ n ]]Cross correlation function R thereoff,g[m]At point m, L has a maximum value, i.e. g n]Left shift L point and f [ n ]]The sequences are most similar, and the time slot corresponding to one sampling point is 1/fsThus, two sequences of delays are solved:
Δt=L/fs
fig. 5 shows the cross-correlation results of two sound sequences with frame length 512 and sampling rate 40.96KHz, and the actual time delay difference is 1.2489 × 10-4s, time delay Δ t solved by an algorithmsComprises the following steps:
Figure BDA0002042585390000042
it can be seen that the actual result matches well with the calculated result.
Step 7, DOA estimation
Because the delay difference of the sound signal reaching each microphone depends on the position of the speaker and the geometrical shape of the microphone array, the DOA can be estimated after the delay difference is obtained and the geometrical shape of the microphone array is known.
Let r be the distance from the sound source to the center of the array, L be the length of the array, and the highest frequency f of normal speech of the personmaxAround 4KHz, the wavelength is about λ ═ c/fmax85 cm. Based on the knowledge of the array antenna, if a low sidelobe or deep null complex beam is to be realized, the low sidelobe or deep null complex beam is neededSatisfy r ≥ 10L2And/lambda. In practical application scenarios where r is between 2m and 10m, the array length can be calculated to be between 0.13m and 0.29m, and since the array pitch is very small compared to r, a far-field model can be used, and the DOA of each microphone is considered to be equal, as shown in fig. 7.
As can be seen from fig. 6, the time delay difference Δ t of the sound signals between adjacent microphones is dsin θ/c, and since it can be solved through cross-correlation operation to obtain a specific value, and the array element spacing d and the sound velocity c are also known, the arrival angle θ of each array element is arcsin (c Δ t/d), and finally, the average is obtained.
Figure 7 shows that the algorithm solves for an angle of arrival of 20.6 degrees at an actual DOA (relative to the geometric center of the array) of 20.2 degrees. The deviation between the solution value and the actual value is small, and the reliability of the algorithm is proved.
Step 8, weight coefficient generation
After the arrival angle is solved, the knowledge about the array antenna needs to perform phase compensation on each signal on the array element through weighting in order to make the beam align with the signal at the angle and suppress the signals arriving in other directions. In the existing digital beam forming technology, there are many ways of generating weight coefficients to implement spatial filtering, such as LCMV filtering algorithm, which can enhance the signal in one direction and generate nulls in another interference direction at the same time; also like the autocorrelation matrix inversion (SMI) method, etc., therefore, the choice of the weight coefficient generation manner is very free.
The basic array antenna spatial filtering algorithm is explained in detail below.
When the arrival angle is theta, the phase difference between each array element k of the uniform linear array and the sound signal of the reference array element is as follows:
Figure BDA0002042585390000051
at this time, weighting coefficients are taken at each array element
Figure BDA0002042585390000052
I.e. the main lobe of the beam is aligned with the direction of the angle of arrival theta.
Fig. 8 shows the effect of the weighting coefficients used when θ is 0 degrees and 30 degrees, respectively, on the gain effect in the array direction under the example algorithm. It can be seen that the prediction fits well with the reality.
In this case, the wide main lobe of the 4-element linear array makes up for the deviation of the wavelength estimation. In a conference or teaching scene, only one speaker can talk at the same time, and for the noise environment nearby the speaker, the volume of the speaker is the largest, so the wide main lobe does not influence the spatial filtering effect. If a more diverse array is used, it is more effective, but it is not very meaningful for a meeting or classroom environment.
Step 9, weighted summation
And weighting the sound frame signals of each array element according to the weight coefficient of each array element calculated in the step 8, and then adding and summing the weighted sound frame signals.
Fig. 9 compares the difference in the effect of the weighted sum and the direct sum, the source of the sound signal being a sample point of human laughter provided by MATLAB. The left column is a time domain diagram of the signal, and the right side is a frequency spectrum diagram of the complete signal; the uppermost row represents original digital signals, the middle row represents digital signals processed by the classroom speaker sound following enhancement system based on the microphone array, and the last row represents digital signals directly added without any processing by adopting the same microphone array. It is apparent from the figure that the system proposed by the present invention can effectively enhance the signal while maintaining the original shape of the sound signal spectrum; the system, which adds directly without any processing, although it also enhances the signal, produces significant distortion in the frequency spectrum of the sound signal and produces a poor sound quality from the loudspeaker.
In addition, because the system adopts the microphone array, the final result is the weighted summation of each array element, so that the whole system can not be directly paralyzed due to the damage of a single microphone, and the self-healing capability is strong.
Step 10, D/A conversion
And converting the weighted and summed digital signals into analog signals, and outputting the analog signals to a loudspeaker to realize the enhanced playing of the voice of the speaker.
Step 11, determining whether to close the system according to the session status
If the conversation is not finished, the system processes the next frame signal immediately; if the conversation is over, the manager can close the system power supply and stop the system operation.

Claims (1)

1. An indoor sound source following and enhancing method based on a microphone array is characterized by comprising the following steps:
step 1, performing anti-aliasing low-pass filtering on analog sound signals picked up by each microphone, and performing A/D sampling and frame division on the filtered analog signals;
step 2, carrying out frame level detection on each path of sound signal, dividing a frame into a sound frame signal and a noise frame signal, if the sound frame is a noise frame, storing the sound frame signal, and if the sound frame is a sound frame, carrying out spectrum cancellation with the previous noise frame;
step 3, selecting a reference path signal, estimating a wavelength according to a reference path sound frame, and performing cross-correlation operation on each path of sound frame and a current reference path sound frame to further estimate a wave arrival angle;
step 4, determining a weight coefficient corresponding to each sound frame according to the estimated sound wave wavelength and the estimated arrival angle and by combining the geometric shape of the microphone array;
step 5, carrying out weighted summation on each path of sound frame to determine an enhanced voice digital signal, carrying out D/A conversion on the obtained voice digital signal, generating an analog signal and transmitting the analog signal to a loudspeaker;
step 6, repeating the steps 1-5, and processing the next frame of sound signals until the sound pickup is finished;
in the step 1, the cut-off frequency band of the anti-aliasing low-pass filtering is a signal above 20KHz which can not be heard by human ears; the divided frame duration should not exceed the masking effect time, the frame interval should be less than the frame length, the common human ear forward masking effect is shortest 5ms, if the sampling rate is fsIf the energy efficiency response duration is within fs200 points, taking the point number requirement of fast Fourier transform into consideration
Figure FDA0003268700760000011
The point is an optimal frame, wherein
Figure FDA0003268700760000012
Represents an integer less than and closest to x;
in step 2, taking the absolute value of all points in the frame and averaging to obtain the judgment level of each frame signal, and comparing the judgment level with a threshold level so as to divide the frame into sound frames or noise frames; carrying out subtraction on the FFT of the sound frame signal and the FFT of the noise frame signal of the nearest section to carry out spectrum cancellation;
in step 3, the FFT result of the sound signal frame is analyzed, and the central point of the frequency band range exceeding the set threshold in the FFT result is selected as the main frequency festCombined with the speed of sound c, backward wavelength λest
λest=c/fest
In step 3, after fourier transform is performed on the cross-correlation function, inverse fourier transform is performed again, since the function reaches the peak value when the sequences are similar, after the peak position of the cross-correlation result is obtained, the time delay difference between the sequences can be obtained by inverse solution, and the specific calculation formula is as follows:
Rf,g[n]=IFFT{F[n]G*[n]}
wherein F [ n ]]And G [ n ]]Is f [ n ]]And g [ n ]]Discrete Fourier transform of G*[n]Represents a pair sequence G [ n ]]Taking conjugation, and using IFFT to represent inverse Fourier transform operation;
in step 3, setting the distance between any array element and a reference array element as d, the sound velocity as c and the time delay difference between sequences as delta t, so that the arrival angle of the array element is theta arcsin (c delta t/d), and averaging the arrival angles of all the array elements to obtain the arrival angle of the whole array;
in step 4, determining a weight coefficient corresponding to each path of sound frame according to a digital beam forming algorithm;
in step 4, for the uniform linear array, setting the phase difference between each array element k and the sound signal of the reference array element as follows:
Figure FDA0003268700760000021
in order to make the main lobe of the wave beam align with the direction of arrival angle theta, weighting coefficients are taken at each array element
Figure FDA0003268700760000022
CN201910346841.0A 2019-04-27 2019-04-27 Microphone array-based indoor sound source following and enhancing method Active CN110111807B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910346841.0A CN110111807B (en) 2019-04-27 2019-04-27 Microphone array-based indoor sound source following and enhancing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910346841.0A CN110111807B (en) 2019-04-27 2019-04-27 Microphone array-based indoor sound source following and enhancing method

Publications (2)

Publication Number Publication Date
CN110111807A CN110111807A (en) 2019-08-09
CN110111807B true CN110111807B (en) 2022-01-11

Family

ID=67487060

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910346841.0A Active CN110111807B (en) 2019-04-27 2019-04-27 Microphone array-based indoor sound source following and enhancing method

Country Status (1)

Country Link
CN (1) CN110111807B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110687523B (en) * 2019-08-29 2023-07-11 中国科学技术大学 An obstacle detection system, method and storage medium
TWI738532B (en) * 2019-10-27 2021-09-01 英屬開曼群島商意騰科技股份有限公司 Apparatus and method for multiple-microphone speech enhancement
CN111665422A (en) * 2020-06-08 2020-09-15 郑州精铖电力设备有限公司 FPGA-based microphone array non-invasive type broadband sound wave real-time imaging detection system
CN114613383B (en) * 2022-03-14 2023-07-18 中国电子科技集团公司第十研究所 Multi-input voice signal beam forming information complementation method in airborne environment
CN114624652B (en) * 2022-03-16 2022-09-30 浙江浙能技术研究院有限公司 Sound source positioning method under strong multipath interference condition

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1866356A (en) * 2005-08-15 2006-11-22 华为技术有限公司 Broadband wave beam forming method and apparatus
WO2007123047A1 (en) * 2006-04-20 2007-11-01 Nec Corporation Adaptive array control device, method, and program, and its applied adaptive array processing device, method, and program
CN105788607A (en) * 2016-05-20 2016-07-20 中国科学技术大学 Speech enhancement method applied to dual-microphone array
CN109239667A (en) * 2018-10-26 2019-01-18 深圳市友杰智新科技有限公司 A kind of sound localization method based on two-microphone array
CN109313909A (en) * 2018-08-22 2019-02-05 深圳市汇顶科技股份有限公司 Assess method, unit and the system of microphone array consistency

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1866356A (en) * 2005-08-15 2006-11-22 华为技术有限公司 Broadband wave beam forming method and apparatus
WO2007123047A1 (en) * 2006-04-20 2007-11-01 Nec Corporation Adaptive array control device, method, and program, and its applied adaptive array processing device, method, and program
CN105788607A (en) * 2016-05-20 2016-07-20 中国科学技术大学 Speech enhancement method applied to dual-microphone array
CN109313909A (en) * 2018-08-22 2019-02-05 深圳市汇顶科技股份有限公司 Assess method, unit and the system of microphone array consistency
CN109239667A (en) * 2018-10-26 2019-01-18 深圳市友杰智新科技有限公司 A kind of sound localization method based on two-microphone array

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Optimal two-layer directive microphone array with application in near-field acoustical holography;Mingsian R. Bai .etal;《Acoustical Society of America》;20120831;第132卷(第2期);第862–871页 *
基于麦克风阵列的语音增强与干扰抑制算法;王义圆等;《电声技术》;20181231;第42卷(第2期);全文 *
基于麦克风阵列的近场环境下语音增强算法研究;余洋;《中国优秀硕士学位论文全文数据库信息科技辑》;20190115(第1期);全文 *

Also Published As

Publication number Publication date
CN110111807A (en) 2019-08-09

Similar Documents

Publication Publication Date Title
CN110111807B (en) Microphone array-based indoor sound source following and enhancing method
US10290311B2 (en) Vector noise cancellation
CN105165026B (en) Use the filter and method of the informed space filtering of multiple instantaneous arrival direction estimations
CN108172235B (en) LS wave beam forming reverberation suppression method based on wiener post filtering
US9503818B2 (en) Method and apparatus for processing signals of a spherical microphone array on a rigid sphere used for generating an ambisonics representation of the sound field
EP2647221B1 (en) Apparatus and method for spatially selective sound acquisition by acoustic triangulation
EP1856948B1 (en) Position-independent microphone system
CN109490822B (en) Voice DOA estimation method based on ResNet
US20120093344A1 (en) Optimal modal beamformer for sensor arrays
CN102324237A (en) Microphone array voice wave beam formation method, speech signal processing device and system
CN101852846A (en) Signal handling equipment, signal processing method and program
US9601133B2 (en) Vector noise cancellation
US10021508B2 (en) Method and apparatus for processing signals of a spherical microphone array on a rigid sphere used for generating an ambisonics representation of the sound field
CN114255777A (en) Mixing method and system for removing reverberation of real-time voice
CN111157949A (en) Voice recognition and sound source positioning method
Mittal et al. Low Latency Two Stage Beamforming with Distributed Microphone Arrays Using a Planewave Decomposition
Meng et al. Deep Kronecker Product Beamforming for Large-Scale Microphone Arrays
Zhao et al. Adapting GCC-PHAT to co-prime circular microphone arrays for speech direction of arrival estimation using neural networks
Pedamallu Microphone Array Wiener Beamforming with emphasis on Reverberation
Natarajan et al. Revolutionizing Speech Clarity: Unveiling a Novel Approach with Hybrid Coherent-to-Diffuse Power Ratio and Recursive Least Square Algorithm for Reverberation Removal
Li et al. Subband gradient flow acoustic source separation for moderate reverberation environment
Yermeche et al. Moving source speech enhancement using time-delay estimation
CN119785811A (en) A multi-channel speech enhancement method for remote conferencing
Amerineni Multi Channel Sub Band Wiener Beamformer
Lin et al. An adaptive low-complexity coherence-based beamformer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant