USRE43570E1 - Method and apparatus for improved weighting filters in a CELP encoder - Google Patents
Method and apparatus for improved weighting filters in a CELP encoder Download PDFInfo
- Publication number
- USRE43570E1 USRE43570E1 US12/157,945 US15794508A USRE43570E US RE43570 E1 USRE43570 E1 US RE43570E1 US 15794508 A US15794508 A US 15794508A US RE43570 E USRE43570 E US RE43570E
- Authority
- US
- United States
- Prior art keywords
- signal
- speech
- error
- speech signal
- weighted
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime, expires
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 230000005284 excitation Effects 0.000 claims abstract description 70
- 230000006854 communication Effects 0.000 claims description 5
- 238000004891 communication Methods 0.000 claims description 5
- 238000001228 spectrum Methods 0.000 description 13
- 230000015572 biosynthetic process Effects 0.000 description 12
- 238000003786 synthesis reaction Methods 0.000 description 12
- 230000003044 adaptive effect Effects 0.000 description 10
- 238000010586 diagram Methods 0.000 description 10
- 230000007774 longterm Effects 0.000 description 7
- 230000001755 vocal effect Effects 0.000 description 5
- 238000001914 filtration Methods 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000000695 excitation spectrum Methods 0.000 description 3
- 230000000737 periodic effect Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000007175 bidirectional communication Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000001308 synthesis method Methods 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/08—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
- G10L19/12—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters the excitation function being a code excitation, e.g. in code excited linear prediction [CELP] vocoders
Definitions
- the present invention relates generally to digital voice encoding and, more particularly, to a method and apparatus for improved weighting filters in a CELP encoder.
- FIG. 1A A general diagram of a CELP encoder 100 is shown in FIG. 1A .
- a CELP encoder uses a model of the human vocal tract to reproduce a speech input signal. The parameters for the model are actually extracted from the speech signal being reproduced, and it is these parameters that are sent to a decoder 114 , which is illustrated in FIG. 1B . Decoder 114 uses the parameters to reproduce the speech signal.
- synthesis filter 104 is a linear predictive filter and serves as the vocal tract model for CELP encoder 100 .
- Synthesis filter 114 takes an input excitation signal ⁇ (n) and synthesizes a speech signal s′(n) by modeling the correlations introduced into speech by the vocal tract and applying them to the excitation signal ⁇ (n).
- CELP encoder 100 speech is broken up into frames, usually 20 ms each, and parameters for synthesis filter 104 are determined for each frame. Once the parameters are determined, an excitation signal ⁇ (n) is chosen for that frame. The excitation signal is then synthesized, producing a synthesized speech signal s′(n). The synthesized frame s′(n) is then compared to the actual speech input frame s(n) and a difference or error signal e(n) is generated by subtractor 106 . The subtraction function is typically accomplished via an adder or similar functional component as those skilled in the art will be aware. Actually, excitation signal ⁇ (n) is generated from a predetermined set of possible signals by excitation generator 102 .
- CELP encoder 100 all possible signals in the predetermined set are tried in order to find the one that produces the smallest error signal e(n). Once this particular excitation signal ⁇ (n) is found, the signal and the corresponding filter parameters are sent to decoder 112 , which reproduces the synthesized speech signal s′(n). Signal s′(n) is reproduced in decoder 112 using an excitation signal ⁇ (n), as generated by decoder excitation generator 114 , and synthesizing it using decoder synthesis filter 116 .
- CELP encoder 100 includes a feedback path that incorporates error weighting filter 108 .
- the function of error weighting filter 108 is to shape the spectrum of error signal e(n) so that the noise spectrum is concentrated in areas of high voice content.
- the shape of the noise spectrum associated with the weighted error signal e w (n) tracks the spectrum of the synthesized speech signal s(n), as illustrated in FIG. 2 by curve 206 . In this manner, the SNR is improved and the quality of the reproduced speech is increased.
- the weighted error signal e w (n) is also used to minimize the error signal by controlling the generation of excitation signal ⁇ (n)
- signal e w (n) actually controls the selection of signal ⁇ (n) and the gain associated with signal ⁇ (n).
- the energy associated with s′(n) be as stable or constant as possible. Energy stability is controlled by the gain associated with ⁇ (n) and requires a less aggressive weighting filter 108 .
- a speech encoder comprising a first weighting means for performing an error weighting on a speech input.
- the first weighting means is configured to reduce an error signal resulting from a difference between a first synthesized speech signal and the speech input.
- the speech encoder includes a means for generating the first synthesized speech signal from a first excitation signal, and a second weighting means for performing an error weighting on the first synthesized speech signal.
- the second weighting means is also configured to reduce the error signal resulting from the difference between the speech input and the first synthesized speech signal.
- the speech encoder also includes a means for generating a second synthesized speech signal from a second excitation signal, and a third weighting means for performing an error weighting on the second synthesized speech signal.
- the third weighting means is configured to reduce a second error signal resulting from the difference between the first weighted error signal and the second synthesized speech signal.
- a second difference means for taking the difference between the second synthesized speech signal and the first error signal, where the second difference means is configured to produce a second weighted error signal.
- a feedback means for using the second weighted error signal to control the selection of the first excitation signal, and the selection of the second excitation signal.
- a transmitter that includes a speech encoder such as the one described above and a method for speech encoding.
- FIG. 1A is a block diagram illustrating a CELP encoder.
- FIG. 1B is a block diagram illustrating a decoder that works in conjunction with the encoder of FIG. 1A .
- FIG. 2 is a graph illustrating the signal to noise ratio of a synthesized speech signal and a weighted error signal in the encoder illustrated in FIG. 1A .
- FIG. 3 is a second block diagram of a CELP encoder.
- FIG. 4 is a block diagram illustrating one embodiment of a speech encoder in accordance with the invention.
- FIG. 5 is a graph illustrating the pitch of a speech signal.
- FIG. 6 is a block diagram of a second embodiment of a speech encoder in accordance with the invention.
- FIG. 7A is a diagram illustrating the concentration of energy of the speech signal in the low frequency portion of the spectrum.
- FIG. 7B is a diagram illustrating the concentration of energy of the speech signal in the high frequency portion of the spectrum.
- FIG. 8 is a block diagram, illustrating a transmitter that includes a speech encoder such as the speech encoder illustrated in FIG. 4 or FIG. 6 .
- FIG. 9 is a process flow diagram illustrating a method of speech encoding such in accordance with the invention.
- excitation signal ⁇ (n) is generated from a large vector quantizer codebook such as codebook 302 in encoder 300 .
- Multiplier 308 multiplies the signal selected from codebook 302 by gain term (g c ) in order to control the power of excitation signal ⁇ (n).
- Equation (2) represents a prediction error filter determined by minimizing the energy of a residual signal produced when the original signal is passed through synthesis filter 312 .
- Synthesis filter 312 is designed to model the vocal tract by applying the correlation normally introduced into speech by the vocal tract to excitation signal ⁇ (n). The result of passing excitation signal ⁇ (n) through synthesis filter 312 is synthesized speech signal s′(n).
- Synthesized speech signal s′(n) is passed through error weighting filter 314 , producing weighted synthesized speech signal s′ w (n).
- Speech input s(n) is also passed through an error weighting filter 318 , producing weighted speech signal s w (n).
- Weighted synthesized speech signal s′ w (n) is subtracted from weighted speech signal s w (n), which produces an error signal.
- the function of the error weighting filters 314 and 318 is to shape the spectrum of the error signal so that the noise spectrum of the error signal is concentrated in areas of high voice content. Therefore, the error signal generated by subtractor 316 is actually a weighted error signal e w (n).
- Weighted error signal e w (n) is feedback to control the selection of the next excitation signal from codebook 302 and also to control the gain term (g c ) applied thereto. Without the feedback, every entry in codebook 302 would need to be passed through synthesis filter 302 and subtractor 316 to find the entry that produced the smallest error signal. But by using error weighting filters 314 and 318 and feeding weighted error signal e w (n) back, the selection process can be streamlined and the correct entry found much quicker.
- Codebook 302 is used to track the short term variations in speech signal s(n); however, speech is characterized by long-term periodicities that are actually very important to effective reproduction of speech signal s(n).
- an adaptive codebook 304 may be included so that the excitation signal ⁇ (n) will include a component of the form G ⁇ (n ⁇ ), where ⁇ is the estimated pitch period. Pitch is the term used to describe the long-term periodicity.
- the adaptive codebook selection is multiplied by gain factor (g p ) in multiplier 306 .
- the selection from adaptive codebook 304 and the selection from codebook 302 are then combined in adder 310 to create excitation signal ⁇ (n).
- synthesis filter 312 may include a pitch filter to model the long-term periodicity present in the voiced speech.
- Encoder 400 uses parallel signal paths for an excitation signal ⁇ 1 (n), from adaptive codebook 402 , and for an excitation signal ⁇ 2 (n) from fixed codebook 404 .
- Each excitation signal ⁇ 1 (n) and ⁇ 2 (n) are multiplied by independent gain terms (g p ) and (g c ) respectively.
- Independent synthesis filters 410 and 412 generate synthesized speech signals s′ 1 (n) and s′ 2 (n) from excitation signals ⁇ 1 (n) and ⁇ 2 (n) and independent error weighting filters 414 and 416 generate weighted synthesized speech signals s′ w1 (n) and s′ w2 (n), respectively.
- Weighted synthesized speech signal s′ w1 (n) is subtracted in subtractor 420 from weighted speech signal s w (n), which is generated from speech signal s(n) by error weighting filter 418 .
- Equation (4) is essentially the same as the equation for e w (n) in encoder 300 of FIG. 3 .
- the error weighting and gain terms applied to the selections from the codebooks are independent and can either be independently controlled through feedback or independently initialized.
- weighted error signal e w (n) in encoder 400 is used to independently control the selection from fixed codebook 404 and the gain (g c ) applied thereto, and the selection from a adaptive codebook 402 and the gain (g p ) applied thereto.
- different error weighting can be used for each error weighting filter 414 , 416 , and 418 .
- the speech input source may be a microphone or a telephone line, such as a telephone line used for an Internet connection.
- the speech input can, therefore, vary from very noisy to relatively calm.
- a set of optimum error weighting parameters for each type of input is determined by the testing.
- the type of input used in encoder 400 is then the determining factor for selecting the appropriate set of parameters to be used for error weighting filters 414 , 416 , and 418 .
- the selection of optimum error weighting parameters combined with independent control of the codebook selections and gains applied thereto, allows for effective balancing of energy stability and excitation spectrum flatness.
- the performance of encoder 400 is improved with regard to both.
- pitch estimator 424 may be incorporated into encoder 400 .
- pitch estimator 424 generates a speech pitch estimate s p (n), which is used to further control the selection from adaptive codebook 402 . This further control is designed to ensure that the long-term periodicity of speech input s(n) is correctly replicated in the selections from adaptive codebook 402 .
- FIG. 5 illustrates a speech sample 502 .
- the short-term variation in the speech signal can change drastically from point to point along speech sample 502 .
- the long-term variation tends to be very periodic.
- the period of speech sample 502 is denoted as (T) in FIG. 5 .
- Period (T) represents the pitch of speech sample 502 ; therefore, if the pitch is not estimated accurately, then the reproduced speech signal may not sound like the original speech signal.
- Filter 602 In order to improve the speech pitch estimation s p (n) encoder 600 of FIG. 6 includes an additional filter 602 .
- Filter 602 generates a filtered weighted speech signal s′′ w (n), which is used by pitch estimator 424 , from weighted speech signal s w (n).
- filter 602 is a low pass filter (LPF). This is because the low frequency portion of speech input s(n) will be more periodic than the high frequency portion. Therefore, filter 602 will allow pitch estimator 424 to make a more accurate pitch estimation by emphasizing the periodicity of speech input s(n).
- LPF low pass filter
- filter 602 is an adaptive filter. Therefore, as illustrated in FIG. 7A , when the energy in speech input s(n) is concentrated in the low frequency portion of the spectrum, very little or no filtering is applied by filter 602 . This is because the low frequency portion and thus the periodicity of speech input s(n) is already emphasized. If, however, the energy in speech input s(n) is concentrated in the higher frequency portion of the spectrum ( FIG. 7B ), then a more aggressive low pass filtering is applied by filter 602 . By varying the degree of filtering applied by filter 602 according to the energy concentration of speech input s(n), a more optimized speech input estimation s p (n) is maintained.
- the input to filter 602 is speech input s(n).
- filter 602 will incorporate a fourth error weighting filter to perform error weighting on speech input s(n).
- This configuration enables the added flexibility of making the error weighting filter incorporated in filter 602 different from error weighting filter 418 , in particular, as well as from filters 414 and 416 . Therefore, the implementation illustrated in FIG. 6 allows for each of four error weighting filters to be independently configured so as to provide the optimum error weighting of each of the four input signals. The result is a highly optimized estimation of speech input s(n).
- filter 602 may take its input from the output of error weighting filter 418 .
- error weighting filter 418 provides the error weighting for s′′w(n)
- filter 602 does not incorporate a fourth error weighting filter.
- This implementation is illustrated by the dashed line in FIG. 6 . This implementation may be used when different error weighting for s′′w(n) and sw(n) is not required.
- the resulting implementation of filter 602 only incorporates the LDF function and is easier to design and implement relative to the previous implementation.
- Transmitter 800 comprises a voice input means 802 , which is typically a microphone.
- Speech input means 802 is coupled to a speech encoder 804 , which encodes speech input provided by speech input means 802 for transmission by transmitter 800 .
- Speech encoder 804 is an encoder such as encoder 400 or encoder 600 as illustrated in FIG. 4 and FIG. 6 , respectively.
- the encoded data generated by speech encoder 804 comprises information relating to the selection for codebooks 402 and 404 and for gain terms (g p ) and (g c ), as well as parameters for synthesis filters 410 and 412 .
- a device which receives the transmission from transmitter 800 , will use these parameters to reproduce the speech input provided by speech input means 802 .
- a device may include a decoder as described in U.S. patent application Ser. No. 09/624,187, filed Jul. 25, 2000, now U.S. Pat. No. 6,466,904, titled “Method and Apparatus Using Harmonic Modeling in an Improved Speech Decoder,” which is incorporated herein by reference in its entirety.
- Speech encoder 804 is coupled to a transceiver 806 , which converts the encoded data from speech encoder 804 into a signal that can be transmitted.
- transmitter 800 will include an antenna 810 .
- transceiver 806 will convert the data from speech encoder 804 into an RF signal for transmission via antenna 810 .
- Other implementations, however, will have a fixed line interface such as a telephone interface 808 .
- Telephone interface 808 may be an interface to a PSTN or ISDN line, for example, and may be accomplished via a coaxial cable connection, a regular telephone line, or the like. In a typical implementation, telephone interface 808 is used for connecting to the Internet.
- Transceiver 806 will typically be interfaced to a decoder as well for bidirectional communication; however, such a decoder is not illustrated in FIG. 8 , because it is not particularly relevant to the invention.
- Transmitter 800 is capable of implementation in a variety of communication devices.
- transmitter 800 may, depending on the implementation, be included in a telephone, a cellular/PCS mobile phone, a cordless phone, a digital answering machine, or a personal digital assistant.
- step 902 error weighting is performed on a speech signal.
- the error weighting may be performed on a speech signal sent by an error weighting filter 418 .
- step 904 a first synthesized speech signal is generated from a first excitation signal multiplied by a first gain term. For example, s′(n) as generated from ⁇ 1 (n) multiplied by gain term (g p ) in FIG. 4 .
- step 906 error weighting is then performed on the first synthesized speech signal to create a weighted first synthesized speech signal, such as s′ w1 (n) illustrated in FIG. 4 .
- step 408 a first error signal is generated by taking the difference between the weighted speech signal and the weighted first synthesized speech signal.
- a second synthesized speech signal is generated from a second excitation signal multiplied by a second gain term. For example, s′ 2 (n) as generated in FIG. 4 by multiplying ⁇ 2 (n) by (g c ).
- error weighting is performed on the second synthesized speech signal to create a weighted second synthesized speech signal, such as s′ w2 (n) in FIG. 4 .
- a second weighted error signal is generated by taking the difference between the first weighted error signal and the weighted second synthesized speech signal.
- This second weighted error signal is then used, in step 916 , to control the generation of subsequent first and second synthesized speech signals.
- the second weighted error signal is used as feedback to control subsequent values of the second weighted error signal. For example, such feedback is illustrated by the feedback of e w (n) in FIG. 4 .
- pitch estimation is performed on the speech signal as illustrated in FIG. 4 by optional step 918 .
- the pitch estimation is then used to control the generation of at least one of the first and second synthesized speech signals.
- a pitch estimation s p (n) is generated by pitch estimator 424 as illustrated in FIG. 4 .
- a filter is used to optimize the pitch estimation. Therefore, as illustrated by optional step 920 in FIG. 4 , the speech signal is filtered and a filtered version of the speech signal is used for the pitch estimation in step 918 .
- a filter 602 as illustrated in FIG. 6 , may be used to generate a filtered speech signal s′′ w (n).
- the filtering is adaptive based on the energy spectrum of the speech signal.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
Abstract
Description
H(z)=I/A(z) (1)
Where
A(z)=1−Σi=1 Pαiz−1
ew(n)=sw(n)−s′w1(n)−s′w2(n) (3)
ew(n)=sw(n)−(s′w1(n)+s′w2(n)) (4)
Claims (33)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/157,945 USRE43570E1 (en) | 2000-07-25 | 2008-06-13 | Method and apparatus for improved weighting filters in a CELP encoder |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/625,088 US7013268B1 (en) | 2000-07-25 | 2000-07-25 | Method and apparatus for improved weighting filters in a CELP encoder |
US10/628,904 US7062432B1 (en) | 2000-07-25 | 2003-07-28 | Method and apparatus for improved weighting filters in a CELP encoder |
US12/157,945 USRE43570E1 (en) | 2000-07-25 | 2008-06-13 | Method and apparatus for improved weighting filters in a CELP encoder |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/628,904 Reissue US7062432B1 (en) | 2000-07-25 | 2003-07-28 | Method and apparatus for improved weighting filters in a CELP encoder |
Publications (1)
Publication Number | Publication Date |
---|---|
USRE43570E1 true USRE43570E1 (en) | 2012-08-07 |
Family
ID=35998889
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/625,088 Ceased US7013268B1 (en) | 2000-07-25 | 2000-07-25 | Method and apparatus for improved weighting filters in a CELP encoder |
US10/628,904 Expired - Lifetime US7062432B1 (en) | 2000-07-25 | 2003-07-28 | Method and apparatus for improved weighting filters in a CELP encoder |
US12/157,945 Expired - Lifetime USRE43570E1 (en) | 2000-07-25 | 2008-06-13 | Method and apparatus for improved weighting filters in a CELP encoder |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/625,088 Ceased US7013268B1 (en) | 2000-07-25 | 2000-07-25 | Method and apparatus for improved weighting filters in a CELP encoder |
US10/628,904 Expired - Lifetime US7062432B1 (en) | 2000-07-25 | 2003-07-28 | Method and apparatus for improved weighting filters in a CELP encoder |
Country Status (1)
Country | Link |
---|---|
US (3) | US7013268B1 (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7013268B1 (en) * | 2000-07-25 | 2006-03-14 | Mindspeed Technologies, Inc. | Method and apparatus for improved weighting filters in a CELP encoder |
US7171355B1 (en) * | 2000-10-25 | 2007-01-30 | Broadcom Corporation | Method and apparatus for one-stage and two-stage noise feedback coding of speech and audio signals |
JP3404016B2 (en) * | 2000-12-26 | 2003-05-06 | 三菱電機株式会社 | Speech coding apparatus and speech coding method |
US7752052B2 (en) * | 2002-04-26 | 2010-07-06 | Panasonic Corporation | Scalable coder and decoder performing amplitude flattening for error spectrum estimation |
US20080208575A1 (en) * | 2007-02-27 | 2008-08-28 | Nokia Corporation | Split-band encoding and decoding of an audio signal |
JP4871894B2 (en) | 2007-03-02 | 2012-02-08 | パナソニック株式会社 | Encoding device, decoding device, encoding method, and decoding method |
Citations (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4720861A (en) | 1985-12-24 | 1988-01-19 | Itt Defense Communications A Division Of Itt Corporation | Digital speech coding circuit |
US5195137A (en) | 1991-01-28 | 1993-03-16 | At&T Bell Laboratories | Method of and apparatus for generating auxiliary information for expediting sparse codebook search |
US5293449A (en) | 1990-11-23 | 1994-03-08 | Comsat Corporation | Analysis-by-synthesis 2,4 kbps linear predictive speech codec |
US5491771A (en) | 1993-03-26 | 1996-02-13 | Hughes Aircraft Company | Real-time implementation of a 8Kbps CELP coder on a DSP pair |
US5495555A (en) | 1992-06-01 | 1996-02-27 | Hughes Aircraft Company | High quality low bit rate celp-based speech codec |
US5633982A (en) | 1993-12-20 | 1997-05-27 | Hughes Electronics | Removal of swirl artifacts from celp-based speech coders |
US5717824A (en) | 1992-08-07 | 1998-02-10 | Pacific Communication Sciences, Inc. | Adaptive speech coder having code excited linear predictor with multiple codebook searches |
US5864798A (en) | 1995-09-18 | 1999-01-26 | Kabushiki Kaisha Toshiba | Method and apparatus for adjusting a spectrum shape of a speech signal |
US6246979B1 (en) | 1997-07-10 | 2001-06-12 | Grundig Ag | Method for voice signal coding and/or decoding by means of a long term prediction and a multipulse excitation signal |
US6470309B1 (en) | 1998-05-08 | 2002-10-22 | Texas Instruments Incorporated | Subframe-based correlation |
US6493665B1 (en) * | 1998-08-24 | 2002-12-10 | Conexant Systems, Inc. | Speech classification and parameter weighting used in codebook search |
US6556966B1 (en) | 1998-08-24 | 2003-04-29 | Conexant Systems, Inc. | Codebook structure for changeable pulse multimode speech coding |
US6738739B2 (en) | 2001-02-15 | 2004-05-18 | Mindspeed Technologies, Inc. | Voiced speech preprocessing employing waveform interpolation or a harmonic model |
US6804218B2 (en) | 2000-12-04 | 2004-10-12 | Qualcomm Incorporated | Method and apparatus for improved detection of rate errors in variable rate receivers |
US6925435B1 (en) * | 2000-11-27 | 2005-08-02 | Mindspeed Technologies, Inc. | Method and apparatus for improved noise reduction in a speech encoder |
US7062432B1 (en) * | 2000-07-25 | 2006-06-13 | Mindspeed Technologies, Inc. | Method and apparatus for improved weighting filters in a CELP encoder |
US7124078B2 (en) * | 1998-01-09 | 2006-10-17 | At&T Corp. | System and method of coding sound signals using sound enhancement |
JP2010181889A (en) | 1998-08-24 | 2010-08-19 | Mindspeed Technologies Inc | Selection of scalar quantization (sq) and vector quantization (vq) for speech encoding |
-
2000
- 2000-07-25 US US09/625,088 patent/US7013268B1/en not_active Ceased
-
2003
- 2003-07-28 US US10/628,904 patent/US7062432B1/en not_active Expired - Lifetime
-
2008
- 2008-06-13 US US12/157,945 patent/USRE43570E1/en not_active Expired - Lifetime
Patent Citations (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4720861A (en) | 1985-12-24 | 1988-01-19 | Itt Defense Communications A Division Of Itt Corporation | Digital speech coding circuit |
US5293449A (en) | 1990-11-23 | 1994-03-08 | Comsat Corporation | Analysis-by-synthesis 2,4 kbps linear predictive speech codec |
US5195137A (en) | 1991-01-28 | 1993-03-16 | At&T Bell Laboratories | Method of and apparatus for generating auxiliary information for expediting sparse codebook search |
US5495555A (en) | 1992-06-01 | 1996-02-27 | Hughes Aircraft Company | High quality low bit rate celp-based speech codec |
US5717824A (en) | 1992-08-07 | 1998-02-10 | Pacific Communication Sciences, Inc. | Adaptive speech coder having code excited linear predictor with multiple codebook searches |
US5491771A (en) | 1993-03-26 | 1996-02-13 | Hughes Aircraft Company | Real-time implementation of a 8Kbps CELP coder on a DSP pair |
US5633982A (en) | 1993-12-20 | 1997-05-27 | Hughes Electronics | Removal of swirl artifacts from celp-based speech coders |
US5864798A (en) | 1995-09-18 | 1999-01-26 | Kabushiki Kaisha Toshiba | Method and apparatus for adjusting a spectrum shape of a speech signal |
US6246979B1 (en) | 1997-07-10 | 2001-06-12 | Grundig Ag | Method for voice signal coding and/or decoding by means of a long term prediction and a multipulse excitation signal |
US7124078B2 (en) * | 1998-01-09 | 2006-10-17 | At&T Corp. | System and method of coding sound signals using sound enhancement |
US6470309B1 (en) | 1998-05-08 | 2002-10-22 | Texas Instruments Incorporated | Subframe-based correlation |
JP2010181889A (en) | 1998-08-24 | 2010-08-19 | Mindspeed Technologies Inc | Selection of scalar quantization (sq) and vector quantization (vq) for speech encoding |
US6556966B1 (en) | 1998-08-24 | 2003-04-29 | Conexant Systems, Inc. | Codebook structure for changeable pulse multimode speech coding |
US6493665B1 (en) * | 1998-08-24 | 2002-12-10 | Conexant Systems, Inc. | Speech classification and parameter weighting used in codebook search |
JP2010181893A (en) | 1998-08-24 | 2010-08-19 | Mindspeed Technologies Inc | Codebook sharing for line spectrum frequency quantization |
JP2010181890A (en) | 1998-08-24 | 2010-08-19 | Mindspeed Technologies Inc | Open-loop pitch processing for speech encoding |
JP2010181891A (en) | 1998-08-24 | 2010-08-19 | Mindspeed Technologies Inc | Control of adaptive codebook gain for speech encoding |
JP2010181892A (en) | 1998-08-24 | 2010-08-19 | Mindspeed Technologies Inc | Gain smoothing for speech coding |
EP2259255A1 (en) | 1998-08-24 | 2010-12-08 | Mindspeed Technologies Inc | Speech encoding method and system |
US7062432B1 (en) * | 2000-07-25 | 2006-06-13 | Mindspeed Technologies, Inc. | Method and apparatus for improved weighting filters in a CELP encoder |
US6925435B1 (en) * | 2000-11-27 | 2005-08-02 | Mindspeed Technologies, Inc. | Method and apparatus for improved noise reduction in a speech encoder |
US6804218B2 (en) | 2000-12-04 | 2004-10-12 | Qualcomm Incorporated | Method and apparatus for improved detection of rate errors in variable rate receivers |
US7590096B2 (en) | 2000-12-04 | 2009-09-15 | Qualcomm Incorporated | Method and apparatus for improved detection of rate errors in variable rate receivers |
US6738739B2 (en) | 2001-02-15 | 2004-05-18 | Mindspeed Technologies, Inc. | Voiced speech preprocessing employing waveform interpolation or a harmonic model |
Non-Patent Citations (20)
Title |
---|
Changchun, Two Kinds of Pitch Predictors in Speech Compressing Coding, Journal of Electronics, vol. 14 No. 3 (Jul. 1997). |
Coding of Speech at 9 kbit/s Using Conjugate-Structure Algebraic-Code-Excited Linear-Prediction (CS-ACELP), ITU-T Recommendation G.729, 1-35 (Mar. 1996). |
Defendants' Disclosure of Claim Terms and Proposed Constructions, Case 3:09-cv-00447-REP, Document 188, Filed Dec. 14, 2009, pp. 1-8. |
Digital cellular telecommunications system (Phase 2); Enhanced Full Rate (EFR) speech transcoding; (GSM 06.60 version 4.1.0), European Telecommunications Standards Institute Draft EN 301 245 V4.1.0, 1-47 (Jun. 1998). |
Dual Rate Speech Coder for Multimedia Communications Transmitting at 5.3 and 6.3 kbit/s, ITU-T Recommendation G.723.1, 1-27 (Mar. 1996). |
Enhanced Variable Rate Codec, Speech Service Option 3 for Wideband Spread Spectrum Digital Systems, TIA/EIA/IS-127 (Jan. 1997). |
File History for U.S. Appl. No. 09/663,002. |
File History for U.S. Appl. No. 11/251,179. |
File History for U.S. Appl. No. 12/069,973. |
File History for U.S. Appl. No. 12/218,242. |
File History for U.S. Appl. No. 12/220,480. |
File History for U.S. Appl. No. 12/321,934. |
File History for U.S. Appl. No. 60/097,569. |
LeBlanc, Efficient Search and Design Procedures for Robust Multi-Stage VQ of LPC Parameters for 4 kb/s Speech Coding, IEEE Transactions on Speech and Audio Processing, vol. 1, No. 4, (Oct. 1993). |
Masaaki Honda, "Speech Coding Using Waveform Matching Based on LPC Residual Phase Equalization", International Conference on Acoustics, Speech & Signal Processing (ICASSP '90), vol. 1, pp. 213-216. |
Ney, Dynamic Programming Algorithm for Optimal Estimation of Speech Parameter Contours, IEEE Transactions on Systems, Man, and Cybernetics, vol. SMC-13, No. 3 (Mar./Apr. 1983). |
Shlomot, Delayed Decision Switched Prediction Multi-Stage LSF Quantization, Rockwell Telecommunication. |
TIA/EIA IS-641-A, TDMA Cellular/PCS-Radio Interface Enhanced Full-Rate Voice Codec, Revision A. |
U.S Civil Docket Index for Case #: 3:09-cv-00447-REP, As of: Mar. 14, 2011 05:44 PM EDT, pp. 1-38. |
Vien V. Nguyen, Vladimir Goncharoff, and John Damoulakis, "Correcting Spectral Envelope Shifts in Linear Predictive Speech Compression Systems", Proceedings of the Military Communications Conference (Milcom '90), vol. 1, 1990, pp. 354-358. |
Also Published As
Publication number | Publication date |
---|---|
US7013268B1 (en) | 2006-03-14 |
US7062432B1 (en) | 2006-06-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6466904B1 (en) | Method and apparatus using harmonic modeling in an improved speech decoder | |
JP3490685B2 (en) | Method and apparatus for adaptive band pitch search in wideband signal coding | |
RU2262748C2 (en) | Multi-mode encoding device | |
RU2469422C2 (en) | Method and apparatus for generating enhancement layer in audio encoding system | |
US5778335A (en) | Method and apparatus for efficient multiband celp wideband speech and music coding and decoding | |
JP4662673B2 (en) | Gain smoothing in wideband speech and audio signal decoders. | |
JP4550289B2 (en) | CELP code conversion | |
JP3678519B2 (en) | Audio frequency signal linear prediction analysis method and audio frequency signal coding and decoding method including application thereof | |
JP3653826B2 (en) | Speech decoding method and apparatus | |
AU714752B2 (en) | Speech coder | |
US9530423B2 (en) | Speech encoding by determining a quantization gain based on inverse of a pitch correlation | |
JP4302978B2 (en) | Pseudo high-bandwidth signal estimation system for speech codec | |
USRE43570E1 (en) | Method and apparatus for improved weighting filters in a CELP encoder | |
KR20030046451A (en) | Codebook structure and search for speech coding | |
US8457953B2 (en) | Method and arrangement for smoothing of stationary background noise | |
JPH09152896A (en) | Sound path prediction coefficient encoding/decoding circuit, sound path prediction coefficient encoding circuit, sound path prediction coefficient decoding circuit, sound encoding device and sound decoding device | |
JP4963965B2 (en) | Scalable encoding apparatus, scalable decoding apparatus, and methods thereof | |
US20080255832A1 (en) | Scalable Encoding Apparatus and Scalable Encoding Method | |
RU2707144C2 (en) | Audio encoder and audio signal encoding method | |
JPH11504733A (en) | Multi-stage speech coder by transform coding of prediction residual signal with quantization by auditory model | |
Kataoka et al. | A 16-kbit/s wideband speech codec scalable with g. 729. | |
JP3481027B2 (en) | Audio coding device | |
JP4820954B2 (en) | Harmonic noise weighting in digital speech encoders | |
JP3785363B2 (en) | Audio signal encoding apparatus, audio signal decoding apparatus, and audio signal encoding method | |
CA2504174A1 (en) | Transcoder and code conversion method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MINDSPEED TECHNOLOGIES, INC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:WIAV SOLUTIONS LLC;REEL/FRAME:025717/0206 Effective date: 20100928 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
AS | Assignment |
Owner name: JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT Free format text: SECURITY INTEREST;ASSIGNOR:MINDSPEED TECHNOLOGIES, INC.;REEL/FRAME:032495/0177 Effective date: 20140318 |
|
AS | Assignment |
Owner name: MINDSPEED TECHNOLOGIES, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:032861/0617 Effective date: 20140508 Owner name: GOLDMAN SACHS BANK USA, NEW YORK Free format text: SECURITY INTEREST;ASSIGNORS:M/A-COM TECHNOLOGY SOLUTIONS HOLDINGS, INC.;MINDSPEED TECHNOLOGIES, INC.;BROOKTREE CORPORATION;REEL/FRAME:032859/0374 Effective date: 20140508 |
|
AS | Assignment |
Owner name: MINDSPEED TECHNOLOGIES, LLC, MASSACHUSETTS Free format text: CHANGE OF NAME;ASSIGNOR:MINDSPEED TECHNOLOGIES, INC.;REEL/FRAME:039645/0264 Effective date: 20160725 |
|
AS | Assignment |
Owner name: MACOM TECHNOLOGY SOLUTIONS HOLDINGS, INC., MASSACH Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MINDSPEED TECHNOLOGIES, LLC;REEL/FRAME:044791/0600 Effective date: 20171017 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553) Year of fee payment: 12 |