US20040181411A1 - Voicing index controls for CELP speech coding - Google Patents
Voicing index controls for CELP speech coding Download PDFInfo
- Publication number
- US20040181411A1 US20040181411A1 US10/799,503 US79950304A US2004181411A1 US 20040181411 A1 US20040181411 A1 US 20040181411A1 US 79950304 A US79950304 A US 79950304A US 2004181411 A1 US2004181411 A1 US 2004181411A1
- Authority
- US
- United States
- Prior art keywords
- input speech
- index
- linear prediction
- voicing
- voicing index
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/26—Pre-filtering or post-filtering
- G10L19/265—Pre-filtering, e.g. high frequency emphasis prior to encoding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/005—Correction of errors induced by the transmission channel, if related to the coding algorithm
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/08—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
- G10L19/087—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters using mixed excitation models, e.g. MELP, MBE, split band LPC or HVXC
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/08—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
- G10L19/12—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters the excitation function being a code excitation, e.g. in code excited linear prediction [CELP] vocoders
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/16—Vocoder architecture
- G10L19/18—Vocoders using multiple modes
- G10L19/20—Vocoders using multiple modes using sound class specific coding, hybrid encoders or object based coding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/038—Speech enhancement, e.g. noise reduction or echo cancellation using band spreading techniques
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/90—Pitch determination of speech signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/08—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
- G10L19/09—Long term prediction, i.e. removing periodical redundancies, e.g. by using adaptive codebook or pitch predictor
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L21/0216—Noise filtering characterised by the method used for estimating noise
- G10L21/0232—Processing in the frequency domain
Definitions
- the present invention relates generally to speech coding and, more particularly, to Code Excited Linear Prediction (CELP) speech coding.
- CELP Code Excited Linear Prediction
- a speech signal can be band-limited to about 10 kHz without affecting its perception.
- the speech signal bandwidth is usually limited much more severely.
- the telephone network limits the bandwidth of the speech signal to between 300 Hz to 3400 Hz, which is known as the “narrowband”.
- Such band-limitation results in the characteristic sound of telephone speech.
- Both the lower limit at 300 Hz and the upper limit at 3400 Hz affect the speech quality.
- the speech signal is sampled at 8 kHz, resulting in a maximum signal bandwidth of 4 kHz.
- the signal is usually band-limited to about 3600 Hz at the high-end.
- the cut-off frequency is usually between 50 Hz and 200 Hz.
- the narrowband speech signal which requires a sampling frequency of 8 kb/s, provides a speech quality referred to as toll quality.
- this toll quality is sufficient for telephone communications, for emerging applications such as teleconferencing, multimedia services and high-definition television, an improved quality is necessary.
- the communications quality can be improved for such applications by increasing the bandwidth. For example, by increasing the sampling frequency to 16 kHz, a wider bandwidth, ranging from 50 Hz to about 7000 Hz can be accommodated, which is referred to as the “wideband”. Extending the lower frequency range to 50 Hz increases naturalness, presence and comfort. At the other end of the spectrum, extending the higher frequency range to 7000 Hz increases intelligibility and makes it easier to differentiate between fricative sounds.
- ABS Analysis-By-Synthesis
- CELP Code Excited Linear Prediction
- LPC linear predictive coding
- the present invention addresses the above analysis-by-synthesis voiced speech issue.
- a voicing index is used to control and improve ABS type speech coding, which indicates the periodicity degree of the speech signal.
- the periodicity degree can significantly vary for different voiced speech segments, and this variation causes an unstable perceptual quality in analysis-by-synthesis type speech coding, such as CELP.
- the voicing index can be used to improve the quality stability by controlling encoder and/or decoder, for example, in the following areas: (a) fixed-codebook short-term enhancement including the spectrum tilt, (b) perceptual weighting filter, (c) sub-fixed codebook determination, (d) LPC interpolation, (e) fixed-codebook pitch enhancement, (f) post-pitch enhancement, (g) noise injection into the high-frequency band at decoder, (h) LTP Sinc window, (i) signal decomposition, etc.
- the voicing index may be based on a normalized pitch correlation.
- FIG. 1 is an illustration of the frequency domain characteristics of a sample speech signal.
- FIG. 2 is an illustration of a voicing index classification available to both the encoder and the decoder.
- FIG. 3 is an illustration of a basic CELP coding block diagram.
- FIG. 4 is an illustration of a CELP coding process with an additional adaptive weighting filter for speech enhancement in accordance with an embodiment of the present invention.
- FIG. 5 is an illustration of a decoder implementation with post filter configuration in accordance with an embodiment of the present invention.
- FIG. 6 is an illustration of a CELP coding block diagram with several sub-codebooks.
- FIG. 7A is an illustration of sampling for creation of a Sinc window.
- FIG. 7B is an illustration of a Sinc window.
- the present application may be described herein in terms of functional block components and various processing steps. It should be appreciated that such functional blocks may be realized by any number of hardware components and/or software components configured to perform the specified functions.
- the present application may employ various integrated circuit components, e.g., memory elements, digital signal processing elements, transmitters, receivers, tone detectors, tone generators, logic elements, and the like, which may carry out a variety of functions under the control of one or more microprocessors or other control devices.
- the present application may employ any number of conventional techniques for data transmission, signaling, signal processing and conditioning, tone generation and detection and the like. Such general techniques that may be known to those skilled in the art are not described in detail herein.
- voicing index is traditionally one of the important indexes sent to the decoder for Harmonic speech coding.
- the voicing index generally represents the degree of periodicity and/or periodic harmonic band boundary of voiced speech. voicing index is traditionally not used in CELP coding systems. However, embodiments of the present invention use the voicing index to provide control and improve the quality of synthesized speech in a CELP or other analysis-by-synthesis type coder.
- FIG. 1 is an illustration of the frequency domain characteristics of a sample speech signal.
- the spectrum domain in the wideband extends from slightly above 0 Hz to around 7 kHz.
- 8 kHz i.e. Nyquist folding frequency
- the speech signal is quite harmonic at lower frequencies, but at higher frequencies the speech signal does not remain as harmonic because the probability of having noisy speech signal increases as the frequency increases.
- the speech signal exhibits traits of becoming noisy at the higher frequencies, e.g., above 5.0 kHz.
- This noisy signal makes waveform matching at higher frequencies very difficult.
- techniques like ABS coding e.g. CELP
- the synthesizer is designed to match the original speech signal by minimizing the error between the original speech and the synthesized speech.
- a noisy signal is unpredictable thus making error minimization very difficult.
- embodiments of the present invention use a voicing index which is sent to the decoder, from the encoder, to improve the quality of speech synthesized by an ABS type speech coder, e.g., CELP coder.
- an ABS type speech coder e.g., CELP coder.
- the voicing index which is transmitted by the encoder to the decoder, may represent the periodicity of the voiced speech or the harmonic structure of the signal.
- the voicing index may be represented by three bits thus providing up to eight classes of speech signal.
- FIG. 2 is an illustration of a voicing index classification available to both the encoder and the decoder.
- index 0 i.e. “000” may indicate background noise
- index 1 i.e. “001”
- index 2 i.e. “010”
- indices 3 - 7 i.e. “011” to “111” could each indicate the periodicity of the speech signals.
- index 3 (“011”) may represent the least periodic signal
- index 7 (“111”) may indicate the most periodic signal.
- the voicing index information can be transmitted by the encoder as part of each encoded frame.
- each frame may include the voicing index bits (e.g. three bits), which indicate the periodicity degree of that particular frame.
- the voicing index for CELP may be based on a normalized pitch correlation parameter, Rp, and may be derived from the following equation: 10 log(1 ⁇ Rp) 2 , where ⁇ 1.0 ⁇ Rp ⁇ 1.0.
- the voicing index may be used for fixed codebook short-term enhancement, including the spectrum tilt.
- FIG. 3 is an illustration of a basic CELP coding block diagram. As illustrated, the CELP coding block 300 comprises the Fixed Codebook 301 , gain block 302 , Pitch filter block 303 , and LPC filter 304 . CELP coding block 300 further comprises comparison block 306 , Weighting Filter block 320 , and Mean Squared Error (MSE) computation block 308 .
- MSE Mean Squared Error
- CELP coding The basic idea behind CELP coding is that Input Speech 307 is compared against the synthesized output 305 to generate error 309 , which is the mean squared error. The computation continues in a closed loop sense with selection of a new coding parameters until error 309 is minimal.
- the decoder synthesizes the speech using similar blocks 301 - 304 (see FIG. 5).
- the encoder passes information to the decoder as needed to select the proper codebook entry, gain, and filters, . . . , etc . . .
- the pitch filter (e.g. 303 ) contribution is heavier than the fixed codebook (e.g. 301 ) contribution.
- an embodiment of the present invention may use the voicing index to place more focus in the high frequency region by implementing an adaptive high pass filter, which is controlled by the value of the voicing index.
- An architecture such as the one shown in FIG. 4 may be implemented.
- Adaptive Filter 310 could be an adaptive filter emphasizing the power in the high frequency region.
- the weighting filter 420 may also be an adaptive filter for improving the CELP coding process.
- the voicing index may be used to select the appropriate Post Filter 520 parameters.
- FIG. 5 is an illustration of the decoder implementation with post filter configuration.
- Post Filter 520 may have several configurations saved in a table, which may be selectable using information in the voicing index.
- the voicing index may be used in conjunction with the perceptual weighting filter of CELP.
- the perceptual weighting filter may be represented by Adaptive filter 420 of FIG. 4, for example.
- waveform matching minimizes the error in the most important portion (i.e. the high energy portion) of the speech signal and ignores low energy area by performing a mean squared error minimization.
- Embodiments of the present invention use an adaptive weighting process to enhance the low energy area.
- the voicing index may be used to define the aggressiveness of the weighting filter 420 depending on the periodicity degree of the frame.
- the voicing index may be used to determine the sub-fixed codebook.
- There are possibly several sub-codebooks for the fixed codebook for example, one sub-codebook 601 with less pulses but higher position resolution, one sub-codebook 602 with more pulses but lower position resolutions, and a noise sub-codebook 603 . Therefore, if the voicing index indicates a noisy signal, then the sub-codebook 602 or noisy sub-codebook 603 can be used; if the voicing index does not indicate a noisy signal, then one of the sub-codebooks (e.g. 601 or 602 ) may be used depending on the degree of periodicity of the given frame.
- the gain block (codebook) 302 may also be applied individually to each sub-codebook in one or more embodiments.
- the voicing index may be used in conjunction with the LPC interpolation.
- the previous LPC is equally important as the current LPC if the location of the interpolated LPC is at the middle between the previous one and the current one.
- the voicing index indicates that the previous frame was unvoiced and the present frame is voiced, then during the LPC interpolation, the LPC interpolation algorithm may favor the current frame more than the previous
- the voicing index may also be used for fixed codebook pitch enhancement.
- the previous pitch gain is used to perform pitch enhancement.
- the voicing index provides information relating to the current frame and, thus, could be a better indicator than the previous pitch gain information.
- the magnitude of the pitch enhancement may be determined based on the voicing index. In other words, the more periodic the frame (based on the voicing index value), the higher the magnitude of the enhancement.
- the voicing index may be used in conjunction with the U.S. patent application Ser. No. 09/365,444, filed Aug. 2, 1999, specification of which is incorporated herein by reference, to determine the magnitude of the enhancements in the bi-directional pitch enhancement system defined therein.
- the voicing index may be used in place of pitch gain for post pitch enhancement. This is advantageous, since, as discussed above, the voicing index may be derived from a normalized pitch correlation value, i.e. Rp, which is typically between 0.0 and 1.0; however, pitch gain may exceed 1.0 and can adversely affect the post pitch enhancement process.
- Rp normalized pitch correlation value
- the voicing index may also be used to determine the amount of noise that should be injected in the high frequency band at the decoder side.
- This embodiment may be used when the input speech is decomposed into a voiced portion and a noise portion as discussed in pending U.S. patent application Ser. No. ______, filed concurrently herewith, entitled “SIGNAL DECOMPOSITION OF VOICED SPEECH FOR CELP SPEECH CODING”, specification of which is incorporated herein by reference.
- the voicing index may also be used to control modification of the Sinc window.
- the Sinc window is used to generate an adaptive codebook contribution vector, i.e. LTP excitation vector, with fractional pitch lag for CELP coding.
- LTP excitation vector i.e. LTP excitation vector
- LTP Long-term prediction or LTP produces the harmonics by taking a previous excitation and copying it to a current subframe according to the pitch period. It should be noted that if a pure copy of the previous excitation is made, then the harmonic is replicated all the way to the end spectrum in the frequency domain. However, that would not be an accurate representation of a true voice signal and especially not in wideband speech coding.
- an adaptive low pass filter is applied to the Sinc interpolation window, since there is a high probability of noise in high frequency area.
- the fixed codebook contributes to coding of the noisy or irregular portion of the speech signal
- a pitch adaptive codebook contributes to the voice or regular portion of the speech signal.
- the adaptive codebook contribution is generated using a Sinc window, which is used due to the fact that the pitch lag can be fractional. If the pitch lag were an integer, one excitation signal could be copied to the next; however, because the pitch lag is fractional, straight copying of the previous excitation signal would not work. After the Sinc window is modified, the straight copying would not work even for integer pitch lag. In order to generate pitch contribution, several samples are taken, as shown in FIG.
- the Sinc window which originally has a symmetric shape, as shown in FIG. 7B.
- the shape in practice depends on the fractional portion of the pitch lag and the adaptive lowpass filter applied to the Sinc window.
- Application of the Sinc window is similar to convolution or filtering, but the Sinc window is a non-causal filter.
- a window signal w(n) is convoluted with the signal s(n) in the time domain, which is an equivalent representation to spectrum of the window W(w) multiplied by the spectrum of the signal S(w) in the frequency domain:
- low passing of the Sinc window is equivalent to low passing the final adaptive codebook contribution (UACB (n)) or excitation signal; however, low passing of the Sinc window is advantageous due to the fact that the Sinc window is shorter than the excitation. Thus, it is easier to modify the Sinc window than the excitation; further more, the filtering of the Sinc window can be pre-calculated and memorized.
- the voicing index may be used to provide information to control modification of the low pass filter for the Sinc window.
- the voicing index may provide information as to whether the harmonic structure is strong or weak. If the harmonic structure is strong, then a weak low pass filter is applied to the Sinc window, and if the harmonic structure is weak, then a strong low pass filter is applied to the Sinc window.
Abstract
Description
- The present application claims the benefit of U.S. provisional application serial No. 60/455,435, filed Mar. 15, 2003, which is hereby fully incorporated by reference in the present application.
- The following co-pending and commonly assigned U.S. patent applications have been filed on the same day as this application, and are incorporated by reference in their entirety:
- U.S. patent application Ser. No. ______, “SIGNAL DECOMPOSITION OF VOICED SPEECH FOR CELP SPEECH CODING,” Attorney Docket Number: 0160112.
- U.S. patent application Ser. No. ______, “SIMPLE NOISE SUPPRESSION MODEL,” Attorney Docket Number: 0160114.
- U.S. patent application Ser. No. ______, “ADAPTIVE CORRELATION WINDOW FOR OPEN-LOOP PITCH,” Attorney Docket Number: 0160115.
- U.S. patent application Ser. No. ______, “RECOVERING AN ERASED VOICE FRAME WITH TIME WARPING,” Attorney Docket Number: 0160116.
- 1. Field of the Invention
- The present invention relates generally to speech coding and, more particularly, to Code Excited Linear Prediction (CELP) speech coding.
- 2. Related Art
- Generally, a speech signal can be band-limited to about 10 kHz without affecting its perception. However, in telecommunications, the speech signal bandwidth is usually limited much more severely. It is known that the telephone network limits the bandwidth of the speech signal to between 300 Hz to 3400 Hz, which is known as the “narrowband”. Such band-limitation results in the characteristic sound of telephone speech. Both the lower limit at 300 Hz and the upper limit at 3400 Hz affect the speech quality.
- In most digital speech coders, the speech signal is sampled at 8 kHz, resulting in a maximum signal bandwidth of 4 kHz. In practice, however, the signal is usually band-limited to about 3600 Hz at the high-end. At the low-end, the cut-off frequency is usually between 50 Hz and 200 Hz. The narrowband speech signal, which requires a sampling frequency of 8 kb/s, provides a speech quality referred to as toll quality. Although this toll quality is sufficient for telephone communications, for emerging applications such as teleconferencing, multimedia services and high-definition television, an improved quality is necessary.
- The communications quality can be improved for such applications by increasing the bandwidth. For example, by increasing the sampling frequency to 16 kHz, a wider bandwidth, ranging from 50 Hz to about 7000 Hz can be accommodated, which is referred to as the “wideband”. Extending the lower frequency range to 50 Hz increases naturalness, presence and comfort. At the other end of the spectrum, extending the higher frequency range to 7000 Hz increases intelligibility and makes it easier to differentiate between fricative sounds.
- Digitally, speech is synthesized by a well-known approach known as Analysis-By-Synthesis (ABS). Analysis-By-Synthesis is also referred to as closed-loop approach or waveform-matching approach. It offers relatively better speech coding quality than other approaches for medium to high bit rates. A known ABS approach is the so-called Code Excited Linear Prediction (CELP). In CELP coding, speech is synthesized by using encoded excitation information to excite a linear predictive coding (LPC) filter. The output of the LPC filter is compared against the voiced speech and used to adjust the filter parameters in a closed loop sense until the best parameters based upon the least error is found. One of the facts influencing CELP coding is that voicing degree can significantly vary for different voiced speech segments thus causing an unstable perceptual quality in the speech coding.
- The present invention addresses the above analysis-by-synthesis voiced speech issue.
- In accordance with the purpose of the present invention as broadly described herein, there is provided systems and methods for improving quality of synthesized speech by using a voicing index to control the speech coding process.
- According to one embodiment of the present invention, a voicing index is used to control and improve ABS type speech coding, which indicates the periodicity degree of the speech signal. The periodicity degree can significantly vary for different voiced speech segments, and this variation causes an unstable perceptual quality in analysis-by-synthesis type speech coding, such as CELP.
- The voicing index can be used to improve the quality stability by controlling encoder and/or decoder, for example, in the following areas: (a) fixed-codebook short-term enhancement including the spectrum tilt, (b) perceptual weighting filter, (c) sub-fixed codebook determination, (d) LPC interpolation, (e) fixed-codebook pitch enhancement, (f) post-pitch enhancement, (g) noise injection into the high-frequency band at decoder, (h) LTP Sinc window, (i) signal decomposition, etc. In one embodiment for CELP speech coding, the voicing index may be based on a normalized pitch correlation.
- These and other aspects of the present invention will become apparent with further reference to the drawings and specification, which follow. It is intended that all such additional systems, methods, features and advantages be included within this description, be within the scope of the present invention, and be protected by the accompanying claims.
- FIG. 1 is an illustration of the frequency domain characteristics of a sample speech signal.
- FIG. 2 is an illustration of a voicing index classification available to both the encoder and the decoder.
- FIG. 3 is an illustration of a basic CELP coding block diagram.
- FIG. 4 is an illustration of a CELP coding process with an additional adaptive weighting filter for speech enhancement in accordance with an embodiment of the present invention.
- FIG. 5 is an illustration of a decoder implementation with post filter configuration in accordance with an embodiment of the present invention.
- FIG. 6 is an illustration of a CELP coding block diagram with several sub-codebooks.
- FIG. 7A is an illustration of sampling for creation of a Sinc window.
- FIG. 7B is an illustration of a Sinc window.
- The present application may be described herein in terms of functional block components and various processing steps. It should be appreciated that such functional blocks may be realized by any number of hardware components and/or software components configured to perform the specified functions. For example, the present application may employ various integrated circuit components, e.g., memory elements, digital signal processing elements, transmitters, receivers, tone detectors, tone generators, logic elements, and the like, which may carry out a variety of functions under the control of one or more microprocessors or other control devices. Further, it should be noted that the present application may employ any number of conventional techniques for data transmission, signaling, signal processing and conditioning, tone generation and detection and the like. Such general techniques that may be known to those skilled in the art are not described in detail herein.
- Voicing index is traditionally one of the important indexes sent to the decoder for Harmonic speech coding. The voicing index generally represents the degree of periodicity and/or periodic harmonic band boundary of voiced speech. Voicing index is traditionally not used in CELP coding systems. However, embodiments of the present invention use the voicing index to provide control and improve the quality of synthesized speech in a CELP or other analysis-by-synthesis type coder.
- FIG. 1 is an illustration of the frequency domain characteristics of a sample speech signal. In this illustration, the spectrum domain in the wideband extends from slightly above 0 Hz to around 7 kHz. Although the highest possible frequency in the spectrum ends at 8 kHz (i.e. Nyquist folding frequency) for a speech signal sampled at 16 kHz, this illustration shows that the energy is almost zero in the area between 7.0 kHz to 8 kHz. It should be apparent to those of skill in the arts that the ranges of signals used herein are for illustration purposes only and that the principles expressed herein are applicable to other signal bands.
- As illustrated in FIG. 1, the speech signal is quite harmonic at lower frequencies, but at higher frequencies the speech signal does not remain as harmonic because the probability of having noisy speech signal increases as the frequency increases. For instance, in this illustration the speech signal exhibits traits of becoming noisy at the higher frequencies, e.g., above 5.0 kHz. This noisy signal makes waveform matching at higher frequencies very difficult. Thus, techniques like ABS coding (e.g. CELP) becomes unreliable if high quality speech is desired. For example, in a CELP coder, the synthesizer is designed to match the original speech signal by minimizing the error between the original speech and the synthesized speech. A noisy signal is unpredictable thus making error minimization very difficult.
- Given the above problem, embodiments of the present invention use a voicing index which is sent to the decoder, from the encoder, to improve the quality of speech synthesized by an ABS type speech coder, e.g., CELP coder.
- The voicing index, which is transmitted by the encoder to the decoder, may represent the periodicity of the voiced speech or the harmonic structure of the signal. In another example embodiment, the voicing index may be represented by three bits thus providing up to eight classes of speech signal. For instance, FIG. 2 is an illustration of a voicing index classification available to both the encoder and the decoder. In this illustration, index0 (i.e. “000”) may indicate background noise, index 1 (i.e. “001”) may indicate noise-like or unvoiced speech signal, index 2 (i.e. “010”) may indicate irregular voiced signal such as voiced signal during onset, and indices 3-7 (i.e. “011” to “111”) could each indicate the periodicity of the speech signals. For instance, index 3 (“011”) may represent the least periodic signal and index 7 (“111”) may indicate the most periodic signal.
- The voicing index information can be transmitted by the encoder as part of each encoded frame. In other words, each frame may include the voicing index bits (e.g. three bits), which indicate the periodicity degree of that particular frame. In one embodiment, the voicing index for CELP may be based on a normalized pitch correlation parameter, Rp, and may be derived from the following equation: 10 log(1−Rp)2, where −1.0<Rp<1.0.
- In one example, the voicing index may be used for fixed codebook short-term enhancement, including the spectrum tilt. FIG. 3 is an illustration of a basic CELP coding block diagram. As illustrated, the
CELP coding block 300 comprises the FixedCodebook 301,gain block 302,Pitch filter block 303, andLPC filter 304.CELP coding block 300 further comprisescomparison block 306,Weighting Filter block 320, and Mean Squared Error (MSE)computation block 308. - The basic idea behind CELP coding is that
Input Speech 307 is compared against thesynthesized output 305 to generateerror 309, which is the mean squared error. The computation continues in a closed loop sense with selection of a new coding parameters untilerror 309 is minimal. - On the receiving side, the decoder synthesizes the speech using similar blocks301-304 (see FIG. 5). Thus, the encoder passes information to the decoder as needed to select the proper codebook entry, gain, and filters, . . . , etc . . .
- In a CELP speech coding system, when the speech signal is more periodic, the pitch filter (e.g.303) contribution is heavier than the fixed codebook (e.g. 301) contribution. As a result, an embodiment of the present invention may use the voicing index to place more focus in the high frequency region by implementing an adaptive high pass filter, which is controlled by the value of the voicing index. An architecture such as the one shown in FIG. 4 may be implemented. For instance,
Adaptive Filter 310 could be an adaptive filter emphasizing the power in the high frequency region. In the illustration, theweighting filter 420 may also be an adaptive filter for improving the CELP coding process. - On the decoder side, the voicing index may be used to select the
appropriate Post Filter 520 parameters. FIG. 5 is an illustration of the decoder implementation with post filter configuration. In one or more embodiments,Post Filter 520 may have several configurations saved in a table, which may be selectable using information in the voicing index. - In another example, the voicing index may be used in conjunction with the perceptual weighting filter of CELP. The perceptual weighting filter may be represented by
Adaptive filter 420 of FIG. 4, for example. As is well known, waveform matching minimizes the error in the most important portion (i.e. the high energy portion) of the speech signal and ignores low energy area by performing a mean squared error minimization. Embodiments of the present invention use an adaptive weighting process to enhance the low energy area. For instance, the voicing index may be used to define the aggressiveness of theweighting filter 420 depending on the periodicity degree of the frame. - In yet another embodiment, as illustrated in FIG. 6, the voicing index may be used to determine the sub-fixed codebook. There are possibly several sub-codebooks for the fixed codebook, for example, one sub-codebook601 with less pulses but higher position resolution, one sub-codebook 602 with more pulses but lower position resolutions, and a
noise sub-codebook 603. Therefore, if the voicing index indicates a noisy signal, then the sub-codebook 602 ornoisy sub-codebook 603 can be used; if the voicing index does not indicate a noisy signal, then one of the sub-codebooks (e.g. 601 or 602) may be used depending on the degree of periodicity of the given frame. Note that the gain block (codebook) 302 may also be applied individually to each sub-codebook in one or more embodiments. - Further, the voicing index may be used in conjunction with the LPC interpolation. For example, during linear interpolation, the previous LPC is equally important as the current LPC if the location of the interpolated LPC is at the middle between the previous one and the current one. Thus, if the voicing index, for example, indicates that the previous frame was unvoiced and the present frame is voiced, then during the LPC interpolation, the LPC interpolation algorithm may favor the current frame more than the previous
- The voicing index may also be used for fixed codebook pitch enhancement. Typically, the previous pitch gain is used to perform pitch enhancement. However, the voicing index provides information relating to the current frame and, thus, could be a better indicator than the previous pitch gain information. The magnitude of the pitch enhancement may be determined based on the voicing index. In other words, the more periodic the frame (based on the voicing index value), the higher the magnitude of the enhancement. For example, the voicing index may be used in conjunction with the U.S. patent application Ser. No. 09/365,444, filed Aug. 2, 1999, specification of which is incorporated herein by reference, to determine the magnitude of the enhancements in the bi-directional pitch enhancement system defined therein.
- As a further example, the voicing index may be used in place of pitch gain for post pitch enhancement. This is advantageous, since, as discussed above, the voicing index may be derived from a normalized pitch correlation value, i.e. Rp, which is typically between 0.0 and 1.0; however, pitch gain may exceed 1.0 and can adversely affect the post pitch enhancement process.
- As another example, the voicing index may also be used to determine the amount of noise that should be injected in the high frequency band at the decoder side. This embodiment may be used when the input speech is decomposed into a voiced portion and a noise portion as discussed in pending U.S. patent application Ser. No. ______, filed concurrently herewith, entitled “SIGNAL DECOMPOSITION OF VOICED SPEECH FOR CELP SPEECH CODING”, specification of which is incorporated herein by reference.
- The voicing index may also be used to control modification of the Sinc window. The Sinc window is used to generate an adaptive codebook contribution vector, i.e. LTP excitation vector, with fractional pitch lag for CELP coding. In wideband speech coding, it is known that strong harmonics appear in the low frequency area of the band and the noisy signals appear in the high frequency area.
- Long-term prediction or LTP produces the harmonics by taking a previous excitation and copying it to a current subframe according to the pitch period. It should be noted that if a pure copy of the previous excitation is made, then the harmonic is replicated all the way to the end spectrum in the frequency domain. However, that would not be an accurate representation of a true voice signal and especially not in wideband speech coding.
- In one embodiment, for wideband speech signal when the previous signal is used to represent the current signal, an adaptive low pass filter is applied to the Sinc interpolation window, since there is a high probability of noise in high frequency area.
- In CELP coding, the fixed codebook contributes to coding of the noisy or irregular portion of the speech signal, and a pitch adaptive codebook contributes to the voice or regular portion of the speech signal. The adaptive codebook contribution is generated using a Sinc window, which is used due to the fact that the pitch lag can be fractional. If the pitch lag were an integer, one excitation signal could be copied to the next; however, because the pitch lag is fractional, straight copying of the previous excitation signal would not work. After the Sinc window is modified, the straight copying would not work even for integer pitch lag. In order to generate pitch contribution, several samples are taken, as shown in FIG. 7A, which are weighted and then added together, where the weights for the samples is called the Sinc window, which originally has a symmetric shape, as shown in FIG. 7B. The shape in practice depends on the fractional portion of the pitch lag and the adaptive lowpass filter applied to the Sinc window. Application of the Sinc window is similar to convolution or filtering, but the Sinc window is a non-causal filter. In the representation shown below, a window signal w(n) is convoluted with the signal s(n) in the time domain, which is an equivalent representation to spectrum of the window W(w) multiplied by the spectrum of the signal S(w) in the frequency domain:
- U ACB(n)=w(n)*s(n)←→W(w)S(w).
- According to the above representation, low passing of the Sinc window is equivalent to low passing the final adaptive codebook contribution (UACB (n)) or excitation signal; however, low passing of the Sinc window is advantageous due to the fact that the Sinc window is shorter than the excitation. Thus, it is easier to modify the Sinc window than the excitation; further more, the filtering of the Sinc window can be pre-calculated and memorized.
- In one embodiment of the present invention, the voicing index may be used to provide information to control modification of the low pass filter for the Sinc window. For instance, the voicing index may provide information as to whether the harmonic structure is strong or weak. If the harmonic structure is strong, then a weak low pass filter is applied to the Sinc window, and if the harmonic structure is weak, then a strong low pass filter is applied to the Sinc window.
- Although the above embodiments of the present application are described with reference to wideband speech signals, the present invention is equally applicable to narrowband speech signals.
- The methods and systems presented above may reside in software, hardware, or firmware on the device, which can be implemented on a microprocessor, digital signal processor, application specific IC, or field programmable gate array (“FPGA”), or any combination thereof, without departing from the spirit of the invention. Furthermore, the present invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described embodiments are to be considered in all respects only as illustrative and not restrictive.
Claims (45)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/799,503 US20040181411A1 (en) | 2003-03-15 | 2004-03-11 | Voicing index controls for CELP speech coding |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US45543503P | 2003-03-15 | 2003-03-15 | |
US10/799,503 US20040181411A1 (en) | 2003-03-15 | 2004-03-11 | Voicing index controls for CELP speech coding |
Publications (1)
Publication Number | Publication Date |
---|---|
US20040181411A1 true US20040181411A1 (en) | 2004-09-16 |
Family
ID=33029999
Family Applications (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/799,504 Expired - Lifetime US7024358B2 (en) | 2003-03-15 | 2004-03-11 | Recovering an erased voice frame with time warping |
US10/799,505 Active 2026-07-14 US7379866B2 (en) | 2003-03-15 | 2004-03-11 | Simple noise suppression model |
US10/799,460 Active 2025-04-08 US7155386B2 (en) | 2003-03-15 | 2004-03-11 | Adaptive correlation window for open-loop pitch |
US10/799,503 Abandoned US20040181411A1 (en) | 2003-03-15 | 2004-03-11 | Voicing index controls for CELP speech coding |
US10/799,533 Active 2026-03-14 US7529664B2 (en) | 2003-03-15 | 2004-03-11 | Signal decomposition of voiced speech for CELP speech coding |
Family Applications Before (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/799,504 Expired - Lifetime US7024358B2 (en) | 2003-03-15 | 2004-03-11 | Recovering an erased voice frame with time warping |
US10/799,505 Active 2026-07-14 US7379866B2 (en) | 2003-03-15 | 2004-03-11 | Simple noise suppression model |
US10/799,460 Active 2025-04-08 US7155386B2 (en) | 2003-03-15 | 2004-03-11 | Adaptive correlation window for open-loop pitch |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/799,533 Active 2026-03-14 US7529664B2 (en) | 2003-03-15 | 2004-03-11 | Signal decomposition of voiced speech for CELP speech coding |
Country Status (4)
Country | Link |
---|---|
US (5) | US7024358B2 (en) |
EP (2) | EP1604352A4 (en) |
CN (1) | CN1757060B (en) |
WO (5) | WO2004084180A2 (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070038440A1 (en) * | 2005-08-11 | 2007-02-15 | Samsung Electronics Co., Ltd. | Method, apparatus, and medium for classifying speech signal and method, apparatus, and medium for encoding speech signal using the same |
US20070213977A1 (en) * | 2006-03-10 | 2007-09-13 | Matsushita Electric Industrial Co., Ltd. | Fixed codebook searching apparatus and fixed codebook searching method |
US20130218578A1 (en) * | 2012-02-17 | 2013-08-22 | Huawei Technologies Co., Ltd. | System and Method for Mixed Codebook Excitation for Speech Coding |
US20150051905A1 (en) * | 2013-08-15 | 2015-02-19 | Huawei Technologies Co., Ltd. | Adaptive High-Pass Post-Filter |
US20160372125A1 (en) * | 2015-06-18 | 2016-12-22 | Qualcomm Incorporated | High-band signal generation |
US10847170B2 (en) | 2015-06-18 | 2020-11-24 | Qualcomm Incorporated | Device and method for generating a high-band signal from non-linearly processed sub-ranges |
US11545143B2 (en) | 2021-05-18 | 2023-01-03 | Boris Fridman-Mintz | Recognition or synthesis of human-uttered harmonic sounds |
Families Citing this family (88)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7742927B2 (en) * | 2000-04-18 | 2010-06-22 | France Telecom | Spectral enhancing method and device |
US20030187663A1 (en) | 2002-03-28 | 2003-10-02 | Truman Michael Mead | Broadband frequency translation for high frequency regeneration |
JP4178319B2 (en) * | 2002-09-13 | 2008-11-12 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Phase alignment in speech processing |
US7933767B2 (en) * | 2004-12-27 | 2011-04-26 | Nokia Corporation | Systems and methods for determining pitch lag for a current frame of information |
US7702502B2 (en) | 2005-02-23 | 2010-04-20 | Digital Intelligence, L.L.C. | Apparatus for signal decomposition, analysis and reconstruction |
US20060282264A1 (en) * | 2005-06-09 | 2006-12-14 | Bellsouth Intellectual Property Corporation | Methods and systems for providing noise filtering using speech recognition |
EP1772855B1 (en) * | 2005-10-07 | 2013-09-18 | Nuance Communications, Inc. | Method for extending the spectral bandwidth of a speech signal |
US7720677B2 (en) * | 2005-11-03 | 2010-05-18 | Coding Technologies Ab | Time warped modified transform coding of audio signals |
KR100900438B1 (en) * | 2006-04-25 | 2009-06-01 | 삼성전자주식회사 | Apparatus and method for voice packet recovery |
US8010350B2 (en) * | 2006-08-03 | 2011-08-30 | Broadcom Corporation | Decimated bisectional pitch refinement |
US8239190B2 (en) * | 2006-08-22 | 2012-08-07 | Qualcomm Incorporated | Time-warping frames of wideband vocoder |
EP2063418A4 (en) * | 2006-09-15 | 2010-12-15 | Panasonic Corp | Audio encoding device and audio encoding method |
GB2444757B (en) * | 2006-12-13 | 2009-04-22 | Motorola Inc | Code excited linear prediction speech coding |
US7521622B1 (en) | 2007-02-16 | 2009-04-21 | Hewlett-Packard Development Company, L.P. | Noise-resistant detection of harmonic segments of audio signals |
ES2394515T3 (en) * | 2007-03-02 | 2013-02-01 | Telefonaktiebolaget Lm Ericsson (Publ) | Methods and adaptations in a telecommunications network |
GB0704622D0 (en) * | 2007-03-09 | 2007-04-18 | Skype Ltd | Speech coding system and method |
CN101320565B (en) * | 2007-06-08 | 2011-05-11 | 华为技术有限公司 | Perception weighting filtering wave method and perception weighting filter thererof |
CN101321033B (en) * | 2007-06-10 | 2011-08-10 | 华为技术有限公司 | Frame compensation process and system |
US20080312916A1 (en) * | 2007-06-15 | 2008-12-18 | Mr. Alon Konchitsky | Receiver Intelligibility Enhancement System |
US8868417B2 (en) * | 2007-06-15 | 2014-10-21 | Alon Konchitsky | Handset intelligibility enhancement system using adaptive filters and signal buffers |
US8015002B2 (en) | 2007-10-24 | 2011-09-06 | Qnx Software Systems Co. | Dynamic noise reduction using linear model fitting |
US8326617B2 (en) | 2007-10-24 | 2012-12-04 | Qnx Software Systems Limited | Speech enhancement with minimum gating |
US8606566B2 (en) * | 2007-10-24 | 2013-12-10 | Qnx Software Systems Limited | Speech enhancement through partial speech reconstruction |
US8296136B2 (en) * | 2007-11-15 | 2012-10-23 | Qnx Software Systems Limited | Dynamic controller for improving speech intelligibility |
EP2242047B1 (en) * | 2008-01-09 | 2017-03-15 | LG Electronics Inc. | Method and apparatus for identifying frame type |
CN101483495B (en) * | 2008-03-20 | 2012-02-15 | 华为技术有限公司 | Background noise generation method and noise processing apparatus |
FR2929466A1 (en) * | 2008-03-28 | 2009-10-02 | France Telecom | DISSIMULATION OF TRANSMISSION ERROR IN A DIGITAL SIGNAL IN A HIERARCHICAL DECODING STRUCTURE |
US20090319263A1 (en) * | 2008-06-20 | 2009-12-24 | Qualcomm Incorporated | Coding of transitional speech frames for low-bit-rate applications |
US8768690B2 (en) | 2008-06-20 | 2014-07-01 | Qualcomm Incorporated | Coding scheme selection for low-bit-rate applications |
US20090319261A1 (en) * | 2008-06-20 | 2009-12-24 | Qualcomm Incorporated | Coding of transitional speech frames for low-bit-rate applications |
MY154452A (en) * | 2008-07-11 | 2015-06-15 | Fraunhofer Ges Forschung | An apparatus and a method for decoding an encoded audio signal |
EP2410522B1 (en) | 2008-07-11 | 2017-10-04 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio signal encoder, method for encoding an audio signal and computer program |
CN101836253B (en) * | 2008-07-11 | 2012-06-13 | 弗劳恩霍夫应用研究促进协会 | Apparatus and method for calculating bandwidth extension data using a spectral tilt controlling framing |
US8532983B2 (en) * | 2008-09-06 | 2013-09-10 | Huawei Technologies Co., Ltd. | Adaptive frequency prediction for encoding or decoding an audio signal |
WO2010028297A1 (en) | 2008-09-06 | 2010-03-11 | GH Innovation, Inc. | Selective bandwidth extension |
US8515747B2 (en) * | 2008-09-06 | 2013-08-20 | Huawei Technologies Co., Ltd. | Spectrum harmonic/noise sharpness control |
WO2010028299A1 (en) * | 2008-09-06 | 2010-03-11 | Huawei Technologies Co., Ltd. | Noise-feedback for spectral envelope quantization |
WO2010031003A1 (en) | 2008-09-15 | 2010-03-18 | Huawei Technologies Co., Ltd. | Adding second enhancement layer to celp based core layer |
WO2010031049A1 (en) * | 2008-09-15 | 2010-03-18 | GH Innovation, Inc. | Improving celp post-processing for music signals |
CN101599272B (en) * | 2008-12-30 | 2011-06-08 | 华为技术有限公司 | Keynote searching method and device thereof |
GB2466668A (en) * | 2009-01-06 | 2010-07-07 | Skype Ltd | Speech filtering |
CN102016530B (en) * | 2009-02-13 | 2012-11-14 | 华为技术有限公司 | Method and device for pitch period detection |
KR101344435B1 (en) | 2009-07-27 | 2013-12-26 | 에스씨티아이 홀딩스, 인크. | System and method for noise reduction in processing speech signals by targeting speech and disregarding noise |
AU2010309894B2 (en) | 2009-10-20 | 2014-03-13 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Multi-mode audio codec and CELP coding adapted therefore |
KR101666521B1 (en) * | 2010-01-08 | 2016-10-14 | 삼성전자 주식회사 | Method and apparatus for detecting pitch period of input signal |
US8321216B2 (en) * | 2010-02-23 | 2012-11-27 | Broadcom Corporation | Time-warping of audio signals for packet loss concealment avoiding audible artifacts |
US8538035B2 (en) | 2010-04-29 | 2013-09-17 | Audience, Inc. | Multi-microphone robust noise suppression |
US8473287B2 (en) | 2010-04-19 | 2013-06-25 | Audience, Inc. | Method for jointly optimizing noise reduction and voice quality in a mono or multi-microphone system |
US8798290B1 (en) | 2010-04-21 | 2014-08-05 | Audience, Inc. | Systems and methods for adaptive signal equalization |
US8781137B1 (en) | 2010-04-27 | 2014-07-15 | Audience, Inc. | Wind noise detection and suppression |
US9245538B1 (en) * | 2010-05-20 | 2016-01-26 | Audience, Inc. | Bandwidth enhancement of speech signals assisted by noise reduction |
US8447595B2 (en) * | 2010-06-03 | 2013-05-21 | Apple Inc. | Echo-related decisions on automatic gain control of uplink speech signal in a communications device |
US20110300874A1 (en) * | 2010-06-04 | 2011-12-08 | Apple Inc. | System and method for removing tdma audio noise |
US8447596B2 (en) | 2010-07-12 | 2013-05-21 | Audience, Inc. | Monaural noise suppression based on computational auditory scene analysis |
US9047875B2 (en) | 2010-07-19 | 2015-06-02 | Futurewei Technologies, Inc. | Spectrum flatness control for bandwidth extension |
US8560330B2 (en) | 2010-07-19 | 2013-10-15 | Futurewei Technologies, Inc. | Energy envelope perceptual correction for high band coding |
CN103229235B (en) * | 2010-11-24 | 2015-12-09 | Lg电子株式会社 | Speech signal coding method and voice signal coding/decoding method |
CN102201240B (en) * | 2011-05-27 | 2012-10-03 | 中国科学院自动化研究所 | Harmonic noise excitation model vocoder based on inverse filtering |
US8774308B2 (en) * | 2011-11-01 | 2014-07-08 | At&T Intellectual Property I, L.P. | Method and apparatus for improving transmission of data on a bandwidth mismatched channel |
US8781023B2 (en) | 2011-11-01 | 2014-07-15 | At&T Intellectual Property I, L.P. | Method and apparatus for improving transmission of data on a bandwidth expanded channel |
EP2774145B1 (en) * | 2011-11-03 | 2020-06-17 | VoiceAge EVS LLC | Improving non-speech content for low rate celp decoder |
WO2013096875A2 (en) * | 2011-12-21 | 2013-06-27 | Huawei Technologies Co., Ltd. | Adaptively encoding pitch lag for voiced speech |
CN105976830B (en) | 2013-01-11 | 2019-09-20 | 华为技术有限公司 | Audio-frequency signal coding and coding/decoding method, audio-frequency signal coding and decoding apparatus |
PL2951815T3 (en) * | 2013-01-29 | 2018-06-29 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio encoders, audio decoders, systems, methods and computer programs using an increased temporal resolution in temporal proximity of onsets or offsets of fricatives or affricates |
EP2830053A1 (en) * | 2013-07-22 | 2015-01-28 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Multi-channel audio decoder, multi-channel audio encoder, methods and computer program using a residual-signal-based adjustment of a contribution of a decorrelated signal |
PT3336839T (en) | 2013-10-31 | 2019-11-04 | Fraunhofer Ges Forschung | Audio decoder and method for providing a decoded audio information using an error concealment modifying a time domain excitation signal |
CN104637486B (en) * | 2013-11-07 | 2017-12-29 | 华为技术有限公司 | The interpolating method and device of a kind of data frame |
US9570095B1 (en) * | 2014-01-17 | 2017-02-14 | Marvell International Ltd. | Systems and methods for instantaneous noise estimation |
PL3462448T3 (en) | 2014-01-24 | 2020-08-10 | Nippon Telegraph And Telephone Corporation | Linear predictive analysis apparatus, method, program and recording medium |
CN110415714B (en) * | 2014-01-24 | 2022-11-25 | 日本电信电话株式会社 | Linear prediction analysis device, linear prediction analysis method, and recording medium |
US9524735B2 (en) * | 2014-01-31 | 2016-12-20 | Apple Inc. | Threshold adaptation in two-channel noise estimation and voice activity detection |
US9697843B2 (en) | 2014-04-30 | 2017-07-04 | Qualcomm Incorporated | High band excitation signal generation |
US9467779B2 (en) | 2014-05-13 | 2016-10-11 | Apple Inc. | Microphone partial occlusion detector |
US10149047B2 (en) * | 2014-06-18 | 2018-12-04 | Cirrus Logic Inc. | Multi-aural MMSE analysis techniques for clarifying audio signals |
CN105335592A (en) * | 2014-06-25 | 2016-02-17 | 国际商业机器公司 | Method and equipment for generating data in missing section of time data sequence |
FR3024582A1 (en) | 2014-07-29 | 2016-02-05 | Orange | MANAGING FRAME LOSS IN A FD / LPD TRANSITION CONTEXT |
CN107113357B (en) * | 2014-12-23 | 2021-05-28 | 杜比实验室特许公司 | Improved method and apparatus relating to speech quality estimation |
US11295753B2 (en) | 2015-03-03 | 2022-04-05 | Continental Automotive Systems, Inc. | Speech quality under heavy noise conditions in hands-free communication |
US9685170B2 (en) * | 2015-10-21 | 2017-06-20 | International Business Machines Corporation | Pitch marking in speech processing |
US9734844B2 (en) * | 2015-11-23 | 2017-08-15 | Adobe Systems Incorporated | Irregularity detection in music |
US10643633B2 (en) * | 2015-12-02 | 2020-05-05 | Nippon Telegraph And Telephone Corporation | Spatial correlation matrix estimation device, spatial correlation matrix estimation method, and spatial correlation matrix estimation program |
US10482899B2 (en) | 2016-08-01 | 2019-11-19 | Apple Inc. | Coordination of beamformers for noise estimation and noise suppression |
US10761522B2 (en) * | 2016-09-16 | 2020-09-01 | Honeywell Limited | Closed-loop model parameter identification techniques for industrial model-based process controllers |
EP3324406A1 (en) | 2016-11-17 | 2018-05-23 | Fraunhofer Gesellschaft zur Förderung der Angewand | Apparatus and method for decomposing an audio signal using a variable threshold |
EP3324407A1 (en) * | 2016-11-17 | 2018-05-23 | Fraunhofer Gesellschaft zur Förderung der Angewand | Apparatus and method for decomposing an audio signal using a ratio as a separation characteristic |
US11602311B2 (en) | 2019-01-29 | 2023-03-14 | Murata Vios, Inc. | Pulse oximetry system |
US11404061B1 (en) * | 2021-01-11 | 2022-08-02 | Ford Global Technologies, Llc | Speech filtering for masks |
CN113872566B (en) * | 2021-12-02 | 2022-02-11 | 成都星联芯通科技有限公司 | Modulation filtering device and method with continuously adjustable bandwidth |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5664055A (en) * | 1995-06-07 | 1997-09-02 | Lucent Technologies Inc. | CS-ACELP speech compression system with adaptive pitch prediction filter gain based on a measure of periodicity |
US5732389A (en) * | 1995-06-07 | 1998-03-24 | Lucent Technologies Inc. | Voiced/unvoiced classification of speech for excitation codebook selection in celp speech decoding during frame erasures |
US5734789A (en) * | 1992-06-01 | 1998-03-31 | Hughes Electronics | Voiced, unvoiced or noise modes in a CELP vocoder |
US5774837A (en) * | 1995-09-13 | 1998-06-30 | Voxware, Inc. | Speech coding system and method using voicing probability determination |
US6138092A (en) * | 1998-07-13 | 2000-10-24 | Lockheed Martin Corporation | CELP speech synthesizer with epoch-adaptive harmonic generator for pitch harmonics below voicing cutoff frequency |
US6141638A (en) * | 1998-05-28 | 2000-10-31 | Motorola, Inc. | Method and apparatus for coding an information signal |
US6260010B1 (en) * | 1998-08-24 | 2001-07-10 | Conexant Systems, Inc. | Speech encoder using gain normalization that combines open and closed loop gains |
US6330533B2 (en) * | 1998-08-24 | 2001-12-11 | Conexant Systems, Inc. | Speech encoder adaptively applying pitch preprocessing with warping of target signal |
US20030028373A1 (en) * | 2001-06-04 | 2003-02-06 | Ananthapadmanabhan Kandhadai | Fast code-vector searching |
US6782367B2 (en) * | 2000-05-08 | 2004-08-24 | Nokia Mobile Phones Ltd. | Method and arrangement for changing source signal bandwidth in a telecommunication connection with multiple bandwidth capability |
US20050065785A1 (en) * | 2000-11-22 | 2005-03-24 | Bruno Bessette | Indexing pulse positions and signs in algebraic codebooks for coding of wideband signals |
US20060206317A1 (en) * | 1998-06-09 | 2006-09-14 | Matsushita Electric Industrial Co. Ltd. | Speech coding apparatus and speech decoding apparatus |
Family Cites Families (58)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4831551A (en) * | 1983-01-28 | 1989-05-16 | Texas Instruments Incorporated | Speaker-dependent connected speech word recognizer |
US4989248A (en) * | 1983-01-28 | 1991-01-29 | Texas Instruments Incorporated | Speaker-dependent connected speech word recognition method |
US4751737A (en) * | 1985-11-06 | 1988-06-14 | Motorola Inc. | Template generation method in a speech recognition system |
US5086475A (en) * | 1988-11-19 | 1992-02-04 | Sony Corporation | Apparatus for generating, recording or reproducing sound source data |
US5371853A (en) * | 1991-10-28 | 1994-12-06 | University Of Maryland At College Park | Method and system for CELP speech coding and codebook for use therewith |
US5765127A (en) * | 1992-03-18 | 1998-06-09 | Sony Corp | High efficiency encoding method |
JP3277398B2 (en) * | 1992-04-15 | 2002-04-22 | ソニー株式会社 | Voiced sound discrimination method |
US5574825A (en) * | 1994-03-14 | 1996-11-12 | Lucent Technologies Inc. | Linear prediction coefficient generation during frame erasure or packet loss |
JP3557662B2 (en) * | 1994-08-30 | 2004-08-25 | ソニー株式会社 | Speech encoding method and speech decoding method, and speech encoding device and speech decoding device |
US5699477A (en) * | 1994-11-09 | 1997-12-16 | Texas Instruments Incorporated | Mixed excitation linear prediction with fractional pitch |
FI97612C (en) * | 1995-05-19 | 1997-01-27 | Tamrock Oy | An arrangement for guiding a rock drilling rig winch |
US5706392A (en) * | 1995-06-01 | 1998-01-06 | Rutgers, The State University Of New Jersey | Perceptual speech coder and method |
DE69732746C5 (en) * | 1996-02-15 | 2020-11-19 | Koninklijke Philips N.V. | SIGNAL TRANSMISSION SYSTEM WITH REDUCED COMPLEXITY |
US5809459A (en) * | 1996-05-21 | 1998-09-15 | Motorola, Inc. | Method and apparatus for speech excitation waveform coding using multiple error waveforms |
JPH1091194A (en) * | 1996-09-18 | 1998-04-10 | Sony Corp | Method of voice decoding and device therefor |
JP3707154B2 (en) | 1996-09-24 | 2005-10-19 | ソニー株式会社 | Speech coding method and apparatus |
JP3707153B2 (en) * | 1996-09-24 | 2005-10-19 | ソニー株式会社 | Vector quantization method, speech coding method and apparatus |
US6014622A (en) * | 1996-09-26 | 2000-01-11 | Rockwell Semiconductor Systems, Inc. | Low bit rate speech coder using adaptive open-loop subframe pitch lag estimation and vector quantization |
EP0878790A1 (en) * | 1997-05-15 | 1998-11-18 | Hewlett-Packard Company | Voice coding system and method |
US6233550B1 (en) * | 1997-08-29 | 2001-05-15 | The Regents Of The University Of California | Method and apparatus for hybrid coding of speech at 4kbps |
US6263312B1 (en) * | 1997-10-03 | 2001-07-17 | Alaris, Inc. | Audio compression and decompression employing subband decomposition of residual signal and distortion reduction |
US6169970B1 (en) * | 1998-01-08 | 2001-01-02 | Lucent Technologies Inc. | Generalized analysis-by-synthesis speech coding method and apparatus |
US6182033B1 (en) * | 1998-01-09 | 2001-01-30 | At&T Corp. | Modular approach to speech enhancement with an application to speech coding |
US6272231B1 (en) * | 1998-11-06 | 2001-08-07 | Eyematic Interfaces, Inc. | Wavelet-based facial motion capture for avatar animation |
JP2002515610A (en) * | 1998-05-11 | 2002-05-28 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Speech coding based on determination of noise contribution from phase change |
GB9811019D0 (en) * | 1998-05-21 | 1998-07-22 | Univ Surrey | Speech coders |
US6173257B1 (en) * | 1998-08-24 | 2001-01-09 | Conexant Systems, Inc | Completed fixed codebook for speech encoder |
JP4249821B2 (en) * | 1998-08-31 | 2009-04-08 | 富士通株式会社 | Digital audio playback device |
US6691084B2 (en) * | 1998-12-21 | 2004-02-10 | Qualcomm Incorporated | Multiple mode variable rate speech coding |
US6308155B1 (en) * | 1999-01-20 | 2001-10-23 | International Computer Science Institute | Feature extraction for automatic speech recognition |
US6453287B1 (en) * | 1999-02-04 | 2002-09-17 | Georgia-Tech Research Corporation | Apparatus and quality enhancement algorithm for mixed excitation linear predictive (MELP) and other speech coders |
US7423983B1 (en) * | 1999-09-20 | 2008-09-09 | Broadcom Corporation | Voice and data exchange over a packet based network |
US6889183B1 (en) * | 1999-07-15 | 2005-05-03 | Nortel Networks Limited | Apparatus and method of regenerating a lost audio segment |
US6691082B1 (en) * | 1999-08-03 | 2004-02-10 | Lucent Technologies Inc | Method and system for sub-band hybrid coding |
US6910011B1 (en) * | 1999-08-16 | 2005-06-21 | Haman Becker Automotive Systems - Wavemakers, Inc. | Noisy acoustic signal enhancement |
US6111183A (en) * | 1999-09-07 | 2000-08-29 | Lindemann; Eric | Audio signal synthesis system based on probabilistic estimation of time-varying spectra |
SE9903223L (en) * | 1999-09-09 | 2001-05-08 | Ericsson Telefon Ab L M | Method and apparatus of telecommunication systems |
US6581032B1 (en) * | 1999-09-22 | 2003-06-17 | Conexant Systems, Inc. | Bitstream protocol for transmission of encoded voice signals |
US6959274B1 (en) * | 1999-09-22 | 2005-10-25 | Mindspeed Technologies, Inc. | Fixed rate speech compression system and method |
US6574593B1 (en) | 1999-09-22 | 2003-06-03 | Conexant Systems, Inc. | Codebook tables for encoding and decoding |
US6636829B1 (en) * | 1999-09-22 | 2003-10-21 | Mindspeed Technologies, Inc. | Speech communication system and method for handling lost frames |
KR20010101422A (en) * | 1999-11-10 | 2001-11-14 | 요트.게.아. 롤페즈 | Wide band speech synthesis by means of a mapping matrix |
FI116643B (en) * | 1999-11-15 | 2006-01-13 | Nokia Corp | Noise reduction |
US20070110042A1 (en) * | 1999-12-09 | 2007-05-17 | Henry Li | Voice and data exchange over a packet based network |
US6766292B1 (en) * | 2000-03-28 | 2004-07-20 | Tellabs Operations, Inc. | Relative noise ratio weighting techniques for adaptive noise cancellation |
US7136810B2 (en) * | 2000-05-22 | 2006-11-14 | Texas Instruments Incorporated | Wideband speech coding system and method |
US20020016698A1 (en) * | 2000-06-26 | 2002-02-07 | Toshimichi Tokuda | Device and method for audio frequency range expansion |
US6990453B2 (en) * | 2000-07-31 | 2006-01-24 | Landmark Digital Services Llc | System and methods for recognizing sound and music signals in high noise and distortion |
US6898566B1 (en) * | 2000-08-16 | 2005-05-24 | Mindspeed Technologies, Inc. | Using signal to noise ratio of a speech signal to adjust thresholds for extracting speech parameters for coding the speech signal |
DE10041512B4 (en) * | 2000-08-24 | 2005-05-04 | Infineon Technologies Ag | Method and device for artificially expanding the bandwidth of speech signals |
US6937904B2 (en) * | 2000-12-13 | 2005-08-30 | Alfred E. Mann Institute For Biomedical Engineering At The University Of Southern California | System and method for providing recovery from muscle denervation |
US20020133334A1 (en) * | 2001-02-02 | 2002-09-19 | Geert Coorman | Time scale modification of digitally sampled waveforms in the time domain |
DE60126513T2 (en) * | 2001-04-24 | 2007-11-15 | Nokia Corp. | METHOD FOR CHANGING THE SIZE OF A CITRIC BUFFER FOR TIME ORIENTATION, COMMUNICATION SYSTEM, RECEIVER SIDE AND TRANSCODER |
US6985857B2 (en) * | 2001-09-27 | 2006-01-10 | Motorola, Inc. | Method and apparatus for speech coding using training and quantizing |
SE521600C2 (en) * | 2001-12-04 | 2003-11-18 | Global Ip Sound Ab | Lågbittaktskodek |
US7283585B2 (en) * | 2002-09-27 | 2007-10-16 | Broadcom Corporation | Multiple data rate communication system |
US7519530B2 (en) * | 2003-01-09 | 2009-04-14 | Nokia Corporation | Audio signal processing |
US7254648B2 (en) * | 2003-01-30 | 2007-08-07 | Utstarcom, Inc. | Universal broadband server system and method |
-
2004
- 2004-03-11 US US10/799,504 patent/US7024358B2/en not_active Expired - Lifetime
- 2004-03-11 EP EP04719809A patent/EP1604352A4/en not_active Withdrawn
- 2004-03-11 CN CN2004800060153A patent/CN1757060B/en not_active Expired - Fee Related
- 2004-03-11 WO PCT/US2004/007581 patent/WO2004084180A2/en active Application Filing
- 2004-03-11 US US10/799,505 patent/US7379866B2/en active Active
- 2004-03-11 US US10/799,460 patent/US7155386B2/en active Active
- 2004-03-11 EP EP04719814A patent/EP1604354A4/en not_active Withdrawn
- 2004-03-11 US US10/799,503 patent/US20040181411A1/en not_active Abandoned
- 2004-03-11 US US10/799,533 patent/US7529664B2/en active Active
- 2004-03-11 WO PCT/US2004/007582 patent/WO2004084182A1/en active Application Filing
- 2004-03-11 WO PCT/US2004/007583 patent/WO2004084181A2/en active Application Filing
- 2004-03-11 WO PCT/US2004/007949 patent/WO2004084467A2/en active Application Filing
- 2004-03-11 WO PCT/US2004/007580 patent/WO2004084179A2/en active Application Filing
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5734789A (en) * | 1992-06-01 | 1998-03-31 | Hughes Electronics | Voiced, unvoiced or noise modes in a CELP vocoder |
US5664055A (en) * | 1995-06-07 | 1997-09-02 | Lucent Technologies Inc. | CS-ACELP speech compression system with adaptive pitch prediction filter gain based on a measure of periodicity |
US5732389A (en) * | 1995-06-07 | 1998-03-24 | Lucent Technologies Inc. | Voiced/unvoiced classification of speech for excitation codebook selection in celp speech decoding during frame erasures |
US5774837A (en) * | 1995-09-13 | 1998-06-30 | Voxware, Inc. | Speech coding system and method using voicing probability determination |
US6141638A (en) * | 1998-05-28 | 2000-10-31 | Motorola, Inc. | Method and apparatus for coding an information signal |
US20060206317A1 (en) * | 1998-06-09 | 2006-09-14 | Matsushita Electric Industrial Co. Ltd. | Speech coding apparatus and speech decoding apparatus |
US6138092A (en) * | 1998-07-13 | 2000-10-24 | Lockheed Martin Corporation | CELP speech synthesizer with epoch-adaptive harmonic generator for pitch harmonics below voicing cutoff frequency |
US6260010B1 (en) * | 1998-08-24 | 2001-07-10 | Conexant Systems, Inc. | Speech encoder using gain normalization that combines open and closed loop gains |
US6330533B2 (en) * | 1998-08-24 | 2001-12-11 | Conexant Systems, Inc. | Speech encoder adaptively applying pitch preprocessing with warping of target signal |
US6782367B2 (en) * | 2000-05-08 | 2004-08-24 | Nokia Mobile Phones Ltd. | Method and arrangement for changing source signal bandwidth in a telecommunication connection with multiple bandwidth capability |
US20050065785A1 (en) * | 2000-11-22 | 2005-03-24 | Bruno Bessette | Indexing pulse positions and signs in algebraic codebooks for coding of wideband signals |
US20030028373A1 (en) * | 2001-06-04 | 2003-02-06 | Ananthapadmanabhan Kandhadai | Fast code-vector searching |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070038440A1 (en) * | 2005-08-11 | 2007-02-15 | Samsung Electronics Co., Ltd. | Method, apparatus, and medium for classifying speech signal and method, apparatus, and medium for encoding speech signal using the same |
US8175869B2 (en) | 2005-08-11 | 2012-05-08 | Samsung Electronics Co., Ltd. | Method, apparatus, and medium for classifying speech signal and method, apparatus, and medium for encoding speech signal using the same |
US20090228266A1 (en) * | 2006-03-10 | 2009-09-10 | Panasonic Corporation | Fixed codebook searching apparatus and fixed codebook searching method |
US20090228267A1 (en) * | 2006-03-10 | 2009-09-10 | Panasonic Corporation | Fixed codebook searching apparatus and fixed codebook searching method |
US7519533B2 (en) * | 2006-03-10 | 2009-04-14 | Panasonic Corporation | Fixed codebook searching apparatus and fixed codebook searching method |
US7949521B2 (en) | 2006-03-10 | 2011-05-24 | Panasonic Corporation | Fixed codebook searching apparatus and fixed codebook searching method |
US7957962B2 (en) | 2006-03-10 | 2011-06-07 | Panasonic Corporation | Fixed codebook searching apparatus and fixed codebook searching method |
US20110202336A1 (en) * | 2006-03-10 | 2011-08-18 | Panasonic Corporation | Fixed codebook searching apparatus and fixed codebook searching method |
US20070213977A1 (en) * | 2006-03-10 | 2007-09-13 | Matsushita Electric Industrial Co., Ltd. | Fixed codebook searching apparatus and fixed codebook searching method |
US8452590B2 (en) | 2006-03-10 | 2013-05-28 | Panasonic Corporation | Fixed codebook searching apparatus and fixed codebook searching method |
US9972325B2 (en) * | 2012-02-17 | 2018-05-15 | Huawei Technologies Co., Ltd. | System and method for mixed codebook excitation for speech coding |
US20130218578A1 (en) * | 2012-02-17 | 2013-08-22 | Huawei Technologies Co., Ltd. | System and Method for Mixed Codebook Excitation for Speech Coding |
US20150051905A1 (en) * | 2013-08-15 | 2015-02-19 | Huawei Technologies Co., Ltd. | Adaptive High-Pass Post-Filter |
US9418671B2 (en) * | 2013-08-15 | 2016-08-16 | Huawei Technologies Co., Ltd. | Adaptive high-pass post-filter |
US20160372125A1 (en) * | 2015-06-18 | 2016-12-22 | Qualcomm Incorporated | High-band signal generation |
US9837089B2 (en) * | 2015-06-18 | 2017-12-05 | Qualcomm Incorporated | High-band signal generation |
US10847170B2 (en) | 2015-06-18 | 2020-11-24 | Qualcomm Incorporated | Device and method for generating a high-band signal from non-linearly processed sub-ranges |
US11437049B2 (en) | 2015-06-18 | 2022-09-06 | Qualcomm Incorporated | High-band signal generation |
US11545143B2 (en) | 2021-05-18 | 2023-01-03 | Boris Fridman-Mintz | Recognition or synthesis of human-uttered harmonic sounds |
Also Published As
Publication number | Publication date |
---|---|
EP1604352A2 (en) | 2005-12-14 |
WO2004084181A2 (en) | 2004-09-30 |
US7155386B2 (en) | 2006-12-26 |
CN1757060B (en) | 2012-08-15 |
EP1604354A2 (en) | 2005-12-14 |
US7529664B2 (en) | 2009-05-05 |
WO2004084179A3 (en) | 2006-08-24 |
WO2004084181A3 (en) | 2004-12-09 |
CN1757060A (en) | 2006-04-05 |
US20040181399A1 (en) | 2004-09-16 |
US7024358B2 (en) | 2006-04-04 |
EP1604352A4 (en) | 2007-12-19 |
WO2004084181B1 (en) | 2005-01-20 |
WO2004084180A3 (en) | 2004-12-23 |
US20040181405A1 (en) | 2004-09-16 |
WO2004084180B1 (en) | 2005-01-27 |
WO2004084182A1 (en) | 2004-09-30 |
US20050065792A1 (en) | 2005-03-24 |
EP1604354A4 (en) | 2008-04-02 |
WO2004084179A2 (en) | 2004-09-30 |
WO2004084180A2 (en) | 2004-09-30 |
WO2004084467A3 (en) | 2005-12-01 |
WO2004084467A2 (en) | 2004-09-30 |
US20040181397A1 (en) | 2004-09-16 |
US7379866B2 (en) | 2008-05-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20040181411A1 (en) | Voicing index controls for CELP speech coding | |
Bessette et al. | The adaptive multirate wideband speech codec (AMR-WB) | |
EP0832482B1 (en) | Speech coder | |
US7151802B1 (en) | High frequency content recovering method and device for over-sampled synthesized wideband signal | |
EP1141946B1 (en) | Coded enhancement feature for improved performance in coding communication signals | |
EP0465057B1 (en) | Low-delay code-excited linear predictive coding of wideband speech at 32kbits/sec | |
EP1232494B1 (en) | Gain-smoothing in wideband speech and audio signal decoder | |
US20030200092A1 (en) | System of encoding and decoding speech signals | |
KR20020052191A (en) | Variable bit-rate celp coding of speech with phonetic classification | |
US6678651B2 (en) | Short-term enhancement in CELP speech coding | |
JPH1097296A (en) | Method and device for voice coding, and method and device for voice decoding | |
US6415252B1 (en) | Method and apparatus for coding and decoding speech | |
Wang et al. | Improved excitation for phonetically-segmented VXC speech coding below 4 kb/s | |
WO2002023536A2 (en) | Formant emphasis in celp speech coding | |
CA2224688C (en) | Speech coder |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MINDSPEED TECHNOLOGIES, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GAO, YANG;REEL/FRAME:015091/0842 Effective date: 20040310 |
|
AS | Assignment |
Owner name: CONEXANT SYSTEMS, INC., CALIFORNIA Free format text: SECURITY INTEREST;ASSIGNOR:MINDSPEED TECHNOLOGIES, INC.;REEL/FRAME:015891/0028 Effective date: 20040917 Owner name: CONEXANT SYSTEMS, INC.,CALIFORNIA Free format text: SECURITY INTEREST;ASSIGNOR:MINDSPEED TECHNOLOGIES, INC.;REEL/FRAME:015891/0028 Effective date: 20040917 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |