WO1999001942A2 - A method of noise reduction in speech signals and an apparatus for performing the method - Google Patents

A method of noise reduction in speech signals and an apparatus for performing the method Download PDF

Info

Publication number
WO1999001942A2
WO1999001942A2 PCT/DK1998/000295 DK9800295W WO9901942A2 WO 1999001942 A2 WO1999001942 A2 WO 1999001942A2 DK 9800295 W DK9800295 W DK 9800295W WO 9901942 A2 WO9901942 A2 WO 9901942A2
Authority
WO
WIPO (PCT)
Prior art keywords
spectrum
signal
noise
speech signal
γçó
Prior art date
Application number
PCT/DK1998/000295
Other languages
French (fr)
Other versions
WO1999001942A3 (en
Inventor
Kjeld Hermansen
Original Assignee
Partran Aps
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Partran Aps filed Critical Partran Aps
Priority to US09/462,232 priority Critical patent/US6510408B1/en
Priority to EP98930656A priority patent/EP0997003A2/en
Priority to AU81021/98A priority patent/AU8102198A/en
Publication of WO1999001942A2 publication Critical patent/WO1999001942A2/en
Publication of WO1999001942A3 publication Critical patent/WO1999001942A3/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Quality & Reliability (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Soundproofing, Sound Blocking, And Sound Damping (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Noise Elimination (AREA)

Abstract

The invention relates to a method of noise reduction in a speech signal, wherein a first spectrum is generated on the basis of the speech signal, and a second spectrum is generated as an estimate of the noise power spectrum. A third spectrum is generated by performing a spectral subtraction of said first and second spectra, and a resulting speech signal is generated on the basis of said third spectrum. A model based representation describing the quasi-stationary part of the speech signal is generated on the basis of the third spectrum. Said model based representation is manipulated, and said resulting speech signal is generated using said manipulated model based representation and a second signal derived from said speech signal. The invention further relates to an apparatus adapted to perform the method.

Description

A method of noise reduction in speech signals and an apparatus for performing the method
The present invention relates to noise reduction in speech signals.
Noise when added to a speech signal can impair the quality of the signal, reduce intelligibility, and increase listener fatigue. It is therefore of great importance to reduce noise in a speech signal in relation to hearing aids, but also in relation to telecommunication.
Various methods of noise reduction in a speech signal are known. These methods include spectral subtraction and other filtering methods e.g. Wiener filtering. Spectral subtraction is a technique for reducing noise in speech signals, which operates by converting a time domain representation of the speech signal into the frequency do- main, e.g. by taking the Fourier transform of segments of the speech signal. Hereby a set of signals representing the short term power spectrum of the speech is obtained. During the speech-free periods, an estimate of the noise power spectrum is generated. The obtained noise power spectrum is subtracted from the speech power spectrum signals in order to obtain a noise reduction. A time domain speech signal is reconstructed using the resulting spectrum, e.g. by use of the inverse Fourier transform. Hereby the time-domain signal is reconstructed from the noise-reduced power spectrum and the unmodified phase spectrum.
Even though this method has been found to be useful, it has the drawback that the noise reduction is based on an estimate of the noise spectrum and is therefore dependent on stationarity in the noise signal to perform optimally. As the noise in a speech signal is often non-stationary, the estimated noise spectrum used for spectral subtraction will be different from the actual noise spectrum during speech activity. This error in noise estimation tends to affect small spectral regions of the output, and will result in short duration random tones in the noise reduced signal. Even though these random noise tones are often a low-energy signal compared to the total energy in the speech signal, the random tone noise tends to be very irritating to listen due to psycho-acoustic effects.
The object of the invention is to provide a method which enables noise reduction in a speech signal, and which avoids the above-mentioned drawbacks of the prior art.
This object is achieved by performing the method as stated in the characterizing portion of claim 1.
The invention is based on the circumstance that a model based representation describing the quasi-stationary part of the speech signal can be generated on the basis of a third spectrum, which is generated by spectral subtraction of a first spectrum generated on the basis of the speech signal and a second spectrum generated as an esti- mate of the noise power spectrum. The spectral subtraction enables the use of model based representation for speech signals including noise, and the model based representation of the quasi-stationary part of the speech signal enables an improved noise reduction compared to methods of prior art, as it enables use of a priori knowledge of speech signals.
This unconventional use of a combination of both traditional and model based methods of noise reduction in a speech signal is advantageous, as it permits smooth ma- nipulation of the speech signal in order to obtain im¬ proved noise reduction without artefacts.
As the model based representation is generated dynamically, i.e. on the fly, movements of the formants in said third spectrum will not affect the quality of the noise reduction, and the method according to the invention is therefore advantageous compared to methods of the prior art .
Preferably, as stated in claim 2, said model based representation includes parameters describing one or more formants in said third spectrum. This is advantageous as the formants, i.e. peaks in the signal spectrum, which are related to the speech, in a said third spectrum contains essential features of the speech signal, and as it is possible to manipulate the formants by using said parameters, and hereby to manipulate the resulting speech signal .
As stated in claim 3, said parameters preferably reflect the resonance frequency, the bandwidth, and the gain at the resonance frequency of said formants in said third spectrum.
In a preferred embodiment, as stated in claim 4, said manipulation includes spectral gaining, which is based on a structure parameter S reflecting structure in the spectrum. Spectral gaining attenuates relatively broad for- mants since these cause unwanted artefacts. This method is based on the fact that man-made speech produces narrow formats in the absence of noise.
In accordance with the embodiment stated in claim 5 said structure parameter S is preferably given by S=B*G, where B is the bandwidth ratio of said formants in said third spectrum, and G is the gain ratio of said formants in said third spectrum.
As stated in claim 6, noise reduction is preferably performed in said second signal. This is advantageous as noise will also be present in said second signal, and a noise reduction in this signal will therefore result in a noise reduction in the resulting signal.
In accordance with the embodiment stated in claim 7, said second signal corresponds to said speech signal. This is advantageous in some cases e.g. when the signal/noise ratio approximately equals 0 dB .
In accordance with the embodiment stated in claim 8, said second signal represents the residual signal, i.e. the non-stationary part of the speech signal such as information reflecting the articulation. This is advantageous in some cases e.g. when the signal/noise ratio approximately equals 6 dB.
As stated in claim 9, various signal elements of said second signal, such as pitch pulses, stop consonants and noise transients, are preferably amplified or attenuated. This is advantageous in some cases e.g. when the signal/noise ratio approximately equals -6 dB .
The present invention also relates to an apparatus for noise reduction in speech signals. The apparatus is characterized by the features defined in the claim 10-19, which are advantageous for the same reasons as described previously in relation to the methods. Finally, the present invention also relates to the use of the method in a hearing aid as described in claim 19.
The invention will be explained more fully by the follow- ing description with reference to the drawing, in which:
Fig. 1 shows a schematic diagram of prior art;
Fig. 2 shows a schematic diagram of one preferred embodi- ent of the present invention;
Fig. 3 illustrates some formants of a speech signal along with some parameters describing one formant;
Fig. 4a shows the dependency between the structure parameter, STRUK, and the bandwidth threshold;
Fig. 4b shows the gain attenuation factor as a function of the bandwidth threshold;
Fig. 5a illustrates the block diagram of an apparatus utilizing the method according to the invention;
Fig. 5b shows some aspects from Fig. 5b in a greater de- tail;
The prior art is described with reference to Fig. 1. The figure illustrate an apparatus, where a speech signal S is connected to the input terminal of a spectrum generat- ing means 1. The output terminal of the spectrum generating means 1 is connected to a spectral subtraction means 5. A measured noise signal N is connected to the input terminal of a noise spectrum generating means 2. The output terminal of the noise spectrum generating means 2 is connected to a second input terminal of the spectral subtraction means 5. The output terminal of the spectral subtraction means 5 is connected to the input terminal of a signal generating means 9. The signal generating means 9 is adapted to generate the resulting speech signal RS, which is connected to the output terminal.
At 1 segments of the speech signal including noise, S, in the time domain are transformed into a representation in the frequency domain, e.g. by use of the FFT (Fast Fourier Transform) . During speech free periods an esti- mate of the noise power spectrum is calculated from a background noise signal, N, and stored at 2. The estimate of the noise power is then subtracted from the spectral representation of the speech signal resulting in yet another spectrum with a reduced amount of noise if a good estimate for the noise power spectrum could be obtained and the background noise has not changed that much since. This is done at 5. This procedure is often called 'Spectral Subtraction' . The resulting spectrum is then transformed back into the time domain at 9, e.g. by the in- verse FFT, thereby generating the resulting speech signal, RS.
Fig. 2 schematically shows an improved method according to a preferred embodiment of the present invention. The figure illustrate an apparatus according to the invention, where a speech signal S is connected to the input terminal of a spectrum generating means 12. The output from the spectrum generating means 12 is connected to a first input terminal of a spectral subtraction means 15. The apparatus also includes a noise spectrum generating means 10 having a input terminal, which is connected to a measured noise signal N, and a output terminal, which is connected to a second input terminal of the spectral subtraction means 15. As shown on the figure, the apparatus also includes a model generating means 17, a model manipulating means 18, and a signal generating means 19, which are connected m series. A second signal generating means 14 has an input terminal, which is also connected to the speech signal, and an output terminal which is connected to a second input terminal of the signal gener- atmg means 19. The signal generating means 19 is adapted to generate the resulting speech signal RS .
At 10 an estimate of the noise power spectrum is calculated from a background noise signal, N, during speech free periods. The estimate is stored for later use. This estimate spectrum is called the second spectrum hereinafter. At 12 segments of the speech signal including noise, S, m the time domain are transformed into a spectral representation, e.g. by the FFT, m the frequency domain. This spectrum is called the first spectrum hereinafter. The second spectrum is then subtracted from the first spectrum at 15, resulting m a noise-reduced spectrum, called the third spectrum hereinafter. This result is not always sufficient or satisfactory as mentioned above. So, m accordance with this invention the third spectrum is used for generating a model based description of the speech signal. This is done at 17, and enables the use of the model based description m noisy environments. The combination of spectral subtraction reduces the noise, thereby enabling the use of a model based description to gam even greater noise reduction.
The model based description ensures simple control of formants, and thereby the essential features of the speech signal, through parameters like the resonance frequency (f), the bandwidth (b) and the ga (g) of each formant (see also Fig. 3) . The model can be derived using known methods, e.g. the method used m the Partran Tool, which is described articles by U. Hartmann, K. Herman- sen and F.K. Fink: "Feature extraction for profoundly deaf people", D.S.P. Group, Institute for Electronic Sys- o
terns, Alborg University, September 1993, and by K. Her- mansen, P. Rubak, U. Hartman and F.K. Fink: "Spectral sharpening of speech signals using the partran tool", Alborg University.
These three parameters, f, b, and g, for each relevant formant capture all the essential features of the quasi- stationary part of a speech signal. These parameters are manipulated at 18 in order to reduce artefact sounds, e.g. "bath tub" sounds, and to reduce the noise even further. Artefacts are distorted sounds with a low signal power and will typically not be removed by any methods according to the prior art. However, these sounds have been found to be very disturbing and irritating to the human ear, which is well-known from various psycho- acoustic tests. The manipulated parameters are then used together with a signal S , which is derived from the original speech signal at 14, in order to obtain a time varying speech signal with reduced noise and artefacts. The resulting f, b, and g parameters are used to form the pulse response for the synthesis filter 19. Convolution of signal S2 and said pulse response forms the resulting speech signal RS .
Fig. 3 illustrates the relation between the individual formants and the parameters f, b and g in greater detail.
In a spectrum of a human speech signal there will always be formants present in the absence of noise, and these will typically have the largest (and the most important formant with respect to intelligibility) formant at the lowest frequency, while the additional formants typically have a decreasing amplitude as their resonance frequency gets bigger. The fact that the biggest formant carries quite a lot of the relevant information enables a human being to understand the speech even if all the other formants have 'drowned' m noise.
Due to the fact that human speech incorporates a given structure for physiological reasons, and the fact that 'ordinary' background noise (e.g. white or pink noise) is highly disorganized/unstructured (A spectrum showing 'ordinary' background (e.g. white) noise would consist of all frequencies present with more or less the same ampli- tude) , a given parameter reflecting the structure of a given sound/speech can characterize the amount of noise present that particular sound/speech. If the sound/speech incorporates a high level of structure then the signal does not contain much noise, since noise is unstructured. A parameter is used order to describe the structure m the speech signal. The but one disclosed this embodiment has been found to be a good and reliable choice. This choice is one of perhaps many and should not limit the present invention. The parameter used m this invention is called STRUK and is defined as:
STRUK =
Figure imgf000011_0001
that is the ratio of the maximum to the minimum value of all of the bandwidths for the available formants multiplied by the ratio of the maximum to the minimum value of all of the gain values for the available formants. In this particular embodiment b is given at the 3dB attenuation from the resonance frequency and g is given at the resonance frequency. Other choices will be apparent to one skilled m the art. The basic idea of spectral gaming is to "punish" great bandwidths, as such are indicators of a missing structure. If STRUK is large (e.g. 100) the spectrum holds little noise, and if STRUK is rela¬ tively small (e.g. 5) the spectrum holds much noise.
Fig. 3 shows two formants (the two to the left) with a resulting model description together with two other formants (the two to the right) that are 'drowned' in noise. Due to the fact described above the model description will be perceived as quite good even though only two formants are included in the model. This makes the method according to the present invention robust.
The parameter STRUK gives an easily modifiable one-valued parameter to determine the level of noise still present in the third spectrum. The model description makes it easy to modify the spectrum in order to remove unwanted artefacts and noise. This is done through the complete control of the parameters describing the formants (f, b and g) . One way to reduce the noise is by 'punishing' formants with a relatively broad bandwidth by attenuating these, since it is in the nature of man-made sound that the formants are relatively narrow. The attenuation is done by using the parameter STRUK and the two relations shown in Figs. 4a and 4b, which show a bandwidth threshold as a function of STRUK (Fig. 4a) and the gain at- tenuation as a function of the bandwidth threshold (Fig. 4b) . Here it is shown that for a large value of STRUK (little noise) the bandwidth threshold is relatively large (e.g. 400 Hz), and thus the gain attenuation only attenuates relative broad formants. For a small value of STRUK (much noise) the bandwidth threshold is relatively small (e.g. 200 Hz) and the gain attenuation attenuates formants even when they are not very broad. That broad formants are attenuated can be seen in Fig. 3. Often it will be the case that the low frequency formants will survive the attenuation, which is desirable since these contain the most information relevant to the human ear, removing the broad formants in the process, which is de¬ sirable as well since these broad formants will often be perceived as artefacts by the human ear.
Again the model based approach with its small number of parameters ensures that a modification can be quite sim¬ ple in order to obtain a noise reduction and/or artefact removal. The model based approach further has the advantage that if one has to transmit a speech signal, then the amount of data needed is greatly reduced by only having a small number of parameters describing the formants and thereby the speech signal.
Fig. 5a illustrates an apparatus according to the inven- tion, where a speech signal connected to the input terminal of pre-emphazising means 50. The output terminal is connected to a input terminals of Hamming weighting sig¬ nal means 52, inverse LPC analysis/filtering means 58, and to a first input terminal of the synthesis filter 74, and the post-emphasizing means 79 adapted to compensate for the effect of the pre-emphasizing means 50 mentioned previously. The output terminal of the Hamming weighted signal means 52 is connected in series to the spectrum generating means 60 adapted, diode-rectifying means 62, spectral subtraction means 64, effect means 66, autocorrelation means 68, LPC model parameters determination means 70, the functional block 76, and to a second input terminal of the synthesis filter 74 and to the input terminal of the autocorrelation means 54. The output termi- nal of the autocorrelation means 54 is connected to LPC model parameters determination means 56. The LPC model parameters are connected to the inverse LPC analysis/filtering means 58. The apparatus further comprises a pitch detection means 72 with an input and an output ter- minal connected to the output terminal of the inverse LPC analysis/filtering means 58 and to a third input terminal of the synthesis filter 74 respectively. The synthesis filter 74 is adapted to select a input signal from one of the input terminals dependent on the noise level. The se¬ lected signal is called the second signal hereinafter. The selection can be performed in several ways. Noise reduction means can be used in order to obtain additional noise reduction in said second signal using known methods if desired.
Fig. 5b illustrates in greater detail the functional block 76, where the input signal is connected in series to: pseudo decomposition means 77, spectral gaining means 78, spectral sharpening means 80 and pseudo composition means 82.
Figs. 5a and 5b illustrate a block diagram of an apparatus utilizing the described method. The signal to be processed is given as x = s + n, where s and n is the signal and noise component, respectively. The signal is pre-emphazised at 50 in order to emphasize signal components with a high frequency in order to be able to access the important information present in these signal compo¬ nents that have a relatively low power.
The basis for an improvement in the SNR (signal to noise ratio) of an observed signal is the presence of one observed signal (from one microphone) . The separation of the signal component and the noise component must thus be based on some knowledge of the signal component as well as the noise component. The overall idea of the invention is the utilization of the inertia conditioned partial stationarity of man-made sounds, as regards both articulation and intonation. The additive noise component, n, is assumed to be "white", pink or a combination thereof, and partly stationary in the second order statistics, but does not contain stationary harmonic components. The basic approach is a separation of the articulation and intonation components via inverse LPC analy¬ sis/filtering 58. This ensures that the residual signal becomes maximally "white" and just contains - in terms of information - intonation components whose variation is assumed to be partly stationary, as mentioned before.
The determination of the articulation components depends on the strength of the noise, a distinction being made between three stages, viz. weak, intermediate and strong noise corresponding to an SNR of +6 dB, 0 dB and -6 dB, respectively.
For weak noise, the model parameters (LPC) 56 are determined on the basis of the autocorrelation function de¬ rived directly from the Hamming weighted signal 52 by the autocorrelation means 54, and non-linear spectral gaining is performed (see the following) in the spectral gaining means 78 according to the PARTRAN concept, see EP publication no. 0 796 489.
For the intermediate and strong noise situation, an indi¬ rect method is used for the determination of the autocor- relation function, which is still the basis for the model based description of articulation.
The indirect determination of the autocorrelation function is based on the relationship between power spectrum and autocorrelation (they are the Fourier transforms of each other) . The Hamming weighted signal is Fourier- transformed with 512 points at 60 and diode-rectified at 62 with a given time constant. The minimum value of this signal is determined and subtracted from the diode recti- fied amplitude spectrum, (where the appearance of the noise spectrum is known a priori, arbitrary noise spectra may be subtracted here. The knowledge may be obtained if it is possible to identify phases in which the signal component is not present) thereby generating an amplitude spectral subtracted spectrum 64 which, following squar- ing, is inverse-Fourier-transformed with a view to determining the autocorrelation function 68. An effect means perform said squaring. By using the autocorrelation the LPC coefficients can be determined 70. These coefficients are used in a pseudo decomposition 77 in order to iden- tify the f, b and g parameters. Then non-linear spectral gaining 78 is performed according to the PARTRAN concept followed by spectral sharpening 80 and pseudo composition 82 in order to obtain a spectrum from the model based de¬ scription.
In all three cases of noise a model based (LPC) description of the articulation is provided. This model spectrum forms the basis for the calculation of the characteristic parameters of the energy maxima, viz. f, b and g parame- ters for each formant.
In connection with the weighting of these energy maxima a control parameter STRUK is developed (see above) , indicating the degree of structure in the observed signal. This parameter is used for spectral gaining 78 according to the PARTRAN concept (see EP publ . no. 0 796 489).
The bandwidth threshold for reduction in the gain is controlled by the parameter STRUK as mentioned above.
The bandwidth threshold changes linearly in the region "intermediate". Each energy maximum is now subjected to gain adjustment depending on the current bandwidth and the current bandwidth threshold. Artefacts in the form of the well-known "bath tub sounds" are eliminated hereby. After spectral gaining 78, spec¬ tral sharpening 80 is performed, comprising adjusting the bandwidth of the energy maxima by the factor band fact.
The thus modified f, b and g parameters (f being unchanged here) are used for forming second order resonators with zero points positioned in Z=l and Z=-l. The pulse response of these resonators coupled in parallel and with alternating signs are used as FIR filter coefficients in the synthesis filter 74 (4-fold interpolation is performed) .
Input signals to the synthesis filter 74 depend on the degree of the noise, a distinction being made here again between weak, intermediate and strong noise.
For weak noise, the residual signal from the inverse filtering 58 is used.
For intermediate noise, the input signal to the inverse filter 58 is used (the pre-emphasized observed signal) . This results in a natural/inherent spectral sharpening, beyond the one currently performed in the PARTRAN trans- position.
In case of strong noise, the jitter on the pulse of the residual signal is of such a nature/size that none of the above signals can be used as input to the synthesis fil- ter 74. It is turned to account here that the intonation of man-made sounds is partly stationary, which is utilized in a modified pitch detection 72 based on a long observation window. A voiced sound detection determines whether pitch is present, and if so, a residual signal consisting of unit pulses of mean spacing is phased in. As a result, the jitter is reduced significantly, and the synthesized signal is less corrupted by noise.
The basic ideas of the described method is to focus on quasi-stationary components in the observed signal. The method identifies these components and "locks" to them as long as they have a suitable strength and stationarity . This applies to both articulation and intonation components. Generally, artefacts are avoided hereby in connec- tion with the filtering of the noise components. Many psycho-acoustic tests indicate that it is related methods which man uses inter alia in noisy environments.
As mentioned before, the method has been developed on the assumption of one observed signal. In the situation where two or more microphones are possible, this per se can give a noise reduction for the noise components in the two signals which correlate with each other. The remaining noise components may subsequently be eliminated via the described method.
Although a preferred embodiment of the present invention has been described and shown, the invention is not limited to it, but may also be embodied in other ways within the scope of the subject-matter defined in the appended claims, for example increase in speech intelligibility/speech comfort by manipulation/weighting of the formants in accordance with their strength/frequency or elimination of speaker dependent components in the speech signal, while maintaining speech intelligibility (speaker scrambling/encryption) .

Claims

P a t e n t C l a i m s
1. A method of noise reduction in a speech signal, wherein
ΓÇó a first spectrum is generated on the basis of the speech signal,
ΓÇó a second spectrum is generated as an estimate of the noise power spectrum, ΓÇó a third spectrum is generated by performing a spectral subtraction of said first and second spectra, and
ΓÇó a resulting speech signal is generated on the basis of said third spectrum, c h a r a c t e r i z e d in that ΓÇó a model based representation describing the quasi-stationary part of the speech signal is generated on the basis of the third spectrum,
ΓÇó said model based representation is manipulated, and
ΓÇó said resulting speech signal is generated using said manipulated model based representation and a second signal derived from said speech signal.
2. A method according to claim 1, c h a r a c t e r i z e d in that said model based representation in- eludes parameters describing one or more formants in said third spectrum.
3. A method according to claim 2, c h a r a c t e r i z e d in that said parameters reflect the resonance frequency, the bandwidth, and the gain at the resonance frequency of said formants in said third spectrum.
4. A method according to one or more of the preceding claims, c h a r a c t e r i z e d in that said manipula- tion includes spectral gaining, which is based on a structure parameter S reflecting the structure in the spectrum.
5. A method according to claim 4, c h a r a c t e - r i z e d in that said structure parameter S is given by S=B*G, where B is the bandwidth ratio of said formants in said third spectrum, and G is the gain ratio of said formants in said third spectrum.
6. A method according to one or more of the preceding claims, c h a r a c t e r i z e d in that noise reduction is performed in said second signal.
7. A method according to one or more of the preceding claims, c h a r a c t e r i z e d in that said second signal corresponds to said speech signal.
8. A method according to one or more of the claims 1-6, c h a r a c t e r i z e d in that said second signal represents the residual signal.
9. A method according to claim 8, c h a r a c t e r i z e d in that various signal elements of said second signal, such as pitch pulse, stop consonants and noise transients, are amplified or attenuated.
10. An apparatus for noise reduction in a speech signal, comprising
ΓÇó spectrum generating means (1,12) adapted to generate a first spectrum on the basis of the speech signal,
ΓÇó noise spectrum generating means (2,10) adapted to generate a second spectrum as an estimate of the noise power spectrum, ΓÇó spectral subtraction means (5,15) adapted to generate a third spectrum by performing spectral subtraction of said first and second spectra, and
ΓÇó signal generating means (9,19) adapted to generate a resulting speech signal on the basis of said third spectrum, c h a r a c t e r i z e d m that the apparatus further comprises
ΓÇó model generating means (17) adapted to generate a model based representation describing the quasi-stationary part of the speech signal on the basis of the third spectrum,
ΓÇó model manipulating means (18) adapted to manipulate said model based representation, ΓÇó a second signal generating means (14) adapted to derive a second signal from said speech signal, and m that said signal generating means (19) is adapted to generate the resulting speech signal using said manipulated model based representation and second signal.
11. An apparatus according to claim 10, c h a r a c t e r i z e d that said model generating means (17) is adapted to generate a model which include parameters describing one or more formants m said third spectrum.
12. An apparatus according to claim 11, c h a r a c t e r i z e d that said parameters reflect the resonance frequency, the bandwidth, and the gam at the resonance frequency of said formants m said third spectrum.
13. An apparatus according to one or more of the claims 10-12, c h a r a c t e r i z e d m that said model manipulating means (18) is adapted to form a structure parameter S which reflects the structure m the spectrum, and is adapted to perform spectral gaining based on said structure parameter S .
14. An apparatus according to claim 13, c h a r a c - t e r i z e d in that said structure parameter S is given by S=B*G, where B is the bandwidth ratio of said formants in said third spectrum, and G is the gain ratio of said formants in said third spectrum.
15. An apparatus according to one or more of the claims 10-14, c h a r a c t e r i z e d in that the apparatus further comprises noise reduction means adapted to perform noise reduction in said second signal.
16. An apparatus according to one or more of the claims 10-15, c h a r a c t e r i z e d in that said speech signal is used as said second signal.
17. An apparatus according to one or more of the claims 10-15, c h a r a c t e r i z e d in that the residual signal is used as said second signal.
18. An apparatus according to claim 17, c h a r a c t e r i z e d in that the apparatus further comprises means (72) adapted to amplify or attenuate various signal elements of said second signal, such as pitch pulses, stop consonants and noise transients.
19. Use of a method or an apparatus according to claims 1-14 in a hearing aid.
PCT/DK1998/000295 1997-07-01 1998-07-01 A method of noise reduction in speech signals and an apparatus for performing the method WO1999001942A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US09/462,232 US6510408B1 (en) 1997-07-01 1998-07-01 Method of noise reduction in speech signals and an apparatus for performing the method
EP98930656A EP0997003A2 (en) 1997-07-01 1998-07-01 A method of noise reduction in speech signals and an apparatus for performing the method
AU81021/98A AU8102198A (en) 1997-07-01 1998-07-01 A method of noise reduction in speech signals and an apparatus for performing the method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
DK0776/97 1997-07-01
DK77697 1997-07-01

Publications (2)

Publication Number Publication Date
WO1999001942A2 true WO1999001942A2 (en) 1999-01-14
WO1999001942A3 WO1999001942A3 (en) 1999-03-25

Family

ID=8097425

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/DK1998/000295 WO1999001942A2 (en) 1997-07-01 1998-07-01 A method of noise reduction in speech signals and an apparatus for performing the method

Country Status (4)

Country Link
US (1) US6510408B1 (en)
EP (1) EP0997003A2 (en)
AU (1) AU8102198A (en)
WO (1) WO1999001942A2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6463408B1 (en) 2000-11-22 2002-10-08 Ericsson, Inc. Systems and methods for improving power spectral estimation of speech signals
WO2004059614A2 (en) * 2002-12-31 2004-07-15 Microsound A/S A method and apparatus for enhancing the perceptual quality of synthesized speech signals

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19747885B4 (en) * 1997-10-30 2009-04-23 Harman Becker Automotive Systems Gmbh Method for reducing interference of acoustic signals by means of the adaptive filter method of spectral subtraction
JP4169921B2 (en) * 2000-09-29 2008-10-22 パイオニア株式会社 Speech recognition system
JP4244514B2 (en) * 2000-10-23 2009-03-25 セイコーエプソン株式会社 Speech recognition method and speech recognition apparatus
JP4352790B2 (en) * 2002-10-31 2009-10-28 セイコーエプソン株式会社 Acoustic model creation method, speech recognition device, and vehicle having speech recognition device
US8326621B2 (en) 2003-02-21 2012-12-04 Qnx Software Systems Limited Repetitive transient noise removal
US7949522B2 (en) 2003-02-21 2011-05-24 Qnx Software Systems Co. System for suppressing rain noise
US8271279B2 (en) 2003-02-21 2012-09-18 Qnx Software Systems Limited Signature noise removal
US7725315B2 (en) * 2003-02-21 2010-05-25 Qnx Software Systems (Wavemakers), Inc. Minimization of transient noises in a voice signal
US8073689B2 (en) * 2003-02-21 2011-12-06 Qnx Software Systems Co. Repetitive transient noise removal
US7885420B2 (en) * 2003-02-21 2011-02-08 Qnx Software Systems Co. Wind noise suppression system
US7895036B2 (en) * 2003-02-21 2011-02-22 Qnx Software Systems Co. System for suppressing wind noise
US8086619B2 (en) * 2003-09-05 2011-12-27 Google Inc. System and method for providing search query refinements
US7444192B2 (en) 2004-10-26 2008-10-28 Aerovironment, Inc. Reactive replenishable device management
WO2006128107A2 (en) * 2005-05-27 2006-11-30 Audience, Inc. Systems and methods for audio signal analysis and modification
US7818168B1 (en) * 2006-12-01 2010-10-19 The United States Of America As Represented By The Director, National Security Agency Method of measuring degree of enhancement to voice signal
US8392181B2 (en) * 2008-09-10 2013-03-05 Texas Instruments Incorporated Subtraction of a shaped component of a noise reduction spectrum from a combined signal

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1996016533A2 (en) * 1994-11-25 1996-06-06 Fink Fleming K Method for transforming a speech signal using a pitch manipulator
SE505156C2 (en) * 1995-01-30 1997-07-07 Ericsson Telefon Ab L M Procedure for noise suppression by spectral subtraction

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB8801014D0 (en) * 1988-01-18 1988-02-17 British Telecomm Noise reduction
SG49709A1 (en) * 1993-02-12 1998-06-15 British Telecomm Noise reduction
US5774846A (en) * 1994-12-19 1998-06-30 Matsushita Electric Industrial Co., Ltd. Speech coding apparatus, linear prediction coefficient analyzing apparatus and noise reducing apparatus
FI100840B (en) * 1995-12-12 1998-02-27 Nokia Mobile Phones Ltd Noise attenuator and method for attenuating background noise from noisy speech and a mobile station
US5937060A (en) * 1996-02-09 1999-08-10 Texas Instruments Incorporated Residual echo suppression
US5933495A (en) * 1997-02-07 1999-08-03 Texas Instruments Incorporated Subband acoustic noise suppression
FR2768547B1 (en) * 1997-09-18 1999-11-19 Matra Communication METHOD FOR NOISE REDUCTION OF A DIGITAL SPEAKING SIGNAL
US6175602B1 (en) * 1998-05-27 2001-01-16 Telefonaktiebolaget Lm Ericsson (Publ) Signal noise reduction by spectral subtraction using linear convolution and casual filtering

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1996016533A2 (en) * 1994-11-25 1996-06-06 Fink Fleming K Method for transforming a speech signal using a pitch manipulator
SE505156C2 (en) * 1995-01-30 1997-07-07 Ericsson Telefon Ab L M Procedure for noise suppression by spectral subtraction

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
IEEE TRANSACTIONS ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, Volume 27, No. 2, April 1979, STEVEN F. BOLL, "Suppression of Acoustic Noise in Speech Using Spectral Subtraction". *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6463408B1 (en) 2000-11-22 2002-10-08 Ericsson, Inc. Systems and methods for improving power spectral estimation of speech signals
WO2004059614A2 (en) * 2002-12-31 2004-07-15 Microsound A/S A method and apparatus for enhancing the perceptual quality of synthesized speech signals
WO2004059614A3 (en) * 2002-12-31 2004-09-23 Microsound As A method and apparatus for enhancing the perceptual quality of synthesized speech signals

Also Published As

Publication number Publication date
US6510408B1 (en) 2003-01-21
WO1999001942A3 (en) 1999-03-25
EP0997003A2 (en) 2000-05-03
AU8102198A (en) 1999-01-25

Similar Documents

Publication Publication Date Title
US6510408B1 (en) Method of noise reduction in speech signals and an apparatus for performing the method
EP2144232B1 (en) Apparatus and methods for enhancement of speech
EP1450353B1 (en) System for suppressing wind noise
Gustafsson et al. Spectral subtraction using reduced delay convolution and adaptive averaging
AU676714B2 (en) Noise reduction
Yegnanarayana et al. Enhancement of reverberant speech using LP residual signal
US6643619B1 (en) Method for reducing interference in acoustic signals using an adaptive filtering method involving spectral subtraction
EP0459362B1 (en) Voice signal processor
US20050288923A1 (en) Speech enhancement by noise masking
JPS63500543A (en) noise suppression system
Kim et al. Nonlinear enhancement of onset for robust speech recognition.
Itoh et al. Environmental noise reduction based on speech/non-speech identification for hearing aids
Habets Single-channel speech dereverberation based on spectral subtraction
O'Shaughnessy Enhancing speech degrated by additive noise or interfering speakers
Jamieson et al. Evaluation of a speech enhancement strategy with normal-hearing and hearing-impaired listeners
Mesgarani et al. Speech enhancement based on filtering the spectrotemporal modulations
JPH11265199A (en) Voice transmitter
Lim Speech enhancement
Graupe et al. Blind adaptive filtering of speech from noise of unknown spectrum using a virtual feedback configuration
Krini et al. Model-based speech enhancement
Wolfe et al. Perceptually motivated approaches to music restoration
Whitmal et al. Wavelet-based noise reduction
JP2001249676A (en) Method for extracting fundamental period or fundamental frequency of periodical waveform with added noise
Yang et al. Environment-Aware Reconfigurable Noise Suppression
Krishnamoorthy et al. Temporal and spectral processing of degraded speech

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AL AM AT AT AU AZ BA BB BG BR BY CA CH CN CU CZ CZ DE DE DK DK EE EE ES FI FI GB GE GH GM GW HR HU ID IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SK SL TJ TM TR TT UA UG US UZ VN YU ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW SD SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR NE SN TD TG

AK Designated states

Kind code of ref document: A3

Designated state(s): AL AM AT AT AU AZ BA BB BG BR BY CA CH CN CU CZ CZ DE DE DK DK EE EE ES FI FI GB GE GH GM GW HR HU ID IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SK SL TJ TM TR TT UA UG US UZ VN YU ZW

AL Designated countries for regional patents

Kind code of ref document: A3

Designated state(s): GH GM KE LS MW SD SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR NE SN TD TG

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 1998930656

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 09462232

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: KR

NENP Non-entry into the national phase

Ref country code: JP

Ref document number: 1999506167

Format of ref document f/p: F

WWP Wipo information: published in national office

Ref document number: 1998930656

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

NENP Non-entry into the national phase

Ref country code: CA

WWW Wipo information: withdrawn in national office

Ref document number: 1998930656

Country of ref document: EP