Next Article in Journal
Tuning of PID/PIDD2 Controllers for Second-Order Oscillatory Systems with Time Delays
Previous Article in Journal
A Specialized Database for Autonomous Vehicles Based on the KITTI Vision Benchmark
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Speech Enhancement Based on Enhanced Empirical Wavelet Transform and Teager Energy Operator

by
Piotr Kuwałek
1,* and
Waldemar Jęśko
2,3
1
Institute of Electrical Engineering and Electronics, Poznan University of Technology, 60-965 Poznan, Poland
2
Institute of Computing Science, Poznan University of Technology, 60-965 Poznan, Poland
3
Poznan Supercomputing and Networking Center, 61-139 Poznan, Poland
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(14), 3167; https://doi.org/10.3390/electronics12143167
Submission received: 22 June 2023 / Revised: 18 July 2023 / Accepted: 19 July 2023 / Published: 21 July 2023

Abstract

:
This paper presents a new speech-enhancement approach based on an enhanced empirical wavelet transform, considering the time and scale adaptation of thresholds for individual component signals obtained from the used transform. The time adaptation is performed using the Teager energy operator on the individual component signals, and the scale adaptation of thresholds is performed by the modified level-dependent threshold principle for the individual component signals. The proposed approach does not require an explicit estimation of the noise level or a priori knowledge of the signal-to-noise ratio as is usually needed in most common speech-enhancement methods. The effectiveness of the proposed method has been assessed based on over 1000 speech recordings from the public Librispeech database. The research included various types of noise (among others white, violet, brown, blue, and pink) and various types of disturbance (among others traffic sounds, hair dryer, and fan), which were added to the selected test signals. The score of perceptual evaluation of speech quality, allowing for the assessment of the quality of enhanced speech, and signal-to-noise ratio, allowing for the assessment of the effectiveness of disturbance attenuation, are selected for the evaluation of the resultant effectiveness of the proposed approach. The resultant effectiveness of the proposed approach is compared with other selected speech-enhancement methods or denoising techniques available in the literature. The experimental research results show that the proposed method performs better than conventional methods in many types of high-noise conditions in terms of producing less residual noise and lower speech distortion.

1. Introduction

Noise has been a problem since the first sound recording devices, such as the Edison phonograph, were developed. These devices recorded voices and other sounds. Unfortunately, while listening to the recordings after information acquisition, one could always observe various unpleasant disturbances occurring in them. Some were caused by the environment, and others (e.g., crackling) were the result of very poorly technologically advanced devices [1]. By the end of the 20th century, technological advances had minimized the mechanical factors of devices causing an obvious signal disturbance. Despite this, noise is still present in modern audio systems, albeit at a low level. All recordings, even those made in hermetic rooms, contain some background noise that is picked up by the microphone. Contemporary microphones have high sensitivity, which results in better clarity and accuracy of the recorded information [2]. Another factor that causes noise in an audio recording is that each part of the recording equipment (e.g., a microphone, amplifier, or mixer that processes the recorded signal) adds a certain electronic noise. Its occurrence in the recording is the result of electric fluctuations that occur as a result of the chaotic movement of charge in electronic components [3]. One of the types of such noise is thermal noise [4]. In addition, disturbance can also be caused by low-level magnetic and electrostatic fields in and around buildings from supply circuits. The method to limit this is to use shielded cables with appropriately designed insulation, most often with XLR connectors [5].
Noise reduction is still under investigation in modern audio engineering. The development of semiconductor technology has contributed to the progress of miniaturization, which, in the area of computer technology, has allowed for the collection of more information and faster processing. Currently, recordings of audio signals can be made using a wide range of available devices. The choice of a specific tool is made mainly in terms of the purpose of registrations and considering the conditions under which the registration is carried out. Presently, disturbances affecting the useful signal can have various characteristics. In an open space, inexpedient sounds can be road and industrial noises, while indoors, air conditioning, household appliances, office machines, or other general-use devices can be noise sources. It happens that the sounds that make daily life easier can also be treated under certain conditions as a disturbance degrading the signal, e.g., loud music. In particular, this problem can be noticed in the currently widely developed voice recognition and monitoring systems, where additional noises can have a negative impact on the correctness of detection, verification, identification, or recognition. Hence, it is important to use various speech-enhancement (noise-reduction) techniques. In the context of speech or speaker recognition based on the pattern recognition process that separates analysis from recognition, speech enhancement already has great potential. Speech enhancement can also be used in coding and with a variety of devices such as audio prostheses.
Considering the processing domain, traditional speech-enhancement methods can be divided into three categories, namely time [6,7,8,9,10], frequency [11,12,13,14] and time–frequency [15,16,17,18] domain methods. Despite the significant development of speech-enhancement techniques [19,20,21,22,23,24,25] from particular categories in recent years, there is still a problem of obtaining a high-quality speech signal with high noise attenuation. Typically, if a method maintains high noise attenuation properties, such a method does not always provide a significant improvement in speech quality. On the other hand, many methods with the potential to significantly increase the quality of speech do not allow high attenuation of the noise signal. The abovementioned problem significantly worsens in the case of highly noisy speech signals. Therefore, considering the presented problems, it can be seen that in recent years there has been a growing tendency to solve speech-enhancement problems using machine-learning methods [26,27,28,29], in particular using various deep-learning techniques [30,31,32,33,34]. However, machine-learning approaches, compared to traditional speech-enhancement techniques, often require significant computational capabilities of the hardware on which the approach is implemented, and it is necessary to create a representative dataset that includes possible numerous real-world cases. It is worth noting that despite the creation of a representative training and validation dataset, the over-fitting of the appropriate model can still occur, which cannot be generalized. Therefore, methods of traditional speech enhancement are still being developed, regardless of machine-learning methods. Hence, the paper presents a new traditional method of speech enhancement based on a new high-efficiency decomposition technique for non-stationary noisy signals, i.e., an enhanced empirical wavelet transform [35], which, in combination with the Teager energy operator [36], allows the high suppression of the noise signal (based on the signal-to-noise ratio (SNR) [37]) and allows for a significant improvement in the quality of speech (based on the perceptual evaluation of speech quality (PESQ) measure [38]), especially in the case of a highly noisy signal of speech. The proposed approach does not require an explicit estimation of the noise level or a priori knowledge of the signal-to-noise ratio, as is usually needed in most common speech-enhancement methods. The presented approach was verified and compared with other available approaches in the literature. Verifications of the proposed approach in the paper were carried out on recordings from one microphone. It is worth noting that the proposed approach has the possibility of multidimensional analysis, and thus, it is possible to apply/extend the proposed approach to the analysis of speech signals from many microphones.

2. Proposed Approach

In the proposed approach shown in Figure 1, the processed speech signal x t is first decomposed into a maximum of N component signals x EEWT k t using the enhanced empirical wavelet transform (EEWT) [35]. Each component signal x EEWT k t is successively processed according to the following steps:
  • Determination of Teager Energy Operator x TEO k t [36];
  • Mask construction M k t based on x TEO k t ;
  • Mask processing M k t ;
  • Determination of the time–space adaptation of thresholds λ k t based on M k t ;
  • Realization of soft thresholding x EEWT k t considering λ k t for the component signal x EEWT k t .
In the last step, the individual processed signals x EEWT k t are summed, as a result of which an enhanced speech signal x t is obtained.

2.1. Enhanced Empirical Wavelet Transform (EEWT)

The EEWT algorithm can be represented in the following steps.
  • The use of fast Fourier transform (FFT) to determine the spectrum of the analyzed signal.
  • The calculation of the upper envelope of the analyzed signal using the l-th order statistical filter (OSF). In the enhanced method [35] (in relation to the conventional empirical wavelet transform (EWT) [39]), the envelope is used to identify the trend of spectrum variation. The number of order l is determined according to the relationship:
    l = n tap · L f s ,
    where n tap is the filter order scaling parameter ( n tap = 70 was adopted in the research—see Section 2.6), L is the length of the analyzed signal, and f s is the sampling frequency of the analyzed signal.
  • The determination of spectrum frequency peaks from the designated envelope and the selection of useful ones based on the following criteria: (a) the width of a flat top cannot be shorter than the order statistics filter size; (b) the most representative flat top in the neighbor ones is picked out; (c) the useful flat tops do not appear in the downtrend of the analyzed signal spectrum.
  • The calculation of the spectrum segmentation boundaries based on the flat tops obtained in Step 3.
  • The construction of the empirical scaling function and empirical wavelet as in the EWT method [39], and the decomposition of the analyzed signal into component signals.
Steps 1–4 allow for the segmentation of the spectrum of the analyzed signal. To segment the spectrum, the segmentation boundaries must be determined. For this purpose, the spectrum is normalized to the range [ 0 ; π ] and is divided into up to N intervals (if the number of useful flat tops N u is smaller than the predefined number of component signals N, then the analyzed signal is decomposed into N u of component signals). The predefined number of component signals N of 32 was adopted in the research—see Section 2.6. The individual interval boundaries are designated as ω k , where ω 0 = 0 and ω min N , N u = π . Each sub-range is marked as k = 1 min N , N u Λ k = [ 0 ; π ] . The boundary determination is based on flat tops described in Step 3. Each boundary is the minimum between subsequent flat tops in the analyzed signal spectrum. If the flat tops are designated as F T k , then:
ω k = arg min f ^ ω ω F T k , F T k + 1 ,
where f ^ ω is the analyzed signal spectrum.
Step 5 allows for the construction of the empirical wavelets, allowing for the extraction of individual component signals as described in [39]. For specific intervals, an empirical scaling function Φ k is constructed described by (3) and an empirical Meyer wavelet Ψ k described by (4) [39]:
Φ k = 1 if ω ω k τ k cos π 2 υ 1 2 τ k ω ω k + τ k if ω k τ k < ω < ω k + τ k 0 otherwise ,
Ψ k = 1 if ω k + τ k ω ω k + 1 τ k + 1 cos π 2 υ 1 2 τ k + 1 ω ω k + 1 + τ k + 1 if ω k + 1 τ k + 1 < ω < ω k + 1 + τ k + 1 sin π 2 υ 1 2 τ k ω ω k + τ k if ω k τ k ω ω k + τ k 0 otherwise ,
where υ ( t ) can be described as:
υ ( t ) = t 4 ( 35 84 t + 70 t 2 20 t 3 ) if 0 < t < 1 0 otherwise .
For the defined empirical wavelet, the τ k can be selected in many ways and determines the appropriate width of the spectrum segment. One of the simplest choices is τ k proportional to ω k , so τ k = ζ ω k , where 0 < ζ < 1 .
The approximate coefficients are the scalar product of the processed signal and the empirical scaling function:
W x ε ( 0 , t ) = x , Φ 1 = x ( τ ) Φ 1 ( τ t ) ¯ d τ .
The detail coefficients are the scalar product of the processed signal and the empirical wavelet:
W x ε ( k , t ) = x , Ψ k = x ( τ ) Ψ k ( τ t ) ¯ d τ .
For the defined approximation coefficients and detail coefficients, signal decomposition defined by empirical wavelet transform can be described as:
x EEWT k t = W x ε ( 0 , t ) Φ 1 ( t ) if k = 0 W x ε ( k , t ) Ψ k ( t ) if k = 1 , 2 , , min N , N u .
In the research, the software provided by the authors of the paper [40] was used to prepare the enhanced EWT method. A method that allows for the calculation of the upper envelope of the analyzed signal using the specific order statistical filter was added to the available software. In addition, based on the obtained envelope, a method was added that allows for the determination of spectrum frequency peaks and selection of useful ones based on the criteria characteristic of EEWT.

2.2. Teager Energy Operator (TEO)

For the time–space-adapted threshold construction, the Teager energy operator (TEO) [36] is used, which for the k-th component signal x EEWT k t (in the discrete domain for sampling moments t i ) is given the following relationship:
x TEO k t i = x EEWT k t i · x EEWT k t i x EEWT k t i + 1 · x EEWT k t i 1 .
The use of the Teager energy operator allows for the increase of the ability to recognize a speech signal from a disturbance (noise) signal.

2.3. Mask Construction

The mask construction for the k-th component signal is performed by smoothing the signal x TEO k t (obtained with the Teager energy operator (TEO)) and then is normalized. Thus, the initial construction of the mask M k t is given by the relationship:
M k t = x TEO k t h k t max x TEO k t h k t ,
where h k t is the impulse response for a second-order Butterworth low-pass filter (infinite impulse response (IIR) filter).

2.4. Mask Processing

The time-scale adaptation of thresholds should be adjusted to the speech waveform; therefore, the difference between the local maxima should be reduced. The processing result should provide a compromise between noise attenuation and speech distortion. Thus, the pre-calculated mask is processed according to the relationship:
M k t = M k t S k max M k t S k γ if S k < S thres 0 if S k S thres ,
where S k is the parameter called offset and is determined as the abscissa of the maximum of the amplitude distribution H of the corresponding mask M k t , and is estimated over the analyzed frame:
S k = abscissa H M k t ,
γ is a parameter that affects the compromise between noise attenuation and speech distortion ( γ = 0.125 was adopted in the research—see Section 2.6); S thres is a parameter that affects the ability to discern speech from silence ( S thres = 0.041 was adopted in the research—see Section 2.6).

2.5. Thresholding

As part of the thresholding, the level-dependent threshold is first determined in accordance with the dependence:
λ k = σ k 2 log L log 2 k + 2 ,
where k = 0 , 1 , 2 , , min N , N u , σ k = M A D k / 0.6745 is the noise level, L is the length of the analyzed signal, M A D k is the median absolute deviation value for x EEWT k t .
In the next step, based on the level-dependent threshold λ k and the processed mask M k t , the time–space adaptation of thresholds λ k t is determined according to the dependence:
λ k t = Γ λ k 1 M k t if P k P < P thres λ k 1 M k t if P k P P thres ,
where P k is the power of the k-th component signal x EEWT k t , P is the power of the processed speech signal x t , P thres is the level-dependent threshold λ k scaling trigger ( P thres = 0.170 was adopted in the research—see Section 2.6); Γ is the scaling parameter of the level-dependent threshold λ k ( Γ = 4.837 was adopted in the research—see Section 2.6).
In the last step, soft thresholding [41] is performed for individual k-th component signals x EEWT k t based on the determined time–space adaptation of thresholds λ k t .

2.6. Final Processing

To obtain an enhanced speech signal x t , the individual processed k-th component signals x EEWT k t are summed according to the relationship:
x t = k = 0 min N , N u x EEWT k t .
It is worth noting that the set of hyperparameters adopted in the research for proposed approach— n tap , N , S thres , γ , P thres , Γ —is the result of the optimization process carried out using Monte Carlo analysis for a subset of the selected test signal database with added “white” noise with different SNR u values. The optimization process was focused on maximizing speech quality. The optimization process included many combinations of hyperparameters randomly selected from the space bounded by a hyperplane: 10 , 3 , 0 , 0.125 , 0.01 , 1 and 500 , 100 , 1 , 3 , 0.3 , 5 . The effectiveness of the proposed approach will likely increase if different types of disturbances are included in the optimization process. Nevertheless, by comparing the proposed approach with the adopted hyperparameter values with other speech-enhancement methods available in the literature, better effectiveness is obtained.

3. Materials

In the carried-out research, the well-known, publicly available Librispeech [42] database was used as a source of clean speech. The Librispeech database is a large-scale corpus of read speech in English with 1000 h of recordings. Its structure is divided into separate subsets, which, by default, are used to train, develop, and evaluate automatic speech recognition systems (“train”, “dev”, and “test” sets, respectively). Due to the extensive collection of recordings in the Librispeech database, only the test subset was used in the experiments, or more precisely the “test-clean” set. This collection contains a total of approx. 2600 speech files, the duration of which ranges from 1 to 35 s (a total of over 5 h of recordings), while approx. 400 recordings are equal to or shorter than 3 s, 800 recordings are equal to or shorter than 4 s, 1000 recordings are equal to or shorter than 5 s, and 2000 recordings are equal to or shorter than 10 s. Initial experiments were shown that consider that each of the methods of speech enhancement or noise reduction described in this paper is time-consuming, even on a limited, selected subset of the Librispeech database (“test-clean”). Therefore, it was decided to use a smaller but representative number of recordings and speakers from the subset of the “test-clean” set, in which recordings with a duration of no longer than 5 s (exactly 1093 speech files; 1 h of recordings) in total, including each of (i.e., 40) speakers from the “test-clean” set), were used. Speech files selected in this way from the Librispeech database were used as a reference set of clean speech, to which were added various types of selected noises and ambient sounds. Speech files prepared in such a way, containing noise, were used as the input data for each of the tested methods of speech enhancement or noise reduction.

4. Results and Discussion

The research used the database of speech signals described in Section 3. For the assumed database of test speech signals, the following types of noise [43] were added successively: white, pink, blue, violet, and brown. The following types of disturbances: traffic noise, fan, a hair dryer (ambient sounds recordings from own resources) with SNR u values equal to 20 dB , 10 dB , 0 dB , −10 dB , −20 dB , where SNR u is unprocessed signal-to-noise ratio determined according to the relationship:
SNR u = 10 log i = 1 L x clean t i 2 i = 1 L x noise t i 2 ,
where x clean t i is the non-distorted (clean) speech signal, x noise t i is the added noise (disturbance) signal, t i is the sampling moments, L is the length of the analyzed signal. Color noise is generated by taking uniform white noise and filtering with a coloring filter to obtain the desired noise spectrum with a power spectral density function given by:
S f = L f f α ,
where α is a real number in the interval 2 , 2 and L f is a positive, slowly varying or constant function. Pink noise ( α = 1 ) has equal energy per octave and the power spectral density of pink noise decreases 3 dB per octave. Blue noise ( α = 1 ) has a power spectral density that increases by 3 dB per octave. Violet noise ( α = 2 ) has a power spectral density that increases by 6 dB per octave. Brown noise ( α = 2 ) has a power spectral density that decreases by 6 dB per octave.
For individual considered noisy speech signals, the process of speech enhancement was implemented using the proposed approach and selected methods of noise reduction or speech enhancement available in the literature, i.e., Karhunen–Loeve transform (KLT)—a generalized subspace approach of KLT [44,45], SSboll—spectral subtraction [13,46], WT-TEO—wavelet speech enhancement based on time-scale adaptation [47], NR—noise reduction using spectral gating [48,49], H-SVD—noise-reduction technique based on Hankel matrix and singular value decomposition [50], EEMD-SVD—noise-reduction technique based on ensemble empirical mode decomposition and singular value decomposition [51].
For the individual speech-enhancement results obtained by the considered methods for the considered noisy speech signals, two measures were determined, i.e., the signal-to-noise ratio (SNR) [37] assessing the ability to suppress noise (disturbance) by the considered methods, and the perceptual evaluation of speech quality (PESQ) score [38] assessing the quality of the processed speech signal. The PESQ score was determined using the available Python library [52], and the SNR was calculated according to the relationship [37]:
SNR = 10 log i = 1 L x clean t i 2 i = 1 L x t i x clean t i 2 ,
where x clean t i is the clean signal without added noise, and x t is the result of the considered methods of speech enhancement or noise reduction.
Figure 2a shows a comparison of the effectiveness of the considered methods for the selected speech signal with the “white” noise with the value of SNR u = 0 dB in the time domain. Based on the presented selected signals in the time domain, it can be seen that the best representation of the original signal (e.g., by comparing the envelope signals) is obtained for the proposed approach as well as for KLT and SSbol.
Figure 2b shows a comparison of the effectiveness of the considered methods for the selected speech signal with the “white” noise with the value of SNR u = 0 dB in the time–frequency domain with the use of spectrograms. Based on the presented selected spectrograms, it can be seen that the best recreation of signal features in the time–frequency domain is obtained for the proposed approach. This property is especially important for the speech recognition process [53,54,55,56,57] in which deep learning is used. In such a process, algorithms can learn from features that are not always relevant to the listener. Thus, obtaining a high PESQ/SNR and recreation of the appropriate envelope cannot always be sufficient. Hence, a new approach has been proposed, which also allows for the reconstruction of important features in the time–frequency domain, thanks to which the use of such a block at the beginning of the signal chain in the speech recognition process allows for a significant increase in the accuracy of speech recognition in a situation when the speech signal is noisy or disturbed, which is the subject of a separate publication. The use of other selected speech-enhancement methods in the process of speech recognition based on deep learning, in most cases deteriorated recognition, is due to the removal of features that do not have to be important from the point of view of the listener, but are important from the point of view of the deep-learning system.
Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7 show a comparison of the distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals, considering “white”, “violet”, “brown”, “blue”, and “pink” noise, respectively. The distribution of the obtained results is presented using the boxplot. Due to the readability of the presented research results, the mean value for individual boxplots is not plotted, because it almost always fluctuated around the median or coincided with the median.
Figure 8, Figure 9 and Figure 10 show a comparison of the distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals, considering disturbances such as traffic sounds, hair dryer, and fan. The distribution of the obtained results is presented using the boxplot. Due to the readability of the presented research results, the mean value for individual boxplots is not plotted, because it almost always fluctuated around the median or coincided with the median. Table 1 shows the maximum and median values of PESQ score for individual considered cases. Table 2 shows the maximum and median values of the SNR for individual considered cases.
Based on the research results presented in Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 regarding the PESQ score, it can be seen that the proposed approach provides the best global enhancement of speech quality for various types of noise and various types of disturbances (ambient sounds). The greatest effectiveness of the proposed approach is visible for the SNR u at the level from 20 dB to 0 dB .
Based on the research results presented in Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 regarding the SNR, it can be seen that the proposed approach provides the best noise (disturbance) attenuation for the analyzed speech signals for various types of noise and various types of disturbances (ambient sounds). In the case of this parameter, the high efficiency of the proposed approach is evident for both a low and a high SNR u value. Additionally, it is worth noting that the dispersion of the research results for the proposed approach is small in the case of SNR, i.e., the resultant mean value, median value, lower and upper quartile, and the minimum and maximum value excluding outliers, and almost coincides with the adopted axis scale considering the dispersion for each analyzed speech enhancement or noise-reduction methods. On this basis, it can be concluded that the high noise (disturbance) attenuation ability will be maintained with a high probability for various types of other noises (disturbances). In the case of other considered methods, the dispersion is much more noticeable, which indicates significantly changeable effectiveness.
The highest efficiency is understood as maintaining the highest median value and the mean value of PESQ and SNR for various types of disturbing signals with different SNR u values. In the case of SNR, both the aforementioned median and mean values as well as the other statistical parameters (e.g., minimum and maximum value) are the highest among all the methods considered in the research. In the case of PESQ, other methods have sometimes obtained higher maximum PESQ values (e.g., SSboll for “pink” noise with SNR u = 20 dB ) or higher minimum PESQ values (e.g., KLT for “white” noise with SNR u = 10 dB ). This situation means that sometimes, as a result of the noise (disturbance) attenuation, the speech quality can be distorted from the listener’s point of view, or sometimes it is observable to obtain a speech signal of better quality for other methods than the proposed approach. It is worth noting that extending the process of optimization of hyperparameters (see Section 2.6), considering other selected noises, could make better speech-enhancement results for the proposed method in selected cases for noise other than “white” noise, where other methods provided better results (e.g., “pink”).

5. Conclusions

The paper presents speech enhancement based on the enhanced empirical wavelet transform and the Teager energy operator. The new proposed approach is less sensitive to large disturbances compared to traditional algorithms and can significantly reduce noise as well as other disturbing signals (ambient sounds). The research results show that the proposed method is good at overall better speech quality, especially in the case of low SNR values for different color noise and different types of disturbance (ambient sounds). At the same time, it should be noted that the proposed method, apart from the general enhancement of the speech quality, in most cases provides the greatest noise (disturbance) reduction for each of the considered types of noise and types of disturbance, even in the case of a significant contribution of disturbance. Additionally, it is worth noting that the dispersion of the research results for the proposed approach is small in the case of SNR, i.e., the resultant mean value, median value, lower and upper quartile, and the minimum and maximum value excluding outliers, and almost coincides with the adopted axis scale considering the dispersion for each analyzed speech enhancement or noise-reduction methods. On this basis, it can be concluded that the high noise (disturbance) attenuation ability will be maintained with a high probability for various types of other noises (disturbances). In the case of other considered methods, the dispersion is much more noticeable, which indicates significantly changeable effectiveness.
The proposed approach does not require an explicit estimation of the noise level or a priori knowledge of the signal-to-noise ratio as is usually needed in most common speech-enhancement methods. The proposed approach allows for the reconstruction of important features in the time–frequency domain, thanks to which the use of such a block at the beginning of the signal chain in the speech recognition process allows for a significant increase in the accuracy of speech recognition in a situation when the speech signal is noisy or disturbed (e.g., the proposed approach can be used in a radio-illumination decision support system for the captain of a ship in distress situations for better identification and understanding of speech when calling for help and contacting maritime emergency services [58,59]). An effective speech recognition process in difficult enhancement conditions using the proposed approach and deep-learning methods is the future direction of research.

Author Contributions

Conceptualization, P.K.; methodology, P.K.; software, P.K. and W.J.; validation, W.J.; formal analysis, P.K.; investigation, P.K. and W.J.; resources, P.K. and W.J.; data curation, P.K. and W.J.; writing—original draft preparation, P.K.; writing—review and editing, P.K.; visualization, P.K.; supervision, W.J.; project administration, W.J.; funding acquisition, P.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by statutory funds of the Faculty of Control, Robotics and Electrical Engineering of the Poznan University of Technology.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to University policy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Recording History. Available online: http://www.recording-history.org (accessed on 1 January 2022).
  2. Lindos Electronics. Available online: http://http://www.lindos.co.uk/ (accessed on 1 January 2022).
  3. Shapley, G.J. Sound of Failure: Experimental Electronic Music in Our Post–Digital Era; University of Technology: Sydney, Australia, 2012. [Google Scholar]
  4. Yang, Y.; SooCho, J.; Lee, B.; Kim, S. A Sound Activity Detector Embedded Low-Power MEMS Microphone Readout Interface for Speech Recognition. In Proceedings of the 2019 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED), Lausanne, Switzerland, 29–31 July 2019; pp. 1–6. [Google Scholar] [CrossRef]
  5. Schneider, M. Electromagnetic interference, microphones and cables. AES J. Audio Eng. Soc. 2005, 6339. Available online: https://www.aes.org/e-lib/browse.cfm?elib=13055 (accessed on 1 January 2023).
  6. Gannot, S.; Burshtein, D.; Weinstein, E. Iterative and sequential Kalman filter-based speech enhancement algorithms. IEEE Trans. Acoust. Speech Signal Process. 1998, 6, 373–385. [Google Scholar] [CrossRef] [Green Version]
  7. Vaseghi, S.V. Advanced Digital Signal Processing and Noise Reduction; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2008. [Google Scholar]
  8. Kim, J.B.; Lee, K.; Lee, C. On the applications of the interacting multiple model algorithm for enhancing noisy speech. IEEE Trans. Acoust. Speech Signal Process. 2000, 8, 349–352. [Google Scholar] [CrossRef]
  9. Dendrinos, M.N.; Bakamidis, S.G.; Carayannis, G. Speech enhancement from noise: A regenerative approach. Speech Commun. 1991, 10, 45–57. [Google Scholar] [CrossRef]
  10. Loizou, P.C. Speech Enhancement: Theory and Practice; CRC Press: Boca Raton, FL, USA, 2013. [Google Scholar]
  11. Ephraim, Y.; Malah, D. Speech enhancement using a minimum-mean square error short-time spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 1984, 32, 1109–1121. [Google Scholar] [CrossRef] [Green Version]
  12. Jax, P.; Vary, P. Artificial bandwidth extension of speech signals using MMSE estimation based on a hidden Markov model. In Proceedings of the 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP ’03), Hong Kong, China, 6–10 April 2003; Volume 1. [Google Scholar] [CrossRef]
  13. Boll, S. Suppression of acoustic noise in speech using spectral subtraction. IEEE Trans. Acoust. Speech Signal Process. 1979, 27, 113–120. [Google Scholar] [CrossRef] [Green Version]
  14. Kwan, C.; Chu, S.; Yin, J.; Liu, X.; Kruger, M.; Sityar, I. Enhanced speech in noisy multiple speaker environment. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–6 June 2008; pp. 1640–1643. [Google Scholar] [CrossRef]
  15. Mallat, S. A Wavelet Tour of Signal Processing; Academic Press: Cambridge, MA, USA, 2009. [Google Scholar]
  16. Virag, N. Single channel speech enhancement based on masking properties of the human auditory system. IEEE Trans. Acoust. Speech Signal Process. 1999, 7, 126–137. [Google Scholar] [CrossRef] [Green Version]
  17. Sun, C.; Zhu, Q.; Wan, M. A novel speech enhancement method based on constrained low-rank and sparse matrix decomposition. Speech Commun. 2014, 60, 44–55. [Google Scholar] [CrossRef]
  18. Sun, C.; Xie, J.; Leng, Y. A Signal Subspace Speech Enhancement Approach Based on Joint Low-Rank and Sparse Matrix Decomposition. Arch. Acoust. 2016, 41, 245–254. [Google Scholar] [CrossRef] [Green Version]
  19. Xian, Y.; Sun, Y.; Wang, W.; Naqvi, S.M. A Multi-Scale Feature Recalibration Network for End-to-End Single Channel Speech Enhancement. IEEE J. Sel. Top. Signal Process. 2021, 15, 143–155. [Google Scholar] [CrossRef]
  20. Wood, S.U.N.; Rouat, J. Unsupervised Low Latency Speech Enhancement With RT-GCC-NMF. IEEE J. Sel. Top. Signal Process. 2019, 13, 332–346. [Google Scholar] [CrossRef] [Green Version]
  21. Chakrabarty, S.; Habets, E.A.P. Time-Frequency Masking Based Online Multi-Channel Speech Enhancement With Convolutional Recurrent Neural Networks. IEEE J. Sel. Top. Signal Process. 2019, 13, 787–799. [Google Scholar] [CrossRef]
  22. Lavanya, T.; Nagarajan, T.; Vijayalakshmi, P. Multi-Level Single-Channel Speech Enhancement Using a Unified Framework for Estimating Magnitude and Phase Spectra. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 1315–1327. [Google Scholar] [CrossRef]
  23. Tu, Y.H.; Du, J.; Lee, C.H. Speech Enhancement Based on Teacher-Student Deep Learning Using Improved Speech Presence Probability for Noise-Robust Speech Recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 2080–2091. [Google Scholar] [CrossRef]
  24. Ming, J.; Crookes, D. Speech Enhancement Based on Full-Sentence Correlation and Clean Speech Recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 531–543. [Google Scholar] [CrossRef] [Green Version]
  25. Kim, M.; Shin, J.W. Improved Speech Enhancement Considering Speech PSD Uncertainty. IEEE/ACM Trans. Audio Speech Lang. Process. 2022, 30, 1939–1951. [Google Scholar] [CrossRef]
  26. Saleem, N.; Khattak, M.I.; Ahmad, S.; Ali, M.Y.; Mohmand, M.I. Machine Learning Approach for Improving the Intelligibility of Noisy Speech. In Proceedings of the 2020 17th International Bhurban Conference on Applied Sciences and Technology (IBCAST), Islamabad, Pakistan, 14–18 January 2020; pp. 303–308. [Google Scholar] [CrossRef]
  27. Choudhury, A.; Roy, P.; Bandyopadhyay, S. Review of Various Machine Learning and Deep Learning Techniques for Audio Visual Automatic Speech Recognition. In Proceedings of the 2023 International Conference on Intelligent Systems, Advanced Computing and Communication (ISACC), Taza, Morocco, 26–27 October 2023; pp. 1–10. [Google Scholar] [CrossRef]
  28. Casey, O.; Dave, R.; Seliya, N.; Sowells Boone, E.R. Machine Learning: Challenges, Limitations, and Compatibility for Audio Restoration Processes. In Proceedings of the 2021 International Conference on Computing, Computational Modelling and Applications (ICCMA), Brest, France, 14–16 July 2021; pp. 27–32. [Google Scholar] [CrossRef]
  29. Ayhan, B.; Kwan, C. Robust Speaker Identification Algorithms and Results in Noisy Environments. In Advances in Neural Networks; Huang, T., Lv, J., Sun, C., Tuzikov, A.V., Eds.; Springer: Cham, Switzerland, 2018; pp. 443–450. [Google Scholar]
  30. Rehr, R.; Gerkmann, T. SNR-Based Features and Diverse Training Data for Robust DNN-Based Speech Enhancement. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 1937–1949. [Google Scholar] [CrossRef]
  31. Zhang, Q.; Nicolson, A.; Wang, M.; Paliwal, K.K.; Wang, C. DeepMMSE: A Deep Learning Approach to MMSE-Based Noise Power Spectral Density Estimation. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 1404–1415. [Google Scholar] [CrossRef]
  32. Takeuchi, D.; Yatabe, K.; Koizumi, Y.; Oikawa, Y.; Harada, N. Real-Time Speech Enhancement Using Equilibriated RNN. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Virtual, 4–9 May 2020; pp. 851–855. [Google Scholar] [CrossRef] [Green Version]
  33. Zhu, Q.S.; Zhang, J.; Zhang, Z.Q.; Dai, L.R. A Joint Speech Enhancement and Self-Supervised Representation Learning Framework for Noise-Robust Speech Recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2023, 31, 1927–1939. [Google Scholar] [CrossRef]
  34. Shifas, M.P.; Zorila, C.; Stylianou, Y. End-to-End Neural Based Modification of Noisy Speech for Speech-in-Noise Intelligibility Improvement. IEEE/ACM Trans. Audio, Speech, Lang. Process. 2022, 30, 162–173. [Google Scholar] [CrossRef]
  35. Hu, Y.; Li, F.; Li, H.G.; Liu, C. An enhanced empirical wavelet transform for noisy and non-stationary signal processing. Dig. Signal Process. 2017, 60, 220–229. [Google Scholar] [CrossRef]
  36. Kaiser, J. On a simple algorithm to calculate the ’energy’ of a signal. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, Albuquerque, NM, USA, 3–6 April 1990; Volume 1, pp. 381–384. [Google Scholar] [CrossRef]
  37. Deller, J.; Hansen, J.; Proakis, J. Discrete-Time Processing of Speech Signals; Wiley-IEEE Press: Hoboken, NJ, USA, 2000. [Google Scholar]
  38. Rix, A.; Beerends, J.; Hollier, M.; Hekstra, A. Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs. In Proceedings of the 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing, Proceedings (Cat. No.01CH37221). Salt Lake City, UT, USA, 7–11 May 2001; Volume 2, pp. 749–752. [Google Scholar] [CrossRef]
  39. Gilles, J. Empirical Wavelet Transform. IEEE Trans. Signal Process. 2013, 61, 3999–4010. [Google Scholar] [CrossRef]
  40. Carvalho, V.R.; Moraes, M.F.; Braga, A.P.; Mendes, E.M. Evaluating five different adaptive decomposition methods for EEG signal seizure detection and classification. Biomed. Signal Process. Control 2020, 62, 102073. [Google Scholar] [CrossRef]
  41. Donoho, D.; Johnstone, I. Ideal Spatial Adaptation via Wavelet Shrinkage. Biometrika 1994, 81, 425–455. [Google Scholar] [CrossRef]
  42. Panayotov, V.; Chen, G.; Povey, D.; Khudanpur, S. Librispeech: An ASR corpus based on public domain audio books. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, Australia, 19–24 April 2015; pp. 5206–5210. [Google Scholar] [CrossRef]
  43. Michael; Li, E.X.D. scivision/Soothing-Sounds: Src/Layout, Black Format, Type Anno. 2021. Available online: https://zenodo.org/record/5574886 (accessed on 1 January 2022).
  44. Hu, Y.; Loizou, P. A generalized subspace approach for enhancing speech corrupted by colored noise. IEEE Trans. Speech Audio Process. 2003, 11, 334–341. [Google Scholar] [CrossRef] [Green Version]
  45. Scheibler, R.; Bezzam, E.; Dokmanic, I. Pyroomacoustics: A Python Package for Audio Room Simulation and Array Processing Algorithms. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 351–355. [Google Scholar] [CrossRef] [Green Version]
  46. Djahnine, A. Suppression-of-Acoustic-Noise-in-Speech-Using-Spectral-Subtraction. 2021. Available online: https://github.com/AissamDjahnine/Suppression-of-Acoustic-Noise-in-Speech-Using-Spectral-Subtraction- (accessed on 1 January 2022).
  47. Bahoura, M.; Rouat, J. Wavelet Speech Enhancement Based on Time-Scale Adaptation. Speech Commun. 2006, 48, 1620–1637. [Google Scholar] [CrossRef] [Green Version]
  48. Sainburg, T.; Thielk, M.; Gentner, T.Q. Finding, visualizing, and quantifying latent structure across diverse animal vocal repertoires. PLoS Comput. Biol. 2020, 16, e1008228. [Google Scholar] [CrossRef]
  49. Sainburg, T. timsainb/noisereduce: V1.0. 2019. Available online: https://zenodo.org/record/3243139 (accessed on 1 January 2022). [CrossRef]
  50. Yang, Y.; Rao, J. Robust and Efficient Harmonics Denoising in Large Dataset Based on Random SVD and Soft Thresholding. IEEE Access 2019, 7, 77607–77617. [Google Scholar] [CrossRef]
  51. Yang, Z.X.; Zhong, J.H. A Hybrid EEMD-Based SampEn and SVD for Acoustic Signal Processing and Fault Diagnosis. Entropy 2016, 18, 112. [Google Scholar] [CrossRef] [Green Version]
  52. PESQ (Perceptual Evaluation of Speech Quality). Wrapper for Python Users. Available online: https://pypi.org/project/pesq/ (accessed on 1 January 2022).
  53. Emiru, E.D.; Li, Y.; Xiong, S.; Fesseha, A. Speech Recognition System Based on Deep Neural Network Acoustic Modeling for Low Resourced Language-Amharic. In Proceedings of the 3rd International Conference on Telecommunications and Communication Engineering (ICTCE), Tokyo, Japan, 9–12 November 2019; pp. 141–145. [Google Scholar] [CrossRef]
  54. Cecko, R.; Jamrozy, J.; Jesko, W.; Kusmierek, E.; Lange, M.; Owsianny, M. Automatic Speech Recognition and its Application to Media Monitoring. Comput. Methods Sci. Technol. CMST 2021, 27, 41–55. [Google Scholar] [CrossRef]
  55. Jesko, W. Vocalization Recognition of People with Profound Intellectual and Multiple Disabilities (PIMD) Using Machine Learning Algorithms. In Proceedings of the Interspeech 2021, Brno, Czech Republic, 30 August–3 September 2021; pp. 2921–2925. [Google Scholar] [CrossRef]
  56. Ravanelli, M.; Parcollet, T.; Plantinga, P.; Rouhe, A.; Cornell, S.; Lugosch, L.; Subakan, C.; Dawalatabad, N.; Heba, A.; Zhong, J.; et al. SpeechBrain: A General-Purpose Speech Toolkit. arXiv 2021, arXiv:2106.04624. [Google Scholar]
  57. Pan, J.; Liu, C.; Wang, Z.; Hu, Y.; Jiang, H. Investigation of deep neural networks (DNN) for large vocabulary continuous speech recognition: Why DNN surpasses GMMS in acoustic modeling. In Proceedings of the 2012 8th International Symposium on Chinese Spoken Language Processing, Hong Kong, China, 5–8 December 2012; pp. 301–305. [Google Scholar] [CrossRef]
  58. Banaszek, A.; Lisaj, A. The Concept of Intelligent Radiocommunication System for Support Decision of Yacht Captains in distress situations with use of neural network computer systems. Procedia Comput. Sci. 2022, 207, 398–407. [Google Scholar] [CrossRef]
  59. Banaszek, A.; Lisaj, A. Advanced methodology for multi-way transmission of ship data treatment from mechanical-navigational technical state sensors with using computational neural network computer systems. Procedia Comput. Sci. 2022, 207, 388–397. [Google Scholar] [CrossRef]
Figure 1. Diagram of the proposed speech-enhancement approach.
Figure 1. Diagram of the proposed speech-enhancement approach.
Electronics 12 03167 g001
Figure 2. The comparison of the effectiveness of the considered methods for the selected speech signal with the “white” noise with the value of SNR u = 0 dB in the time (a) and the time–frequency (b) domain.
Figure 2. The comparison of the effectiveness of the considered methods for the selected speech signal with the “white” noise with the value of SNR u = 0 dB in the time (a) and the time–frequency (b) domain.
Electronics 12 03167 g002
Figure 3. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “white” noise.
Figure 3. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “white” noise.
Electronics 12 03167 g003
Figure 4. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “violet” noise.
Figure 4. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “violet” noise.
Electronics 12 03167 g004
Figure 5. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “brown” noise.
Figure 5. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “brown” noise.
Electronics 12 03167 g005
Figure 6. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “blue” noise.
Figure 6. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “blue” noise.
Electronics 12 03167 g006
Figure 7. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “pink” noise.
Figure 7. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with “pink” noise.
Electronics 12 03167 g007
Figure 8. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with ambient sound like “traffic sounds”.
Figure 8. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with ambient sound like “traffic sounds”.
Electronics 12 03167 g008
Figure 9. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with ambient sound like “hair dryer”.
Figure 9. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with ambient sound like “hair dryer”.
Electronics 12 03167 g009
Figure 10. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with ambient sound like “fan”.
Figure 10. The distribution of PESQ and SNR results for the proposed approach and other considered methods for analyzed speech signals with ambient sound like “fan”.
Electronics 12 03167 g010
Table 1. The maximum and median values of PESQ score for individual considered cases.
Table 1. The maximum and median values of PESQ score for individual considered cases.
Disturbance TypeSNRuThe Maximum ValueThe Median Value
EEWT-TOKLTSSbollWT-TONRH-SVDEEMD-SVDEEWT-TOKLTSSbollWT-TONRH-SVDEEMD-SVD
“white”
noise
203.902.004.013.503.812.552.282.971.422.822.572.961.441.51
103.001.912.892.612.712.571.912.311.442.011.931.961.471.40
02.241.501.771.902.182.241.551.711.211.371.451.631.481.24
−101.861.461.241.611.681.871.481.361.221.101.271.311.271.19
−201.491.601.201.391.401.311.421.191.071.051.181.051.081.16
“violet”
noise
204.451.943.933.973.652.582.513.291.512.683.032.741.461.57
103.781.752.493.233.202.562.092.791.401.692.362.281.431.46
03.101.601.862.492.652.571.642.231.241.301.811.871.471.26
−102.051.531.441.681.881.841.481.541.221.151.291.531.231.26
−201.701.501.271.501.251.201.511.321.181.101.261.081.091.19
“brown”
noise
204.501.503.014.363.812.552.974.111.201.713.142.731.441.71
104.501.502.034.283.872.682.944.021.201.343.132.731.441.70
04.001.531.513.743.302.342.133.201.231.152.652.371.121.50
−103.452.051.202.582.851.173.002.371.421.091.781.971.061.84
−202.201.801.141.691.761.111.371.421.251.031.261.321.081.23
“blue”
noise
204.002.003.773.703.592.422.293.061.312.772.772.481.401.50
103.782.132.963.223.252.702.132.831.361.982.332.281.491.47
02.501.751.872.072.312.361.531.871.281.351.581.691.431.21
−101.831.661.411.671.752.051.411.401.321.161.321.351.261.17
−201.521.511.271.201.661.271.511.231.211.101.081.241.081.19
“pink”
noise
204.421.913.454.283.782.832.813.351.281.963.222.721.531.69
103.101.702.032.912.932.662.092.371.281.372.162.061.461.44
02.001.471.471.751.912.491.571.631.141.161.391.511.281.23
−101.551.491.701.421.621.191.521.241.171.191.141.271.081.20
−201.331.431.691.171.651.171.361.151.051.051.061.111.081.11
traffic
sounds
204.171.353.073.843.402.582.763.221.182.082.962.431.471.70
102.891.442.072.792.472.652.242.311.231.512.141.881.521.58
01.921.381.561.851.721.831.611.521.141.181.431.371.141.27
−101.401.301.901.401.261.521.511.271.111.131.151.101.091.15
−201.471.271.601.441.281.331.601.261.081.131.121.101.071.13
hair
dryer
204.001.362.963.703.602.712.643.051.142.122.712.561.491.58
102.801.331.962.382.542.601.912.121.131.411.781.901.451.37
01.911.311.341.741.791.901.541.591.111.131.361.411.401.23
−101.621.851.901.581.411.901.631.421.161.211.161.151.391.13
−201.361.351.441.251.511.601.311.221.081.101.071.111.181.09
fan203.561.332.793.253.172.542.412.691.111.912.492.301.431.59
102.521.321.962.372.362.501.891.951.121.411.741.771.411.39
01.951.371.541.901.972.051.711.551.131.171.421.491.301.30
−101.451.321.111.331.361.601.301.261.121.041.131.131.261.13
−201.301.242.051.241.171.921.241.181.091.121.101.061.281.09
Table 2. The maximum and median values of the SNR for individual considered cases.
Table 2. The maximum and median values of the SNR for individual considered cases.
Disturbance TypeSNRuThe Maximum ValueThe Median Value
EEWT-TOKLTSSbollWT-TONRH-SVDEEMD-SVDEEWT-TOKLTSSbollWT-TONRH-SVDEEMD-SVD
“white”
noise
2022.435.0021.7219.5613.074.7511.4717.180.2517.3213.739.230.621.51
1019.381.1217.3618.3213.473.1311.1415.29−2.0511.7813.698.84−2.51−1.20
016.490.8315.1210.8911.635.546.2911.63−1.689.788.698.99−0.600.02
−105.140.134.461.503.203.50−0.811.11−1.580.78−1.60−0.25−2.94−3.86
−200.14−0.84−1.71−1.30−1.04−1.04−1.97−3.63−3.19−4.05−7.16−5.96−7.20−8.32
“violet”
noise
2027.636.0625.0723.6513.102.2111.2519.821.2819.4513.808.25−3.44−2.42
1021.082.8419.5620.4412.012.3011.4115.48−1.2014.1711.047.74−2.90−1.64
022.762.0815.8118.9412.166.042.7014.71−1.9510.3511.068.63−1.81−2.18
−1013.820.234.426.516.534.19−1.247.36−3.531.394.002.01−4.75−6.06
−201.910.430.361.160.21−1.37−2.080.14−1.33−1.29−0.97−4.15−6.16−8.34
“brown”
noise
2025.023.1120.6318.7312.573.8610.2018.310.0511.5916.017.85−3.35−2.23
1014.98−0.0212.4110.4413.824.1111.7212.64−3.616.249.929.44−4.11−0.04
05.250.003.403.0310.00−0.836.952.59−2.34−2.161.357.24−8.78−1.08
−102.10−3.281.83−0.2710.00−1.290.29−3.26−5.54−7.01−6.935.69−8.88−7.47
−20−1.04−5.16−1.78−9.58−0.61−10.080.00−6.49−7.55−9.95−11.68−3.00−12.89−7.00
“blue”
noise
2024.936.3923.8424.0910.390.709.5018.331.3718.4914.946.06−4.30−3.88
1018.205.5818.0319.2311.762.9310.7314.132.7013.4211.977.88−1.72−0.19
015.021.9912.1410.408.342.301.659.32−0.657.587.105.68−1.99−1.58
−109.982.085.354.014.063.77−2.924.91−0.151.201.020.02−4.65−6.85
−200.480.01−0.300.02−1.16−2.40−3.25−0.81−1.54−2.00−3.62−5.60−7.19−9.03
“pink”
noise
2025.022.1823.4021.3313.502.4411.2519.21−1.1214.8917.778.41−3.30−2.14
1015.020.8214.0412.9213.345.6510.9912.84−1.328.8311.699.22−0.840.90
04.01−0.903.382.847.00−0.795.712.05−3.02−0.820.764.92−8.15−1.87
−103.75−0.870.550.944.73−1.532.12−1.54−3.68−5.44−3.861.93−8.59−2.90
−202.070.000.03−2.781.42−2.841.47−2.98−4.05−8.31−10.22−2.58−10.24−2.39
traffic
sounds
2025.272.4421.6221.4913.695.2912.8620.620.0217.6618.809.06−1.94−0.68
1015.012.1513.0711.9612.084.9611.5612.74−0.249.7111.178.13−1.84−0.82
011.104.077.997.0110.043.338.238.260.814.204.577.01−2.90−1.13
−102.930.00−0.280.121.39−0.900.090.36−2.40−3.74−3.930.00−4.76−4.71
−201.770.23−0.23−4.52−0.23−0.23−0.23−1.56−3.51−8.88−11.78−3.10−10.42−2.19
hair
dryer
2025.231.0722.4921.6513.592.4612.4818.91−1.6516.8717.808.60−4.34−1.91
1017.241.6514.4712.0812.074.279.1013.42−0.919.2710.627.99−2.31−1.35
011.102.277.694.868.074.714.117.23−0.973.632.796.01−3.01−1.50
−106.41−0.20−0.20−0.321.66−2.55−1.29−0.20−0.39−3.04−5.02−2.16−7.35−5.16
−202.211.370.73−1.560.09−1.590.17−2.701.19−2.80−10.32−4.99−6.56−4.93
fan2024.770.8622.3518.9413.242.2512.5018.38−1.7715.1016.018.16−3.37−2.16
1015.001.0412.7610.5910.782.307.8411.89−1.097.3610.096.96−2.53−1.60
010.962.278.045.238.393.444.947.46−0.563.372.495.60−3.79−1.50
−102.78−0.140.27−0.422.45−1.44−0.24−0.25−3.85−0.98−4.40−1.20−4.29−3.90
−200.530.620.53−4.770.680.530.53−1.43−2.89−3.26−12.89−5.57−8.27−4.27
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kuwałek, P.; Jęśko, W. Speech Enhancement Based on Enhanced Empirical Wavelet Transform and Teager Energy Operator. Electronics 2023, 12, 3167. https://doi.org/10.3390/electronics12143167

AMA Style

Kuwałek P, Jęśko W. Speech Enhancement Based on Enhanced Empirical Wavelet Transform and Teager Energy Operator. Electronics. 2023; 12(14):3167. https://doi.org/10.3390/electronics12143167

Chicago/Turabian Style

Kuwałek, Piotr, and Waldemar Jęśko. 2023. "Speech Enhancement Based on Enhanced Empirical Wavelet Transform and Teager Energy Operator" Electronics 12, no. 14: 3167. https://doi.org/10.3390/electronics12143167

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop