• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Speech Intelligibility Enhancement Algorithm Based on Multi-Resolution Power-Normalized Cepstral Coefficients(MRPNCC)for Digital Hearing Aids

    2021-04-26 07:20:58XiaWangXingDengHongmingShenGuodongZhangandShibingZhang

    Xia Wang,Xing Deng,Hongming Shen,*,Guodong Zhang and Shibing Zhang

    1School of Information Science and Technology,Nantong University,Nantong,226019,China

    2School of Automation,Key Laboratory of Measurement and Control for CSE,Ministry of Education,Southeast University,Nanjing,210096,China

    3School of Computer Science,Jiangsu University of Science and Technology,Zhenjiang,212003,China

    ABSTRACT Speech intelligibility enhancement in noisy environments is still one of the major challenges for hearing impaired in everyday life.Recently,Machine-learning based approaches to speech enhancement have shown great promise for improving speech intelligibility.Two key issues of these approaches are acoustic features extracted from noisy signals and classifiers used for supervised learning.In this paper,features are focused.Multi-resolution powernormalized cepstral coefficients(MRPNCC)are proposed as a new feature to enhance the speech intelligibility for hearing impaired.The new feature is constructed by combining four cepstrum at different time-frequency(T-F)resolutions in order to capture both the local and contextual information.MRPNCC vectors and binary masking labels calculated by signals passed through gammatone filterbank are used to train support vector machine(SVM)classifier,which aim to identify the binary masking values of the T-F units in the enhancement stage.The enhanced speech is synthesized by using the estimated masking values and wiener filtered T-F unit.Objective experimental results demonstrate that the proposed feature is superior to other comparing features in terms of HIT-FA,STOI,HASPI and PESQ,and that the proposed algorithm not only improves speech intelligibility but also improves speech quality slightly.Subjective tests validate the effectiveness of the proposed algorithm for hearing impaired.

    KEYWORDS Speech intelligibility enhancement;multi-resolution power-normalized cepstral coefficients;binary masking value;hearing impaired

    1 Introduction

    Noise affects speech intelligibility,especially that of hearing-impaired patients.To achieve the same speech intelligibility in a noisy environment,the signal-to-noise ratio (SNR) required for hearing-impaired patients is generally 3-6 dB higher than that of individuals with normal hearing [1].Traditional noise-suppression algorithms in modern hearing aids can improve the quality of noisy speech,but they have limited capacity to improve intelligibility [2].Speech intelligibility is closely associated with the information contained in the speech.Therefore,improving the intelligibility of speech requires that the target speech signal should not be distorted while noise is removed.However,the speech signals estimated from noisy speech are often inaccurate.Recently,research on speech intelligibility enhancement based on binary masking is emerging [3].The noisy signal is first decomposed into two-dimensional time-frequency (T-F) units,of which one dimension is the T-F units of the target signal and the another dimension is the T-F units of the interference signal.The ideal binary masking (IBM) algorithm indicates that if the energy of the target signal is higher than that of the interference signal,then the element value of the binary matrix corresponding to the T-F unit is 1,otherwise,it is 0.Under certain conditions,the algorithm improves speech intelligibility by removing masking signals that do not contribute to intelligibility [4-7].In general,there are two kinds of research on speech intelligibility enhancement by binary masking.One is based on ideal binary masking,the other is based on estimated binary masking using machine learning.Early research mainly focused on ideal binary masking such as reference [8].The disadvantage of an ideal binary masking algorithm is that the IBM values are known prior,which is hardly satisfied in practice.Motivated by the success of the IBM algorithm for improvement of speech intelligibility,binary masking algorithms based on machine learning is proposed [9,10].Two key issues of these algorithms are acoustic features extracted from noisy signals and classifiers used for supervised learning.While previous studies have emphasized classifiers,in this paper,we focus on features.Our goal is to develop superior features to classify T-F unit accurately and to improve speech intelligibility.

    In this paper,we proposed a new feature called multi-resolution power-normalized cepstral coefficients (MRPNCC) to improve the performance of the speech intelligibility enhancement algorithm for digital hearing aids.This paper is organized as follows.A speech intelligibility enhancement system based on MRPNCC is constructed in Section 2.Section 3 describes the extraction steps of MRPNCC.Systematic evaluation and comparison are presented in Section 4.We conclude this paper in Section 5.

    2 Speech Intelligibility Enhancement System Based on MRPNCC

    2.1 System Overview

    A block diagram of the speech intelligibility enhancement system based on binary masking is shown in Fig.1 consisting a training stage and an enhancement stage.In the training stage,noise and clean speech signal are used to train the Support Vector Machine (SVM) model to classify the speech dominant unit and the noise dominant unit.At first,noisy speech is synthesized by clean speech and noise according to the specified SNR.Then the clean speech,noise and noisy speech are passed through gammatone filterbank to generate sub-band signals,from which features of noisy sub-band signals are extracted and priori SNRs are calculated.The label of the T-F unit is determined by the calculated SNR and predetermined local criteria value.Finally,the extracted feature vector and the label of T-F unit are used to train SVM classifiers.In the enhancement stage,noisy speech is enhanced by the classified label and Wiener filter.Firstly,noisy signal is decomposed into sub-band signals using gammatone filterbank.Then,feature vector of each subband signal is extracted and classified by the SVM model.At last,the estimated masking value is multiplied by the speech spectrum enhanced by the Wiener filter to generate the enhanced speech spectrum which ultimately gives rise to enhanced speech through inverse Fourier transform.

    2.2 Binary Masking Value Estimation Based on SVM

    SVM is a machine learning algorithm that is based on statistical learning theory.Statistical learning theory adopts the rule of structural risk minimization criterion to improve the generalization ability of the model.In addition,Gaussian mixture model (GMM) is one of the commonly used classification models.The output of GMM reflects the similarity between samples of the same class.However,the output of SVM reflects the difference between the heterogeneous samples.It is suitable for classification with small number of samples and nonlinear data.It has the advantages of short training time and good generalization performance.Furthermore,the researchers in [11] found that under the same experimental conditions,the classification performance of SVM is better than that of the GMM model.This is probably because the GMM has more parameters while the training data is inadequate.It was also found in [11] that the classification performance of the SVM is superior to that of the multilayer perceptron (MLP).Therefore,in this paper SVM is adopted to estimate the label of the T-F units.

    Figure 1:Diagram of the intelligibility enhancement system

    During training stage,MRPNCC feature vectors of noisy speech are firstly calculated.Then,the ideal binary masking value of each T-F unit is obtained using the corresponding clean speech and noise to train SVM model.During test stage,the extracted feature vector is used to determine whether the T-F unit is speech-dominant or noise-dominant.The SVM maps the feature vector to a high-dimensional space by applying a kernel function.In the high-dimensional space,a hyperplane is derived to maximize the margin of class separation.Here,the radial basis function(RBF) kernel is chosen:

    The reason that we choose RBF as our kernel function is as follows:The first reason is that this kernel nonlinearly maps samples into a higher dimensional space,so it can handle the case when the relation between class labels and attributes is nonlinear.The second reason is the number of hyperparameters which influences the complexity of model selection.The polynomial kernel has more hyperparameters than the RBF kernel.Finally,the RBF kernel has fewer numerical difficulties.Only when the number of features is very large,the RBF kernel is not suitable.In this paper,the feature dimension is 160,so RBF is the best choice.

    Given a set of pairs (xi,yi),wherexiis a feature vector andyiis the corresponding binary label,the SVM requires to solve the following optimization problem:

    wherewis the weight of the hyperplane,ξis a non-negative variable measuring the deviation of a data point from the hyperplane,Ccontrols the trade-off between complexity of the SVM and the number of nonseparable points andΦrepresents the vector of a set of nonlinear functions that transform the input space to the higher dimensional feature space.bis the bias.After the SVM is well trained,the T-F units can be classified by the model with the following function:

    where SV represents the set of support vector indices in training data andαiis the Lagrange multiplier that can be determined in the training phase.

    2.3 Speech Spectrum Estimation Based on Binary Masking

    The binary masking values of T-F units are classified using the SVM model.The identification result oflabel=0 represents that the T-F unit is dominated by noise,whilelabel=1 represents that the T-F unit is dominated by speech.Therefore,after the binary masking operation,the speech spectrum of thelth channel for themth frame is as follows:

    whereY(m,l)is the frequency spectrum of noisy speech andG(m,l)is the gain of the Wiener filter.The gain is defined below:

    whereSNRprio(m,l)is thea prioriSNR.We calculateSNRprio(m,l)as follows:

    3 MRPNCC Feature Extraction

    Power normalized cepstral coefficients (PNCC) are features based on human auditory processing.PNCC simulate the masking effect of human hearing during the process of feature extraction and have been used in robust speech recognition systems.The researchers in [13] have also achieved a recognition accuracy that is superior to that of the traditional Mel-Frequency Cepstral Coefficients (MFCC) feature.Motivated by the PNCC feature extraction algorithm,in this paper we propose a new feature called multi-resolution PNCC (MRPNCC).The new feature is constructed by combining four PNCCs at different spectrotemporal resolutions in order to capture both the local and contextual information of the signal.The MRPNCC is obtained after completing the asymmetric denoising,temporal masking,average power normalizing,power function nonlinearity,and discrete cosine transforming processes.A flow chart for one-resolution parameter extraction is shown in Fig.2.The other three-resolution parameter extraction methods are similar to it.

    Figure 2:One-resolution PNCC feature extraction flow chart

    3.1 Pre-Processing

    The pre-processing of the signal includes pre-emphasizing,framing,and filtering with the gammatone filterbank,in which the pre-emphasis filter isH(z)=1?0.97z?1.During the framing phase,the pre-emphasized signals are processed at two frame lengths (L1andL2) with the following parameters in this paper:L1=20 ms andL2=200 ms.Signals with the two frame lengths are then filtered using a 40-channel gammatone filterbank,generating their respective subband signals.Then,the multi-resolution power of the sub-band signals is calculated.Let themth frame of signals with the lengths ofL1andL2bex1,m(n)andx2,m(n).The spectra oflth sub-band signals ofx1,m(n)andx2,m(n)are represented asX1(m,l)andX2(m,l),respectively.Then the powers of thelth sub-band signals are denoted as follows:P1(m,l)andP2(m,l)are then smoothed with square windows with the widths ofN1andN2in both the time domain and frequency domain to obtain the context information.After the smoothing,the following formula can be obtained:

    wherel1=max(l?N1+1,1),m1=max(m?N1+1,1),l2=max(l?N2+1,1),andm2=max(m?N2+1,1).Subsequent feature extractions are performed on the basis ofP1(m,l),P2(m,l),P3(m,l)andP4(m,l),which are calculated in parallel.The above four kinds of power spectra are uniformly designated asP(m,l)to make them easy to express.

    3.2 Asymmetric Noise Suppression

    In general,the speech power in the channel changes faster than the background noise power in the same channel,i.e.,speech has a higher modulation frequency than that of noise.Based on this observation,some algorithms have been developed to perform high-pass or band-pass filtering in the modulation spectrum domain to remove the slowly-varying components which typically represent the additive noise [14].In this paper the asymmetric denoising method is used to eliminate the interference of the noise during the feature extraction process.The block diagram of asymmetric noise suppression processing with temporal masking is shown in Fig.3.The relationship between the input and the output of the asymmetric filter is as follows [15]:

    In Formula (12),mis the frame index,lis the channel index,andλaandλbare constants between 0 and 1.In the formula above,the parameters areλa=0.999 andλb=0.5.Q(0,l)is initialized to 0.9·P(m,l).After noise suppression,half-wave rectification is performed on the signal and the output of the rectifier isQ0(m,l).Since the rectified signal still contains noise,a second asymmetric filtering is performed to obtain a lower envelope which is then used as the basic noise level.The envelopeQf(m,l)is obtained as follows:

    Qf(m,l)in Formula (13) is initialized asQ(m,l).As shown in Fig.3,we use the lower envelope of the rectified signalQf(m,l)as floor level forQ1(m,l)after temporal masking:

    whereQtm(m,l)is obtained by performing temporal masking on signalQ0(m,l).Then,Qtm(m,l)and the second asymmetric filtering result are compared,and the higher value is retained.

    Figure 3:Asymmetric noise suppression block diagram

    Whether the sub-band signal is a speech signal or a silent segment is determined based onP(m,l)and the lower envelopeQ(m,l).IfP(m,l)≥cQ(m,l),then the sub-band signal is a speech segment with an asymmetric filtering module output ofQ1(m,l).Otherwise,the sub-band signal is a silent segment with an asymmetric filtering module output ofQf(m,l).These signals can be expressed mathematically as follows:

    In Formula (15),c=2 provides the best performance according to the pre-experiments.

    3.3 Temporal Masking

    Studies have shown that the human auditory system pays more attention to the onset of an incoming power envelope rather than the falling edge of the same power envelope [16,17],which has led to various speech onset enhancement systems [18,19].In this paper,this auditory effect of the human ear has been integrated into the temporal masking algorithm.First,the peak power of each channelQp(m,l)is calculated as follows:

    In Formula (16),λtis the forgetting factor for obtaining the online peak andλt=0.85.The output after temporal masking is as follows:

    In Formula (17),μt=0.2.Therefore,the transfer function of the entire asymmetric denoising process can be expressed as:

    The power spectrum after modulating with the transfer function ofS(m,l)is as follows:

    wherePo(m,l)is the corresponding original spectrum at a different resolution without any smoothing.

    3.4 Mean Power Normalization

    The human ear has an automatic gain adjustment function to mitigate the influence of changes in amplitude of the input signal.To simulate this function of the human ear,the power spectrum is normalized during the MRPNCC parameter extraction process.In this paper,the mean power is accomplished by running average [13]:

    In Formula (20),λμis set to 0.999.The normalized power is obtained directly from the estimated average power:

    3.5 Power Function Non-Linearization

    The nonlinear function used to describe the relationship between input and output of the channel plays an important role in the auditory model.The power law function has been shown to be a good approximation of the transfer function observed in physiological experiments.Athulya et al.[20] found that the use of a power law function with an exponent of 1/15 best suited for noisy conditions.Therefore,the power law factor is set to 1/15 in this paper:

    4 Experiments

    4.1 Simulation Experiment

    MHINT is acorpusdesigned for Mandarin hearing test [21].Sentences in thecorpusrepresent simple and conversational speech among which some words can be predicted with the contextual information.However,subjects with different hearing loss have different cognitive ability which will affect the experimental results.Therefore,experiments were conducted on thecorpusrecorded by ourselves in an anechoic chamber,containing 400 sentences.To reduce the effect of cognition in the hearing test,thecorpuswas designed with relatively low word context.A total of 400 sentences were constructed by a native female speaker in China.The sampling rate of all the utterances was 44.1 kHz and the signals were resampled to 16 kHz.A total of 300 sentences were used to train the SVM model and the remaining 100 sentences were used to test.All these sentences were corrupted by three types of noise at?5,0 and 5 dB SNR.Three types of noise signals were babble noise,factory noise,and speech shaped noise (SSN).

    A baseline system was constructed to determine the parameters for the proposed feature extraction algorithm such as the frame length and the width of the smoothing window.For training and test,clean speech was mixed with babble noise at 0 dB.At first,the training speech was sent to a 40-channel gammatone filterbank.Then,the feature parameters were extracted from the filtered signals to train the SVM model.Lastly,the SVM model generated the classification result based on the feature vectors to determine whether the T-F unit was speech-dominant or noise-dominant.In the model,to determine the category of the T-F unit,the threshold parameter of local SNR criterion was set to?5 dB.Since the task of the system was to classify the T-F unit,the most direct evaluation index should be the classification accuracy.However,it is not reasonable to use the classification accuracy as the evaluation index of the system performance,because it assigns the same importance to the rates of two errors (i.e.,misclassifying the speechdominant T-F unit as the noise-dominated T-F unit and misclassifying the noise-dominated T-F unit as the speech-dominant T-F unit).In fact,studies on speech intelligibility show that the effect of misclassifying the noise-dominated T-F unit as the speech-dominant T-F unit on speech intelligibility is greater than that of misclassifying the speech-dominant T-F unit as the noise-dominated T-F unit [22].So,the estimation of HIT-FA is considered as a best objective evaluation criteria because of its strong correlation to human speech intelligibility [23].HIT refers to the hit rate,which is the percentage that the system correctly classifies the speech-dominated T-F unit.FA refers to the false alarm rate,which is the percentage that the system wrongly classifies the noise-dominated T-F unit.Therefore,in this paper,HIT-FA was used as an objective index for the system performance evaluation.

    The methods for determining the frame length and the smoothing window width in the MRPNCC feature extraction algorithm are as follows:The features extracted fromP1(m,l),P2(m,l),P3(m,l),P4(m,l)were expressed as PNCC1,PNCC2,PNCC3,and PNCC4,respectively.All of these features were cascaded in series to form the feature MRPNCC.Notably,the method used to extract PNCC1-PNCC4 was not exactly the same as that used in Reference [13],so the resulting feature vectors were not identical either.The same symbols are used here because of the convenience of description.First,during calculating PNCC1,the window size was set to 20 ms,the noisy speech with babble noise at 0 dB SNR was used to train the SVM model,and then noisy speech with babble noise at 0 dB SNR was used to identify the T-F units of each channel and the HIT-FA value of the system was calculated.The frame length of PNCC1 was fixed,while that of PNCC2 was varied to compute the HIT-FA.The frame length corresponding to the maximum HIT-FA value was deemed the optimal frame length which was 200 ms.Similarly,the widths of the square sliding windows for PNCC3 and PNCC4 (N1andN2) were determined,and the optimal widths were determined atN1=2 andN2=4,respectively.Fig.4 shows the HIT-FA value of the system after adding contextual information.As can be seen from the Fig.4,adding PNCC2-4 improves the system performance for babble noisy speech at 0 dB SNR.Similar trends are observed for the other noises.

    Figure 4:HIT-FA values of the intelligibility enhancement system after adding different contextual information

    A visualization of the MRPNCC feature is given in Fig.5,where the left plots features extracted from the clean speech and the right from a babble mixture at 0 dB SNR.As shown in Fig.5,PNCC displayed in the first row is the regular cepstrum coefficients,PNCC2 displayed in the second row captures temporal context,PNCC3 displayed in the third row encodes relatively small spectrotemporal context and PNCC4 displayed in the fourth row encodes relatively large spectrotemporal context.By comparing Figs.5a and 5b,we find that MRPNCC features can partially retain spectrotemporal information of speech in noise environment,and that it acquires the contextual information in the signal and that it is thus suitable for the classification of T-F units.This rationale behind MRPNCC is that regardless of the speech-dominant T-F unit or the noise-dominant T-F unit,the T-F units are usually clustered.

    The estimated binary masking values of the noisy speech adding babble noise at 0 dB SNR are shown in Fig.6.Fig.6a shows the ideal binary masking value.It indicates that the speechdominant T-F units or noise-dominant T-F units are presented essentially in clusters,and only occasionally a few of the T-F units wander off the T-F cluster.Fig.6b displays the estimation result of the binary masking value using the SVM model.Using the MRPNCC feature,the SVM identifies most of the speech-dominant T-F units,while a small number of noise-dominant units are misidentified as speech-dominant ones and a small number of speech-dominant units are misidentified as noise-dominant ones.

    Figure 5:MRPNCC feature spectrum (a) clean speech and (b) noisy speech with babble noise at 0 dB SNR

    Figure 6:Binary mask estimation (a) ideal binary masking value and (b) binary masking estimated via SVM

    To compare the waveforms of a speech segment before and after enhancement by the proposed algorithm,the speech waveforms before and after enhancement are shown in Fig.7.Fig.7a shows the waveform of the clean speech signal.Fig.7b shows that of the noisy signal with an SNR of 0 dB synthesized by mixing clean speech with babble noise.Fig.7c shows that of the speech processed by the ideal binary masking enhancement algorithm.Fig.7d shows that of the speech processed by the proposed algorithm.It is indicated that the waveform of the speech processed by the ideal binary masking algorithm is closest to that of the clean speech,and the waveform of the speech processed by the proposed algorithm is close to that of the clean speech.

    Figure 7:Comparison of speech waveforms before and after enhancement (a) clean speech(b) noisy speech (c) enhanced speech by IBM and (d) enhanced speech by proposed method

    Experimental results are presented to demonstrate the superiority of MRPNCC feature over competing ones.The features used for comparison were PLP [24],Relative spectral transform PLP (R-PLP) [24],MFCC [25],MRMFCC which is extend from MFCC,Amplitude Modulation Spectrogram (AMS) [26],and Power Normalized Cepstral Coefficients (PNCC) [13].The HIT-FA values of different features under babble noisy environment at 0 dB SNR are shown in Fig.8.Comparisons of different features indicate that the HIT-FA value of the MRPNCC feature is the highest and that of the AMS is the lowest,while those of the other features are in the middle.The reason why the MRPNCC feature has the best performance is that it utilizes the multi-resolution information of the signal.

    Figure 8:HIT-FA values with different features

    To compare the performance of the MRMFCC and the MRPNCC in 3 noise types,the test results of STOI,HASPI and PESQ are shown in Tab.1.We can see that all of the methods improve the speech intelligibility and speech quality.Results based on MRPNCC is better than that of MRMFCC due to temporal masking during feature extraction.Comparing the results of 3 noise conditions,results of the babble noise are the worst.This is probably because of the similarity between the babble noise and the speech.At the same time,we find that although PESQ is improved,it is improved slightly.

    Table 1:STOI,HASPI and PESQ of the proposed algorithm and contrast algorithm

    4.2 Auditory Experiment

    Ten patients with moderate to moderate-to-severe sensorineural hearing loss participated in the listening test.They were five males and five females aged from 52 to 66 years with an average age of 60.3 years,Pure-tone average (PTA) based on 0.5,1,and 2 kHz ranged from 35 to 65 dB hearing level (HL:average 54.6),and a sloping hearing loss pattern.The pure tone-hearing threshold for these patients is shown in Tab.2.The test utterance was taken from thecorpuswith a total of 100 sentences.The noise types were babble noise,factory noise and SSN with SNRs of?5,0 and 5 dB,respectively.Because the subjects all had hearing loss,the processed signal had to be amplified.Before the amplification,the root mean square level of the signal was corrected to 65 dB,and the signals were amplified using the NAL-NL1 formula and transmitted to the subject through an earphone.The test was conducted in an anechoic chamber.After hearing one sentence,the subject was asked to speak the sentence out immediately,and the tester recorded the content that the subject had just spoken.After the experiment was completed,the percentage of correctly identified words by the subject was used to represent the intelligibility of the speech.During testing,the noise type was randomized and the SNR was blocked.Firstly,the unprocessed speech with babble noise was tested.Secondly,the processed speech at?5 dB was tested.Thirdly,the processed speech at 0 dB was tested.At last,the processed speech at 5 dB was tested.

    Table 2:Subject’s pure tone hearing threshold/dB

    Fig.9 shows the speech intelligibility results of the ten hearing impaired patients under different SNRs and different environments.Because the intelligibility of the processed speech in the?5 dB environment was lower than 50%,it was not shown in the figure.The intelligibility of the unprocessed speech in a noisy environment is indicated by “?”.That of the processed speech is indicated by “▲”,and the improved speech intelligibility is represented by the height of the line.In each figure,the intelligibility of speech at 0 dB SNR is shown in the left and those of speech at 5 dB SNR is shown in the right.Pairedttests confirmed the reliability of the processing benefit in each condition shown in Fig.9 (p<0.001).

    Figure 9:Speech intelligibility under different noise and SNR conditions (a) babble noise(b) factory noise and (c) SSN

    In the babble noise environment,the speech intelligibility for all patients at the SNR of 0 dB was improved.Subject 2 received the most benefit of 49%.Four out of the ten subjects received a benefit of more than 40%.In the babble noise environment,the speech intelligibility for all patients at the SNR of 5 dB was improved.Subject 2 received the most benefit of 36%.Five out of the ten subjects received a benefit of more than 20%.

    In the factory noise environment,the speech intelligibility for all patients at the SNR of 0 dB was improved.Subject 9 received the most benefit of 46%.Five out of the ten subjects received a benefit of more than 35%.In the factory noise environment,all of the speech intelligibility at the SNR of 5 dB was improved.Subject 2 received the most benefit of 37%.Five out of the ten subjects received a benefit of more than 20%.

    In the SSN environment,the speech intelligibility of all the patients at the SNR of 0 dB was improved.Subject 5 received the most benefit of 44%.Four out of the ten subjects received a benefit of more than 40%.In the SSN environment,except for Subject 1,the speech intelligibility at the SNR of 5 dB for other patients was improved.Subject 2 received the most benefit of 37%.Five out of the ten subjects received a benefit of more than 20%.

    A comprehensive comparison of the three panels in Fig.9 indicates that if the intelligibility of the unprocessed speech for the subject is lower,then the intelligibility of the processed speech for the subject improves more.If the intelligibility of the unprocessed speech for the subject is higher,then that of the processed speech for the subject improves less.Because of the high baseline level,it is difficult to make further improvement.In some cases,the intelligibility even decreases.The group-mean improvement of intelligibility in all environments is 25.57%.It indicates that the proposed algorithm is suitable for improving the speech intelligibility.

    5 Conclusions

    In this paper,we proposed the MRPNCC feature for speech intelligibility enhancement system.The signals were firstly decomposed to T-F units using the gammatone filterbank.Then,the novel robust MRPNCC features were extracted from the obtained T-F units,and utilized to train the SVM model together with the known labels.During the recognition stage,the SVM model was used to identify whether the T-F unit was noise-dominated or speech-dominated.If the T-F unit was noise-dominated,then it was removed.Otherwise,it was processed by Wiener filter.Lastly,the processed T-F unit was used to synthesize the enhanced speech.To evaluate the performance of the speech intelligibility enhancement system,experiments under different feature conditions were compared,and results show that the MRPNCC feature outperformed the traditional PLP,R-PLP,MFCC,MRMFCC,AMS,and PNCC features.Finally,subjective auditory experiments results of ten hearing-impaired subjects demonstrated that the intelligibility of the enhanced speech was significantly improved.

    Funding Statement:This work was supported by the National Natural Science Foundation of China (Nos.61902158,61673108),the Science and Technology Program of Nantong (JC2018129,MS12018082) and Top-notch Academic Programs Project of Jiangsu Higher Education Institutions (PPZY2015B135).

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    久久精品国产综合久久久| 亚洲国产看品久久| 一边摸一边做爽爽视频免费| 亚洲色图综合在线观看| 国产国语露脸激情在线看| 久久人妻福利社区极品人妻图片| 亚洲国产精品sss在线观看 | 久久久久久久久久久久大奶| 岛国毛片在线播放| 精品一区二区三卡| 天堂俺去俺来也www色官网| 九色亚洲精品在线播放| 精品久久久精品久久久| 久久青草综合色| www.熟女人妻精品国产| 91麻豆av在线| 制服诱惑二区| 99国产精品一区二区蜜桃av | 国产精品永久免费网站| 高清av免费在线| 18禁黄网站禁片午夜丰满| 久9热在线精品视频| 成人手机av| 亚洲国产欧美一区二区综合| 国产精品久久视频播放| 国产有黄有色有爽视频| 午夜福利在线免费观看网站| 国产精品自产拍在线观看55亚洲 | 多毛熟女@视频| 可以免费在线观看a视频的电影网站| 大片电影免费在线观看免费| 欧美+亚洲+日韩+国产| 黄网站色视频无遮挡免费观看| 一级黄色大片毛片| 曰老女人黄片| 欧美乱码精品一区二区三区| 黄色a级毛片大全视频| 亚洲精品国产一区二区精华液| 精品午夜福利视频在线观看一区| 亚洲免费av在线视频| 很黄的视频免费| 99热网站在线观看| e午夜精品久久久久久久| 很黄的视频免费| 三级毛片av免费| 美女 人体艺术 gogo| 国产极品粉嫩免费观看在线| 波多野结衣一区麻豆| 欧美人与性动交α欧美精品济南到| 欧美人与性动交α欧美精品济南到| 免费观看a级毛片全部| 在线观看午夜福利视频| 亚洲一区二区三区欧美精品| 久久精品成人免费网站| 久久性视频一级片| 波多野结衣一区麻豆| 自拍欧美九色日韩亚洲蝌蚪91| 日韩欧美在线二视频 | 久久狼人影院| 亚洲第一av免费看| 久久精品熟女亚洲av麻豆精品| 80岁老熟妇乱子伦牲交| 精品人妻在线不人妻| 久久国产精品男人的天堂亚洲| 国产又爽黄色视频| av欧美777| 亚洲第一青青草原| 成年人免费黄色播放视频| 老司机影院毛片| 日韩人妻精品一区2区三区| 中文字幕人妻丝袜一区二区| 黄色视频不卡| 另类亚洲欧美激情| 国产欧美日韩精品亚洲av| 久久久水蜜桃国产精品网| 欧美日韩中文字幕国产精品一区二区三区 | 免费在线观看视频国产中文字幕亚洲| 黑人操中国人逼视频| 亚洲一区二区三区不卡视频| 亚洲国产精品合色在线| 麻豆成人av在线观看| 不卡av一区二区三区| 岛国在线观看网站| 亚洲精品成人av观看孕妇| 97人妻天天添夜夜摸| 一区二区三区激情视频| 久久婷婷成人综合色麻豆| 一区二区三区激情视频| 在线观看www视频免费| 亚洲熟女毛片儿| 每晚都被弄得嗷嗷叫到高潮| 91字幕亚洲| 午夜福利乱码中文字幕| 天堂√8在线中文| 90打野战视频偷拍视频| 99国产综合亚洲精品| 日韩欧美免费精品| 免费女性裸体啪啪无遮挡网站| 交换朋友夫妻互换小说| 久久国产精品男人的天堂亚洲| 在线视频色国产色| 亚洲欧美日韩高清在线视频| bbb黄色大片| 他把我摸到了高潮在线观看| 999精品在线视频| 身体一侧抽搐| 精品亚洲成a人片在线观看| 亚洲精品久久成人aⅴ小说| 国产精品免费一区二区三区在线 | 两性夫妻黄色片| 欧美日韩福利视频一区二区| 欧美一级毛片孕妇| 欧美国产精品一级二级三级| 淫妇啪啪啪对白视频| 久久国产亚洲av麻豆专区| 久久精品亚洲精品国产色婷小说| 手机成人av网站| 少妇 在线观看| 热re99久久精品国产66热6| 少妇粗大呻吟视频| 精品久久久久久,| 国产野战对白在线观看| 午夜影院日韩av| 成人18禁在线播放| 久久这里只有精品19| bbb黄色大片| 多毛熟女@视频| 窝窝影院91人妻| 一边摸一边抽搐一进一出视频| 妹子高潮喷水视频| 亚洲少妇的诱惑av| 日韩欧美一区二区三区在线观看 | 欧美成狂野欧美在线观看| 亚洲人成电影免费在线| 欧美国产精品一级二级三级| 欧美乱码精品一区二区三区| 亚洲专区国产一区二区| 欧美日韩亚洲高清精品| www.自偷自拍.com| 欧美在线黄色| 午夜福利一区二区在线看| 亚洲久久久国产精品| 国产精品二区激情视频| 欧美日韩中文字幕国产精品一区二区三区 | 亚洲成av片中文字幕在线观看| 黄色视频不卡| 一夜夜www| 久久精品成人免费网站| 精品久久久久久电影网| 国产激情久久老熟女| xxxhd国产人妻xxx| 亚洲一区二区三区不卡视频| 亚洲欧美激情在线| av线在线观看网站| 亚洲国产看品久久| 99国产精品免费福利视频| 国产成人精品在线电影| 精品免费久久久久久久清纯 | 久久精品亚洲熟妇少妇任你| 国产精品1区2区在线观看. | 欧美激情 高清一区二区三区| 亚洲熟妇熟女久久| 久久香蕉国产精品| av有码第一页| 91av网站免费观看| 午夜福利在线观看吧| 精品无人区乱码1区二区| 欧美精品av麻豆av| 午夜福利影视在线免费观看| 一本综合久久免费| 成人特级黄色片久久久久久久| 中文字幕制服av| 国产xxxxx性猛交| 两个人免费观看高清视频| 在线天堂中文资源库| 91av网站免费观看| 久久九九热精品免费| e午夜精品久久久久久久| 正在播放国产对白刺激| 91精品国产国语对白视频| 亚洲国产精品合色在线| 一边摸一边做爽爽视频免费| 大型黄色视频在线免费观看| 欧美av亚洲av综合av国产av| 另类亚洲欧美激情| 一个人免费在线观看的高清视频| 免费高清在线观看日韩| 极品教师在线免费播放| 欧美亚洲 丝袜 人妻 在线| 老司机午夜福利在线观看视频| 欧美最黄视频在线播放免费 | 久久久久国内视频| 国产亚洲欧美98| 久久热在线av| 两个人看的免费小视频| 韩国av一区二区三区四区| 国产亚洲av高清不卡| 韩国精品一区二区三区| 深夜精品福利| 狠狠狠狠99中文字幕| 国产国语露脸激情在线看| svipshipincom国产片| 精品久久久精品久久久| 欧美日韩亚洲高清精品| 国产在线观看jvid| 亚洲av日韩精品久久久久久密| 午夜福利在线观看吧| x7x7x7水蜜桃| 淫妇啪啪啪对白视频| 老熟妇乱子伦视频在线观看| 老司机深夜福利视频在线观看| 97人妻天天添夜夜摸| 欧美日韩精品网址| 飞空精品影院首页| 欧美午夜高清在线| 精品久久久久久久毛片微露脸| 国产成人影院久久av| 国产区一区二久久| 久久久久精品国产欧美久久久| av不卡在线播放| 身体一侧抽搐| 一a级毛片在线观看| 国产精品av久久久久免费| 男人操女人黄网站| 日韩欧美在线二视频 | 男女床上黄色一级片免费看| 国产成人精品无人区| 老司机午夜十八禁免费视频| 一进一出好大好爽视频| 黑丝袜美女国产一区| 69精品国产乱码久久久| 欧洲精品卡2卡3卡4卡5卡区| 亚洲午夜理论影院| 999精品在线视频| 午夜福利乱码中文字幕| 香蕉丝袜av| 久久香蕉激情| 高清在线国产一区| 无限看片的www在线观看| 欧美+亚洲+日韩+国产| 另类亚洲欧美激情| 狠狠婷婷综合久久久久久88av| 国产精品美女特级片免费视频播放器 | 人成视频在线观看免费观看| 99国产精品一区二区蜜桃av | 亚洲欧美精品综合一区二区三区| 亚洲 国产 在线| 久久天堂一区二区三区四区| 亚洲性夜色夜夜综合| 桃红色精品国产亚洲av| 狂野欧美激情性xxxx| 91在线观看av| 欧美 日韩 精品 国产| 日韩欧美一区二区三区在线观看 | cao死你这个sao货| 亚洲第一欧美日韩一区二区三区| 午夜日韩欧美国产| 丝袜在线中文字幕| 国产精品免费一区二区三区在线 | 精品欧美一区二区三区在线| 一区二区三区国产精品乱码| 欧美成人午夜精品| 日韩欧美一区二区三区在线观看 | 欧美日韩一级在线毛片| 国产精品免费大片| 韩国av一区二区三区四区| 一进一出抽搐gif免费好疼 | 亚洲第一av免费看| 亚洲片人在线观看| 免费人成视频x8x8入口观看| 99久久精品国产亚洲精品| 一区二区三区国产精品乱码| 老汉色∧v一级毛片| 久热这里只有精品99| 成人亚洲精品一区在线观看| 在线观看免费午夜福利视频| 99国产精品99久久久久| 村上凉子中文字幕在线| 视频区欧美日本亚洲| 亚洲aⅴ乱码一区二区在线播放 | 国产淫语在线视频| 高清黄色对白视频在线免费看| 国产亚洲欧美98| 丝袜美足系列| 国产亚洲精品久久久久久毛片 | 成年人黄色毛片网站| av网站在线播放免费| netflix在线观看网站| 久久久国产欧美日韩av| 久久中文字幕一级| 可以免费在线观看a视频的电影网站| 真人做人爱边吃奶动态| 91字幕亚洲| 丝袜美腿诱惑在线| 国产亚洲欧美98| 色综合婷婷激情| 无遮挡黄片免费观看| 国产一区二区激情短视频| 精品一区二区三区四区五区乱码| 国产99久久九九免费精品| 免费在线观看影片大全网站| 久久久水蜜桃国产精品网| 99久久国产精品久久久| 99国产精品99久久久久| 自线自在国产av| 99精国产麻豆久久婷婷| aaaaa片日本免费| 无限看片的www在线观看| 波多野结衣av一区二区av| 久久99一区二区三区| 久久草成人影院| 久久精品亚洲熟妇少妇任你| 亚洲第一av免费看| 男女下面插进去视频免费观看| 可以免费在线观看a视频的电影网站| 免费日韩欧美在线观看| 啦啦啦免费观看视频1| 男女免费视频国产| 亚洲欧美激情综合另类| 天天躁日日躁夜夜躁夜夜| 精品久久久久久,| 男女下面插进去视频免费观看| 欧洲精品卡2卡3卡4卡5卡区| 国产乱人伦免费视频| 久热这里只有精品99| 亚洲精品自拍成人| 91老司机精品| 亚洲精品国产区一区二| 欧美午夜高清在线| 两性夫妻黄色片| 窝窝影院91人妻| 18禁观看日本| 一二三四社区在线视频社区8| 一区在线观看完整版| 精品高清国产在线一区| 韩国精品一区二区三区| 欧美激情极品国产一区二区三区| 久久久国产成人免费| 一级毛片高清免费大全| 伊人久久大香线蕉亚洲五| 黑丝袜美女国产一区| 成人三级做爰电影| 色婷婷久久久亚洲欧美| 色老头精品视频在线观看| 国产在线观看jvid| 久久天躁狠狠躁夜夜2o2o| 亚洲精品国产色婷婷电影| 999精品在线视频| 午夜福利在线免费观看网站| 91麻豆av在线| 亚洲熟女毛片儿| 色94色欧美一区二区| 91麻豆av在线| 国产乱人伦免费视频| 他把我摸到了高潮在线观看| 男人的好看免费观看在线视频 | 黄色视频,在线免费观看| aaaaa片日本免费| 曰老女人黄片| www.999成人在线观看| 亚洲欧美激情综合另类| 超碰成人久久| 少妇被粗大的猛进出69影院| 别揉我奶头~嗯~啊~动态视频| 天天躁日日躁夜夜躁夜夜| 两个人免费观看高清视频| 别揉我奶头~嗯~啊~动态视频| 欧美日韩国产mv在线观看视频| 亚洲第一av免费看| 亚洲一区二区三区欧美精品| xxxhd国产人妻xxx| 日韩制服丝袜自拍偷拍| 亚洲精品中文字幕一二三四区| 久久ye,这里只有精品| 欧美亚洲日本最大视频资源| 久久久水蜜桃国产精品网| 美女 人体艺术 gogo| 日韩免费高清中文字幕av| 国产亚洲精品第一综合不卡| 日韩欧美免费精品| x7x7x7水蜜桃| 看免费av毛片| 午夜视频精品福利| 91精品三级在线观看| 国产欧美日韩综合在线一区二区| 王馨瑶露胸无遮挡在线观看| 欧美av亚洲av综合av国产av| a级毛片在线看网站| 午夜成年电影在线免费观看| 国产一区二区三区视频了| 欧美日本中文国产一区发布| 日本vs欧美在线观看视频| 99精品在免费线老司机午夜| 老熟妇仑乱视频hdxx| 日韩视频一区二区在线观看| 女人精品久久久久毛片| 天天躁日日躁夜夜躁夜夜| 国产野战对白在线观看| 国产欧美亚洲国产| aaaaa片日本免费| 一本综合久久免费| 国产主播在线观看一区二区| 交换朋友夫妻互换小说| 色尼玛亚洲综合影院| 久热爱精品视频在线9| 午夜福利一区二区在线看| 69精品国产乱码久久久| 天天躁日日躁夜夜躁夜夜| 国产91精品成人一区二区三区| 国产成人一区二区三区免费视频网站| 国产精品乱码一区二三区的特点 | 国产成人啪精品午夜网站| 妹子高潮喷水视频| 亚洲精品中文字幕一二三四区| 91老司机精品| 天天影视国产精品| 久久精品亚洲av国产电影网| 咕卡用的链子| 国产欧美日韩一区二区三| 国产成人免费观看mmmm| 一边摸一边做爽爽视频免费| 亚洲中文av在线| 欧美国产精品一级二级三级| 身体一侧抽搐| 最近最新中文字幕大全免费视频| 欧美最黄视频在线播放免费 | 国产精品99久久99久久久不卡| 亚洲精品美女久久av网站| 自拍欧美九色日韩亚洲蝌蚪91| 亚洲精品中文字幕在线视频| 男女免费视频国产| 人妻丰满熟妇av一区二区三区 | 狠狠婷婷综合久久久久久88av| 国产精品 欧美亚洲| 激情视频va一区二区三区| 中文字幕另类日韩欧美亚洲嫩草| 国产精品一区二区免费欧美| 久久久久久人人人人人| 电影成人av| 看片在线看免费视频| 午夜免费观看网址| 校园春色视频在线观看| 亚洲成人手机| 一进一出好大好爽视频| 精品久久久久久电影网| 一二三四在线观看免费中文在| 日韩视频一区二区在线观看| 久久午夜综合久久蜜桃| 国产一区在线观看成人免费| 后天国语完整版免费观看| 亚洲欧美精品综合一区二区三区| 色94色欧美一区二区| 成人手机av| 国产欧美日韩一区二区精品| 王馨瑶露胸无遮挡在线观看| 热99久久久久精品小说推荐| 狂野欧美激情性xxxx| 久久香蕉精品热| 韩国av一区二区三区四区| av在线播放免费不卡| 午夜老司机福利片| 亚洲色图av天堂| 99精国产麻豆久久婷婷| 精品少妇久久久久久888优播| 精品一区二区三卡| 日本wwww免费看| 国产一区二区三区视频了| 国产精品 国内视频| 无限看片的www在线观看| 啦啦啦 在线观看视频| 亚洲欧洲精品一区二区精品久久久| 色在线成人网| 18禁黄网站禁片午夜丰满| 亚洲中文字幕日韩| 久久亚洲真实| 欧美人与性动交α欧美软件| 色94色欧美一区二区| 亚洲欧美日韩高清在线视频| 久久精品aⅴ一区二区三区四区| 最新在线观看一区二区三区| 三上悠亚av全集在线观看| 视频在线观看一区二区三区| 高清欧美精品videossex| 19禁男女啪啪无遮挡网站| 久久人人爽av亚洲精品天堂| 一区福利在线观看| 欧美一级毛片孕妇| av中文乱码字幕在线| 中文字幕人妻熟女乱码| 一二三四在线观看免费中文在| 香蕉丝袜av| 亚洲av成人一区二区三| 黄色视频不卡| 91老司机精品| 热99久久久久精品小说推荐| 人妻久久中文字幕网| 日韩欧美国产一区二区入口| 99香蕉大伊视频| 国产亚洲一区二区精品| 丝袜人妻中文字幕| 国产精品一区二区精品视频观看| 亚洲欧美一区二区三区黑人| 精品少妇久久久久久888优播| 天天躁狠狠躁夜夜躁狠狠躁| 女人久久www免费人成看片| 大型av网站在线播放| 黄片大片在线免费观看| 18禁裸乳无遮挡免费网站照片 | 免费日韩欧美在线观看| 中文字幕人妻丝袜一区二区| 欧美国产精品一级二级三级| 巨乳人妻的诱惑在线观看| 久久久久久亚洲精品国产蜜桃av| 老熟女久久久| 亚洲人成电影免费在线| 国产99白浆流出| 日韩大码丰满熟妇| 欧美日韩精品网址| 亚洲一码二码三码区别大吗| 黑人操中国人逼视频| 国产男女超爽视频在线观看| 看片在线看免费视频| 中文字幕人妻丝袜一区二区| 亚洲精品av麻豆狂野| 日韩欧美一区视频在线观看| 国产精品免费大片| 亚洲美女黄片视频| 极品人妻少妇av视频| 中文字幕制服av| 99精品久久久久人妻精品| 国产一区二区三区视频了| 国产精品偷伦视频观看了| 一二三四在线观看免费中文在| 母亲3免费完整高清在线观看| 中文字幕精品免费在线观看视频| 欧美日韩亚洲综合一区二区三区_| 99久久99久久久精品蜜桃| 国产免费现黄频在线看| 99精品在免费线老司机午夜| 日本精品一区二区三区蜜桃| 久久ye,这里只有精品| 国产不卡av网站在线观看| 中文字幕最新亚洲高清| 一进一出好大好爽视频| 国产亚洲欧美98| 国产野战对白在线观看| 亚洲欧美激情综合另类| 人人妻,人人澡人人爽秒播| 国产成人av教育| 国产精品影院久久| av片东京热男人的天堂| 国产亚洲精品第一综合不卡| 国产成人欧美在线观看 | 国产av又大| 精品少妇一区二区三区视频日本电影| a级毛片在线看网站| 国产精品免费视频内射| 国产成人影院久久av| 黑人巨大精品欧美一区二区mp4| 亚洲av欧美aⅴ国产| 水蜜桃什么品种好| 丝袜人妻中文字幕| 国产视频一区二区在线看| 91国产中文字幕| 悠悠久久av| 国产极品粉嫩免费观看在线| 日韩制服丝袜自拍偷拍| 99久久国产精品久久久| 国产精品影院久久| 国产成人av教育| 女人高潮潮喷娇喘18禁视频| 国产成人一区二区三区免费视频网站| www.自偷自拍.com| 免费在线观看亚洲国产| 无遮挡黄片免费观看| 可以免费在线观看a视频的电影网站| 久久国产精品男人的天堂亚洲| 欧美成人免费av一区二区三区 | 日本a在线网址| 激情在线观看视频在线高清 | 一边摸一边抽搐一进一出视频| 久久久久久久久久久久大奶| 老司机在亚洲福利影院| 免费看十八禁软件| 一区二区三区激情视频| 久久九九热精品免费| 亚洲va日本ⅴa欧美va伊人久久| 久久久久精品人妻al黑| 免费一级毛片在线播放高清视频 | 国产单亲对白刺激| 午夜福利欧美成人| 成人18禁在线播放| 美女高潮喷水抽搐中文字幕| 国产1区2区3区精品| 国产精品一区二区在线观看99| 亚洲自偷自拍图片 自拍| 男女之事视频高清在线观看| 美女国产高潮福利片在线看| 中国美女看黄片| 亚洲精品国产区一区二| 18禁黄网站禁片午夜丰满| 精品人妻1区二区| 成年女人毛片免费观看观看9 | 狠狠狠狠99中文字幕| 少妇 在线观看| 国产精品成人在线| 每晚都被弄得嗷嗷叫到高潮| 热99re8久久精品国产| tube8黄色片| 成人18禁在线播放| 久久久水蜜桃国产精品网| 亚洲午夜理论影院| 亚洲av熟女| 999久久久精品免费观看国产| av福利片在线|