AN EFFICIENT SPEECH GENERATIVE MODEL BASED ON DETERMINISTIC/STOCHASTIC SEPARATION OF SPECTRAL ENVELOPES
https://doi.org/10.35596/1729-7648-2020-18-2-23-29
Abstract
The paper presents a speech generative model that provides an efficient way of generating speech waveform from its amplitude spectral envelopes. The model is based on hybrid speech representation that includes deterministic (harmonic) and stochastic (noise) components. The main idea behind the approach originates from the fact that speech signal has a determined spectral structure that is statistically bound with deterministic/stochastic energy distribution in the spectrum. The performance of the model is evaluated using an experimental low-bitrate wide-band speech coder. The quality of reconstructed speech is evaluated using objective and subjective methods. Two objective quality characteristics were calculated: Modified Bark Spectral Distortion (MBSD) and Perceptual Evaluation of Speech Quality (PESQ). Narrow-band and wide-band versions of the proposed solution were compared with MELP (Mixed Excitation Linear Prediction) speech coder and AMR (Adaptive Multi-Rate) speech coder, respectively. The speech base of two female and two male speakers were used for testing. The performed tests show that overall performance of the proposed approach is speaker-dependent and it is better for male voices. Supposedly, this difference indicates the influence of pitch highness on separation accuracy. In that way, using the proposed approach in experimental speech compression system provides decent MBSD values and comparable PESQ values with AMR speech coder at 6,6 kbit/s. Additional subjective listening testsdemonstrate that the implemented coding system retains phonetic content and speaker’s identity. It proves consistency of the proposed approach.
About the Authors
M. TahaBelarus
Mostafa Taha, Master of Sciences, PhD student of Computer Engineering Department
Minsk
E. S. Azarov
Belarus
Elias S. Azarov, D.Sci., Professor of Computer Engineering Department
Minsk
D. S. Likhachov
Belarus
Likhachov Denis Sergeevich, PhD, Associate Professor of Computer Engineering Department
220013, Republic of Belarus, Minsk, P. Brovki str., 6; tel. +375172938805
A. A. Petrovsky
Belarus
Alexander A. Petrovsky, D.Sci., Professor of Computer Engineering Department
Minsk
References
1. A. van den Oord, Dieleman S., Zen H., Simonyan K., Vinyals O., Graves A., Kalchbrenner N., Senior A., Kavukcuoglu K. WaveNet: A generative model for raw audio, arXiv:1609.03499, 2016.
2. Shen J., Pang R., Weiss R. J., Schuster M., Jaitly N., Yang Z., Chen Z., Zhang Y., Wang Y., Skerrv-Ryan R. “Natural TTS synthesis by conditioning WaveNet on mel spectrogram predictions,” in Proc. International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2018:4779-4783.
3. Arik S., Diamos G., Gibiansky A., Miller J., Peng K., Ping W., Raiman J., and Zhou Y. Deep voice 2: Multi-speaker neural text-to-speech. arXiv:1705.08947, 2017.
4. Valin J.-V., Skoglund J. LPCNet: Improving neural speech synthesis through linear prediction, arXiv:1810.11846
5. Griffin D., Lim J. A new model-based speech analysis/synthesis system. In Proc. International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 1985;10:513-516.
6. Laroche J., Stylianou Y., Moulines E. HNS: Speech modification based on a harmonic+noise model. Proceedings of the ICASSP-93 IEEE International Conference on Acoustics, Speech, and Signal Processing, 1993;2:550-553.
7. Serra X. Musical sound modeling with sinusoids plus noise. Musical Signal Processing (C. Roads, S. Popea, A. Picialli, G. De Poli Eds.). Swets & Zeitlinger Publishers; 1997.
8. Azarov E., Petrovsky A. Instantaneous harmonic analysis for vocal processing. Proceedings of DAFx-09. Como, Italy, September 14; 2009.
Review
For citations:
Taha M., Azarov E.S., Likhachov D.S., Petrovsky A.A. AN EFFICIENT SPEECH GENERATIVE MODEL BASED ON DETERMINISTIC/STOCHASTIC SEPARATION OF SPECTRAL ENVELOPES. Doklady BGUIR. 2020;18(2):23-29. https://doi.org/10.35596/1729-7648-2020-18-2-23-29