SPEECH SYNTHESIS USING PERCEPTUAL LINEAR PREDICTION PARAMETERS

A method for synthesizing human speech using a linear mapping of a small set of coefficients that are speaker-independent. Preferably, the speakerindependent set of coefficients are cepstral coefficients developed during a training session using a perceptual linear predictive analysis. A linear pred...

Full description

Saved in:
Bibliographic Details
Main Authors COX, LOUIS ANTHONY, JR, HERMANSKY, HYNEK
Format Patent
LanguageEnglish
French
Published 12.12.1995
Edition6
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:A method for synthesizing human speech using a linear mapping of a small set of coefficients that are speaker-independent. Preferably, the speakerindependent set of coefficients are cepstral coefficients developed during a training session using a perceptual linear predictive analysis. A linear predictive all-pole model is used to develop corresponding formants and bandwidths to which the cepstral coefficients are mapped by using a separate multiple regression model for each of the five formant frequencies and five formant bandwidths. The dual analysis produces both the cepstral coefficients of the PLP model for the different vowel-like sounds and their true formant frequencies and bandwidths. The separate multiple regression models developed by mapping the cepstral coefficients into the formant frequencies and formant bandwidths can then be applied to cepstral coefficients determined for subsequent speech to produce corresponding formants and bandwidths used to synthesize that speech. Since less data are required for synthesizing each speech segment than in conventional techniques, areduction in the required storage space and/or transmission rate for the data required in the speech synthesis is achieved. In addition, the cepstral coefficients for each speech segment can be used with the regressive model for a different speaker, to produce synthesized speech corresponding to the different speaker.
Bibliography:Application Number: CA19922074418