Embedded knowledge-based speech detectors for real-time recognition tasks

Speech recognition has become common in many application domains, from dictation systems for professional practices to vocal user interfaces for people with disabilities or hands-free system control. However, so far the performance of automatic speech recognition (ASR) systems are comparable to huma...

Full description

Saved in:
Bibliographic Details
Published in2006 International Conference on Parallel Processing Workshops (ICPPW'06) pp. 6 pp. - 360
Main Authors Siniscalchi, S.M., Gennaro, F., Andolina, S., Vitabile, S., Gentile, A., Sorbello, F.
Format Conference Proceeding
LanguageEnglish
Published IEEE 2006
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Speech recognition has become common in many application domains, from dictation systems for professional practices to vocal user interfaces for people with disabilities or hands-free system control. However, so far the performance of automatic speech recognition (ASR) systems are comparable to human speech recognition (HSR) only under very strict working conditions, and in general much lower. Incorporating acoustic-phonetic knowledge into ASR design has been proven a viable approach to raise ASR accuracy. Manner of articulation attributes such as vowel, stop, fricative, approximant, nasal, and silence are examples of such knowledge. Neural networks have already been used successfully as detectors for manner of articulation attributes starting from representations of speech signal frames. In this paper, the full system implementation is described. The system has a first stage for MFCC extraction followed by a second stage implementing a sinusoidal based multi-layer perceptron for speech event classification. Implementation details over a Celoxica RC203 board are given
ISBN:9780769526379
0769526373
ISSN:0190-3918
2332-5690
DOI:10.1109/ICPPW.2006.35