In this paper, we propose design and initial implementation of a robust system which can automatically translates voice into text and text to sign language animations. Sign Language Translation Systems could significantly improve deaf lives especially in communications, exchange of information and employment of machine for translation conversations from one language to another has. Therefore, considering these points, it seems necessary to study the speech recognition. Usually, the voice recognition algorithms address three major challenges. The first is extracting feature form speech and the second is when limited sound gallery are available for recognition, and the final challenge is to improve speaker dependent to speaker independent voice recognition. Extracting feature form speech is an important stage in our method. Different procedures are available for extracting feature form speech. One of the commonest of which used in speech recognition systems is Mel-Frequency Cepstral Coefficients (MFCCs). The algorithm starts with preprocessing and signal conditioning. Next extracting feature form speech using Cepstral coefficients will be done. Then the result of this process sends to segmentation part. Finally recognition part recognizes the words and then converting word recognized to facial animation. The project is still in progress and some new interesting methods are described in the current report.
[1]
Yi-Ching Liaw,et al.
A novel encoding algorithm for vector quantization using transformed codebook
,
2009,
Pattern Recognit..
[2]
Pierre Gançarski,et al.
A global averaging method for dynamic time warping, with applications to clustering
,
2011,
Pattern Recognit..
[3]
Saeed Bagheri Shouraki,et al.
Recognition of human speech phonemes using a novel fuzzy approach
,
2007,
Appl. Soft Comput..
[4]
Xi Xiao,et al.
A hybrid SVM/DDBHMM decision fusion modeling for robust continuous digital speech recognition
,
2007,
Pattern Recognit. Lett..
[5]
James M. McQueen,et al.
Neural mechanisms for voice recognition
,
2010,
NeuroImage.
[6]
Marco Gori,et al.
A survey of hybrid ANN/HMM models for automatic speech recognition
,
2001,
Neurocomputing.
[7]
Alfred Mertins,et al.
Automatic speech recognition and speech variability: A review
,
2007,
Speech Commun..
[8]
Hynek Hermansky,et al.
RASTA-PLP speech analysis technique
,
1992,
[Proceedings] ICASSP-92: 1992 IEEE International Conference on Acoustics, Speech, and Signal Processing.
[9]
Koray Balci.
Xface: MPEG-4 based open source toolkit for 3D Facial Animation
,
2004,
AVI.