posted on 2014-10-22, 10:39authored byAbdulhussain E. Mahdi
A computer-based tool for visualisation of the vocal-tract, during speech articulation, by means of a midsagittal
view of the human head has been designed and developed. The vocal tract graphics are generated using a new approach
that is based on estimating both the area functions and the formant frequencies from the acoustic speech signal.
First, it is assumed that the speech production process is an autoregressive model. Using a linear prediction analysis, the
vocal tract area functions and the first three formants are estimated. The estimated area functions are then mapped to corresponding
mid-sagittal distances and displayed as 2D vocal tract lateral graphics. The mapping process is based on a
simple numerical algorithm and an accurate reference grid derived from x-rays for the pronunciation of a number English
vowels uttered by different speakers. To compensate for possible errors in the estimated area functions due to variation in
vocal tract length between speakers, the first two sectional distances are determined by the three formants. Experimental
results show high correlation with x-ray data and the PARAFAC analysis. The tool also displays other speech parameters
that are closely related to the production of intelligible speech and hence would be useful as a visual feedback aid for
speech training of hearing–impaired individuals and children with speech impediment.
History
Publication
Open Electrical and Electronic Engineering Journal;2, pp 27-32