Source-filter Model
Get Source-filter Model essential facts below. View Videos or join the Source-filter Model discussion. Add Source-filter Model to your topic list for future reference or share this resource on social media.
Source-filter Model

The source-filter model represents speech as a combination of a sound source, such as the vocal cords, and a linear acoustic filter, the vocal tract. While only an approximation, the model is widely used in a number of applications such as speech synthesis and speech analysis because of its relative simplicity. It is also related to linear prediction. The development of the model is due, in large part, to the early work of Gunnar Fant, although others, notably Ken Stevens, have also contributed substantially to the models underlying acoustic analysis of speech and speech synthesis.[1] Fant built off the work of Tsutomu Chiba and Masato Kajiyama, who first showed the relationship between a vowel's acoustic properties and the shape of the vocal tract.[1]

An important assumption that is often made in the use of the source-filter model is the independence of source and filter.[1] In such cases, the model should more accurately be referred to as the "independent source-filter model".[]


In 1942, Chiba and Kajiyama published their research on vowel acoustics and the vocal tract in their book, The Vowel: Its nature and structure. By creating models of the vocal tract using X-ray photography, they were able to predict the formant frequencies of different vowels, establishing a relationship between the two. Gunnar Fant, a pioneering speech scientist, used Chiba and Kajiyama's research involving X-ray photography of the vocal tract to interpret his own data of Russian speech sounds in Acoustic Theory of Speech Production, which established the source-filter model.[2]


To varying degrees, different phonemes can be distinguished by the properties of their source(s) and their spectral shape. Voiced sounds (e.g., vowels) have at least one source due to mostly periodic glottal excitation, which can be approximated by an impulse train in the time domain and by harmonics in the frequency domain, and a filter that depends on, for example, tongue position and lip protrusion.[3] On the other hand, fricatives, such as [s] and [f], have at least one source due to turbulent noise produced at a constriction in the oral cavity or pharynx. So-called voiced fricatives, such as [z] and [v], have two sources - one at the glottis and one at the supra-glottal constriction.

Speech synthesis

In implementation of the source-filter model of speech production, the sound source, or excitation signal, is often modelled as a periodic impulse train, for voiced speech, or white noise for unvoiced speech. The vocal tract filter is, in the simplest case, approximated by an all-pole filter, where the coefficients are obtained by performing linear prediction to minimize the mean-squared error in the speech signal to be reproduced. Convolution of the excitation signal with the filter response then produces the synthesised speech.

Modeling human speech production

One possible combination of source and filter in the human vocal tract.

In human speech production, the sound source is the vocal folds, which can produce a periodic sound when constricted or an aperiodic (white noise) sound when relaxed.[4] The filter is the rest of the vocal tract, which can change shape through manipulation of the pharynx, mouth, and nasal cavity.[3] Fant roughly compares the source and filter to phonation and articulation, respectively. The source produces a number of harmonics of varying amplitudes, which travel through the vocal tract and are either amplified or attenuated to produce a speech sound.[4]

See also


  1. ^ a b c Arai, Takayuki (2004). "History of Chiba and Kajiyama and their influence in modern speech science". From Sound to Sense: 50+ Years of Discoveries in Speech Communication (PDF). pp. 115-120.
  2. ^ Fant, Gunnar. "T. Chiba and M. Kajiyama, Pioneers in Speech Acoustics". Journal of the Phonetic Society of Japan. 5 (2). doi:10.24467/onseikenkyu.5.2_4. Retrieved 2020.
  3. ^ a b Fant, Gunnar (1970). Acoustic Theory of Speech Production with Calculations Based on X-ray Studies of Russian Articulations. De Gruyter.
  4. ^ a b Zsiga, Elizabeth C. (2012). The Sounds of Language: An Introduction to Phonetics and Phonology. John Wiley & Sons. ISBN 978-1-118-34060-8.

  This article uses material from the Wikipedia page available here. It is released under the Creative Commons Attribution-Share-Alike License 3.0.



Music Scenes