Academic literature on the topic 'Automatic speech recognition ; tandem features ; Articulatory features'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Automatic speech recognition ; tandem features ; Articulatory features.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Automatic speech recognition ; tandem features ; Articulatory features"

1

Zlokarnik, Igor. "Adding articulatory features to acoustic features for automatic speech recognition." Journal of the Acoustical Society of America 97, no. 5 (1995): 3246. http://dx.doi.org/10.1121/1.411699.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Lal, Partha, and Simon King. "Cross-Lingual Automatic Speech Recognition Using Tandem Features." IEEE Transactions on Audio, Speech, and Language Processing 21, no. 12 (2013): 2506–15. http://dx.doi.org/10.1109/tasl.2013.2277932.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Ghosh, Prasanta Kumar, and Shrikanth Narayanan. "Automatic speech recognition using articulatory features from subject-independent acoustic-to-articulatory inversion." Journal of the Acoustical Society of America 130, no. 4 (2011): EL251—EL257. http://dx.doi.org/10.1121/1.3634122.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Deng, Li, and Don X. Sun. "A statistical approach to automatic speech recognition using the atomic speech units constructed from overlapping articulatory features." Journal of the Acoustical Society of America 95, no. 5 (1994): 2702–19. http://dx.doi.org/10.1121/1.409839.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Ylinen, Sari, Anna-Riikka Smolander, Reima Karhila, et al. "The Effects of a Digital Articulatory Game on the Ability to Perceive Speech-Sound Contrasts in Another Language." Frontiers in Education 6 (May 20, 2021). http://dx.doi.org/10.3389/feduc.2021.612457.

Full text
Abstract:
Digital and mobile devices enable easy access to applications for the learning of foreign languages. However, experimental studies on the effectiveness of these applications are scarce. Moreover, it is not understood whether the effects of speech and language training generalize to features that are not trained. To this end, we conducted a four-week intervention that focused on articulatory training and learning of English words in 6–7-year-old Finnish-speaking children who used a digital language-learning game app Pop2talk. An essential part of the app is automatic speech recognition that ena
APA, Harvard, Vancouver, ISO, and other styles

Dissertations / Theses on the topic "Automatic speech recognition ; tandem features ; Articulatory features"

1

Lal, Partha. "Cross-lingual automatic speech recognition using tandem features." Thesis, University of Edinburgh, 2011. http://hdl.handle.net/1842/5773.

Full text
Abstract:
Automatic speech recognition requires many hours of transcribed speech recordings in order for an acoustic model to be effectively trained. However, recording speech corpora is time-consuming and expensive, so such quantities of data exist only for a handful of languages — there are many languages for which little or no data exist. Given that there are acoustic similarities between different languages, it may be fruitful to use data from a well-supported source language for the task of training a recogniser in a target language with little training data. Since most languages do not share a com
APA, Harvard, Vancouver, ISO, and other styles
2

Leung, Ka Yee. "Combining acoustic features and articulatory features for speech recognition /." View Abstract or Full-Text, 2002. http://library.ust.hk/cgi/db/thesis.pl?ELEC%202002%20LEUNGK.

Full text
Abstract:
Thesis (M. Phil.)--Hong Kong University of Science and Technology, 2002.<br>Includes bibliographical references (leaves 92-96). Also available in electronic version. Access restricted to campus users.
APA, Harvard, Vancouver, ISO, and other styles
3

Albalkhi, Rahaf. "Articulation modelling of vowels in dysarthric and non-dysarthric speech." Thesis, 2020. http://hdl.handle.net/1828/11771.

Full text
Abstract:
People with motor function disorders that cause dysarthric speech find difficulty using state-of- the-art automatic speech recognition (ASR) systems. These systems are developed based on non- dysarthric speech models, which explains the poor performance when used by individuals with dysarthria. Thus, a solution is needed to compensate for the poor performance of these systems. This thesis examines the possibility of quantifying vowels of dysarthric and non-dysarthric speech into codewords regardless of inter-speaker variability and possible to be implemented on limited- processing-capability m
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "Automatic speech recognition ; tandem features ; Articulatory features"

1

Dohen, Marion, Hélène Loevenbruck, and Harold Hill. "Recognizing Prosody from the Lips." In Visual Speech Recognition. IGI Global, 2009. http://dx.doi.org/10.4018/978-1-60566-186-5.ch014.

Full text
Abstract:
The aim of this chapter is to examine the possibility of extracting prosodic information from lip features. The authors used two lip feature measurement techniques in order to evaluate the “lip pattern” of prosodic focus in French. Two corpora with Subject-Verb-Object (SVO) sentences were designed. Four focus conditions (S, V, O or neutral) were elicited in a natural dialogue situation. In the first set of experiments, they recorded two speakers of French with front and profile video cameras. The speakers wore blue lipstick and facial markers. In the second set, the authors recorded five speakers with a 3D optical tracker. An analysis of the lip features showed that visible articulatory lip correlates of focus exist for all speakers. Two types of patterns were observed: absolute and differential. A potential outcome of this study is to provide criteria for automatic visual detection of prosodic focus from lip data.
APA, Harvard, Vancouver, ISO, and other styles
2

Aleksic, Petar S., and Aggelos K. Katsaggelos. "Lip Feature Extraction and Feature Evaluation in the Context of Speech and Speaker Recognition." In Visual Speech Recognition. IGI Global, 2009. http://dx.doi.org/10.4018/978-1-60566-186-5.ch002.

Full text
Abstract:
There has been significant work on investigating the relationship between articulatory movements and vocal tract shape and speech acoustics (Fant, 1960; Flanagan, 1965; Narayanan &amp; Alwan, 2000; Schroeter &amp; Sondhi, 1994). It has been shown that there exists a strong correlation between face motion, and vocal tract shape and speech acoustics (Grant &amp; Braida, 1991; Massaro &amp; Stork, 1998; Summerfield, 1979, 1987, 1992; Williams &amp; Katsaggelos, 2002; Yehia, Rubin, &amp; Vatikiotis-Bateson, 1998). In particular, dynamic lip information conveys not only correlated but also complimentary information to the acoustic speech information. Its integration into an automatic speech recognition (ASR) system, resulting in an audio-visual (AV) system, can potentially increase the system’s performance. Although visual speech information is usually used together with acoustic information, there are applications where visual-only (V-only) ASR systems can be employed achieving high recognition rates. Such include small vocabulary ASR (digits, small number of commands, etc.) and ASR in the presence of adverse acoustic conditions. The choice and accurate extraction of visual features strongly affect the performance of AV and V-only ASR systems. The establishment of lip features for speech recognition is a relatively new research topic. Although a number of approaches can be used for extracting and representing visual lip information, unfortunately, limited work exists in the literature in comparing the relative performance of different features. In this chapter, the authors describe various approaches for extracting and representing important visual features, review existing systems, evaluate their relative performance in terms of speech and speaker recognition rates, and discuss future research and development directions in this area.
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Automatic speech recognition ; tandem features ; Articulatory features"

1

Cetin, Ozgur, Mathew Magimai-Doss, Karen Livescu, et al. "Monolingual and crosslingual comparison of tandem features derived from articulatory and phone MLPS." In 2007 IEEE Workshop on Automatic Speech Recognition & Understanding (ASRU). IEEE, 2007. http://dx.doi.org/10.1109/asru.2007.4430080.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Zhan, Qingran, Petr Motlicek, Shixuan Du, Yahui Shan, Sifan Ma, and Xiang Xie. "Cross-lingual Automatic Speech Recognition Exploiting Articulatory Features." In 2019 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC). IEEE, 2019. http://dx.doi.org/10.1109/apsipaasc47483.2019.9023195.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Srinivasan, Gokul, Aravind Illa, and Prasanta Kumar Ghosh. "A Study on Robustness of Articulatory Features for Automatic Speech Recognition of Neutral and Whispered Speech." In ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019. http://dx.doi.org/10.1109/icassp.2019.8683103.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!