Academic literature on the topic 'Speech recognition in education'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Speech recognition in education.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Speech recognition in education"

1

Tang, K. Wendy, Ridha Kamoua, Victor Sutan, et al. "Speech Recognition Technology for Disabilities Education." Journal of Educational Technology Systems 33, no. 2 (2004): 173–84. http://dx.doi.org/10.2190/k6k8-78k2-59y7-r9r2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Xian, Weijia. "Speech Emotion Recognition Application for Education." BCP Education & Psychology 7 (November 7, 2022): 378–83. http://dx.doi.org/10.54691/bcpep.v7i.2691.

Full text
Abstract:
Based on convolutional neural networks, a speech recognition application capable of analyzing human emotions is designed. This speech emotion recognition can better assist teachers to understand students' emotional status in the learning process and enable them to improve their teaching methods with the help of the system, thus achieving the goal of improving students' learning efficiency. The application is based on PAD dimension, convolutional neural network to extract deep speech emotion features, and Least squares support vector machine for emotion recognition, thus improving the recognition accuracy of this application.
APA, Harvard, Vancouver, ISO, and other styles
3

Abdelhamid, Abdelaziz A. "Speech Emotions Recognition for Online Education." Fusion: Practice and Applications 10, no. 1 (2023): 78–87. http://dx.doi.org/10.54216/fpa.100104.

Full text
Abstract:
The severe circumstances caused by COVID-19 make online education the best replacement for regular face-to-face education for continuing the education process. One year ago, and till now most schools adopted online learning during this pandemic shutdown, which indicates the applicability of this teaching methodology. However, the efficiency of this method needs to be improved to guarantee its effectiveness. Although face-to-face teaching has many advantages over online education, there is a chance to promote online learning by utilizing the recent techniques of artificial intelligence. From this perspective, we propose a framework to detect and recognize emotions in the speech of students during virtual classes to keep instructors updated with the feelings of students so and can behave accordingly. The approach of detecting emotions from the speech is much more helpful for cases when turning on the cameras at the student's side could be embarrassing. This case is very common, especially for schools in Middle East countries. The proposed framework can also be applied to other similar scenarios such as online meetings.
APA, Harvard, Vancouver, ISO, and other styles
4

Taylor, H. Rosemary. "Book Review: Speech Synthesis and Recognition Systems, Speech Synthesis and Recognition." International Journal of Electrical Engineering & Education 26, no. 4 (1989): 366. http://dx.doi.org/10.1177/002072098902600409.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

M Tasbolatov, N. Mekebayev, O. Mamyrbayev, M. Turdalyuly, D. Oralbekova,. "Algorithms and architectures of speech recognition systems." Psychology and Education Journal 58, no. 2 (2021): 6497–501. http://dx.doi.org/10.17762/pae.v58i2.3182.

Full text
Abstract:
Digital processing of speech signal and the voice recognition algorithm is very important for fast and accurate automatic scoring of the recognition technology. A voice is a signal of infinite information. The direct analysis and synthesis of a complex speech signal is due to the fact that the information is contained in the signal.
 Speech is the most natural way of communicating people. The task of speech recognition is to convert speech into a sequence of words using a computer program.
 This article presents an algorithm of extracting MFCC for speech recognition. The MFCC algorithm reduces the processing power by 53% compared to the conventional algorithm. Automatic speech recognition using Matlab.
APA, Harvard, Vancouver, ISO, and other styles
6

Shen, Hexue. "Application of Transfer Learning Algorithm and Real Time Speech Detection in Music Education Platform." Scientific Programming 2021 (October 11, 2021): 1–7. http://dx.doi.org/10.1155/2021/1093698.

Full text
Abstract:
Artificial intelligence (AI), particularly machine learning (ML) and neural networks (NN), has various applications and has sparked a lot of interest in the recent years due to its superior performance in a variety of tasks. Automatic speech recognition (ASR) is a technique that is becoming more important with the passage of time and is being used in our daily lives. Speech recognition is an important application of ML and NN, which is the auditory system of machines that realize the communication between humans and machines. In general, speech recognition methods are divided into three types, i.e., based on the channel model and speech knowledge method, template matching scheme, and the use of NN method. The main problem associated with the existing speech recognition methods is the low recognition accuracy and more computation time. In order to overcome the problem of low recognition accuracy of existing speech recognition techniques, a speech recognition technology based on the combination of deep convolution neural network (DCNN) algorithm and transfer learning techniques, i.e., VGG-16, is proposed in this study. Due to the limited application range of DCNN, when the input and output parameters are changed, it is necessary to reconstruct the model that leads to a long training time of the architecture. Therefore, the migration learning method is conducive to reducing the size of the dataset. Various experiments have been performed using different dataset constructs. The simulation results show that transfer learning is not only suitable for the comparison between the source dataset and the target dataset, but also suitable for two different datasets. The application of small datasets not only reduces the time and cost of dataset generation, but also reduces the training time and the requirement of computing power. From the experimental results, it is quite obvious that the proposed system performed better than the existing speech recognition methods, and its performance is superior in terms of recognition accuracy than the other approaches.
APA, Harvard, Vancouver, ISO, and other styles
7

McCrocklin, Shannon, and Idée Edalatishams. "Revisiting Popular Speech Recognition Software for ESL Speech." TESOL Quarterly 54, no. 4 (2020): 1086–97. http://dx.doi.org/10.1002/tesq.3006.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Taylor, H. Rosemary. "Book Review: Electronic Speech Recognition." International Journal of Electrical Engineering & Education 25, no. 2 (1988): 140. http://dx.doi.org/10.1177/002072098802500211.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Miller, Margaret K., Lauren Calandruccio, Emily Buss, et al. "Masked English Speech Recognition Performance in Younger and Older Spanish–English Bilingual and English Monolingual Children." Journal of Speech, Language, and Hearing Research 62, no. 12 (2019): 4578–91. http://dx.doi.org/10.1044/2019_jslhr-19-00059.

Full text
Abstract:
Purpose The purpose of this study was to compare masked English speech recognition thresholds between Spanish–English bilingual and English monolingual children and to evaluate effects of age, maternal education, and English receptive language abilities on individual differences in masked speech recognition. Method Forty-three Spanish–English bilingual children and 42 English monolingual children completed an English sentence recognition task in 2 masker conditions: (a) speech-shaped noise and (b) 2-talker English speech. Two age groups of children, younger (5–6 years) and older (9–10 years), were tested. The predictors of masked speech recognition performance were evaluated using 2 mixed-effects regression models. In the 1st model, fixed effects were age group (younger children vs. older children), language group (bilingual vs. monolingual), and masker type (speech-shaped noise vs. 2-talker speech). In the 2nd model, the fixed effects of receptive English vocabulary scores and maternal education level were also included. Results Younger children performed more poorly than older children, but no significant difference in masked speech recognition was observed between bilingual and monolingual children for either age group when English proficiency and maternal education were also included in the model. English language abilities fell within age-appropriate norms for both groups, but individual children with larger receptive vocabularies in English tended to show better recognition; this effect was stronger for younger children than for older children. Speech reception thresholds for all children were lower in the speech-shaped noise masker than in the 2-talker speech masker. Conclusions Regardless of age, similar masked speech recognition was observed for Spanish–English bilingual and English monolingual children tested in this study when receptive English language abilities were accounted for. Receptive English vocabulary scores were associated with better masked speech recognition performance for both bilinguals and monolinguals, with a stronger relationship observed for younger children than older children. Further investigation involving a Spanish-dominant bilingual sample is warranted given the high English language proficiency of children included in this study.
APA, Harvard, Vancouver, ISO, and other styles
10

Zhang, Siqi. "Exploration of the Application of Speech Recognition Technology Based on Artificial Intelligence in Daily Life." World Journal of Innovation and Modern Technology 8, no. 2 (2025): 1–4. https://doi.org/10.53469/wjimt.2025.08(02).01.

Full text
Abstract:
Speech recognition technology provides users with a more convenient and efficient interactive experience by automatically recognizing and processing speech signals. Starting from the acquisition and processing of speech signals, the foundation of speech recognition models and algorithms, and the application research of deep learning and neural networks, this article focuses on exploring the specific applications of speech recognition technology in fields such as smart homes, smart transportation, mobile devices and smartphones, education and learning, and healthcare. By analyzing practical application scenarios in various fields, the important role of speech recognition technology in improving convenience, efficiency, and promoting intelligent development has been demonstrated.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Speech recognition in education"

1

Cai, Carrie Jun. "Adapting existing games for education using speech recognition." Thesis, Massachusetts Institute of Technology, 2013. http://hdl.handle.net/1721.1/82184.

Full text
Abstract:
Thesis (S.M.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2013.<br>This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.<br>Cataloged from PDF student-submitted version of thesis.<br>Includes bibliographical references (p. 73-77).<br>Although memory exercises and arcade-style games are alike in their repetitive nature, memorization tasks like vocabulary drills tend to be mundane and tedious while arcade-style games are popular, intense and broadly addictive. The repetitive structure of arcade games suggests an opportunity to modify these well-known games for the purpose of learning. Arcade-style games like Tetris and Pac-man are often difficult to adapt for educational purposes because their fast-paced intensity and keystroke-heavy nature leave little room for simultaneous practice of other skills. Incorporating spoken language technology could make it possible for users to learn as they play, keeping up with game speed through multimodal interaction. Two challenges exist in this research: first, it is unclear which learning strategy would be most eective when incorporated into an already fast-paced, mentally demanding game. Secondly, it remains difficult to augment fast-paced games with speech interaction because the frustrating effect of recognition errors highly compromises entertainment. In this work, we designed and implemented Tetrilingo, a modified version of Tetris with speech recognition to help students practice and remember word-picture mappings. With our speech recognition prototype, we investigated the extent to which various forms of memory practice impact learning and engagement, and found that free-recall retrieval practice was less enjoyable to slower learners despite producing signicant learning benefits over alternative learning strategies. Using utterances collected from learners interacting with Tetrilingo, we also evaluated several techniques to increase speech recognition accuracy in fast-paced games by leveraging game context. Results show that, because false negative recognition errors are self-perpetuating and more prevalent than false positives, relaxing the constraints of the speech recognizer towards greater leniency may enhance overall recognition performance.<br>by Carrie Jun Cai.<br>S.M.
APA, Harvard, Vancouver, ISO, and other styles
2

Emeeshat, Janah S. "Isolated Word Speech Recognition System for Children with Down Syndrome." Youngstown State University / OhioLINK, 2017. http://rave.ohiolink.edu/etdc/view?acc_num=ysu150400900840969.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Madsen, Mikayla Nicole. "Speech Perception of Global Acoustic Structure in Children With Speech Delay, With and Without Dyslexia." BYU ScholarsArchive, 2020. https://scholarsarchive.byu.edu/etd/8937.

Full text
Abstract:
Children with speech delay (SD) have underlying deficits in speech perception that may be related to reading skill. Children with SD and children with dyslexia have previously shown deficits for distinct perceptual characteristics, including segmental acoustic structure and global acoustic structure. In this study, 35 children (ages 7-9 years) with SD, SD + dyslexia, and/or typically developing were presented with a vocoded speech recognition task to investigate their perception of global acoustic speech structure. Findings revealed no differences in vocoded speech recognition between groups, regardless of SD or dyslexia status. These findings suggest that in children with SD, co-occurring dyslexia does not appear to influence speech perception of global acoustic structure. We discuss these findings in the context of previous research literature and also discuss limitations of the current study and future directions for follow-up investigations.
APA, Harvard, Vancouver, ISO, and other styles
4

Garrett, Jennifer Tumlin. "Using Speech Recognition Software to Increase Writing Fluency for Individuals with Physical Disabilities." Digital Archive @ GSU, 2007. http://digitalarchive.gsu.edu/epse_diss/46.

Full text
Abstract:
Writing is an important skill that is necessary throughout school and life. Many students with physical disabilities, however, have difficulty with writing skills due to disability-specific factors, such as motor coordination problems. Due to the difficulties these individuals have with writing, assistive technology is often utilized. One piece of assistive technology, speech recognition software, may help remove the motor demand of writing and help students become more fluent writers. Past research on the use of speech recognition software, however, reveals little information regarding its impact on individuals with physical disabilities. Therefore, this study involved students of high school age with physical disabilities that affected hand use. Using an alternating treatments design to compare the use of word processing with the use of speech recognition software, this study analyzed first-draft writing samples in the areas of fluency, accuracy, type of word errors, recall of intended meaning, and length. Data on fluency, calculated in words correct per minute (wcpm) indicated that all participants wrote much faster with speech recognition compared to word processing. However, accuracy, calculated as percent correct, was much lower when participants used speech recognition compared to word processing. Word errors and recall of intended meaning were coded based on type and varied across participants. In terms of length, all participants wrote longer drafts when using speech recognition software, primarily because their fluency was higher, and they were able, therefore, to write more words. Although the results of this study indicated that participants wrote more fluently with speech recognition, because their accuracy was low, it is difficult to determine whether or not speech recognition is a viable solution for all individuals with physical disabilities. Therefore, additional research is needed that takes into consideration the editing and error correction time when using speech recognition software.
APA, Harvard, Vancouver, ISO, and other styles
5

Quinlan, Thomas H. "Speech recognition technology and the writing processes of students with writing difficulties : improving fluency /." Thesis, Connect to this title online; UW restricted, 2002. http://hdl.handle.net/1773/7841.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Sánchez, Cortina Isaías. "Confidence Measures for Automatic and Interactive Speech Recognition." Doctoral thesis, Universitat Politècnica de València, 2016. http://hdl.handle.net/10251/61473.

Full text
Abstract:
[EN] This thesis work contributes to the field of the {Automatic Speech Recognition} (ASR). And particularly to the {Interactive Speech Transcription} and {Confidence Measures} (CM) for ASR. The main goals of this thesis work can be summarised as follows: 1. To design IST methods and tools to tackle the problem of improving automatically generated transcripts. 2. To assess the designed IST methods and tools on real-life tasks of transcription in large educational repositories of video lectures. 3. To improve the reliability of the IST by improving the underlying (CM). Abstracts: The {Automatic Speech Recognition} (ASR) is a crucial task in a broad range of important applications which could not accomplished by means of manual transcription. The ASR can provide cost-effective transcripts in scenarios of increasing social impact such as the {Massive Open Online Courses} (MOOC), for which the availability of accurate enough is crucial even if they are not flawless. The transcripts enable search-ability, summarisation, recommendation, translation; they make the contents accessible to non-native speakers and users with impairments, etc. The usefulness is such that students improve their academic performance when learning from subtitled video lectures even when transcript is not perfect. Unfortunately, the current ASR technology is still far from the necessary accuracy. The imperfect transcripts resulting from ASR can be manually supervised and corrected, but the effort can be even higher than manual transcription. For the purpose of alleviating this issue, a novel {Interactive Transcription of Speech} (IST) system is presented in this thesis. This IST succeeded in reducing the effort if a small quantity of errors can be allowed; and also in improving the underlying ASR models in a cost-effective way. In other to adequate the proposed framework into real-life MOOCs, another intelligent interaction methods involving limited user effort were investigated. And also, it was introduced a new method which benefit from the user interactions to improve automatically the unsupervised parts ({Constrained Search} for ASR). The conducted research was deployed into a web-based IST platform with which it was possible to produce a massive number of semi-supervised lectures from two different well-known repositories, videoLectures.net and poliMedia. Finally, the performance of the IST and ASR systems can be easily increased by improving the computation of the {Confidence Measure} (CM) of transcribed words. As so, two contributions were developed: a new particular {Logistic Regresion} (LR) model; and the speaker adaption of the CM for cases in which it is possible, such with MOOCs.<br>[ES] Este trabajo contribuye en el campo del {reconocimiento automático del habla} (RAH). Y en especial, en el de la {transcripción interactiva del habla} (TIH) y el de las {medidas de confianza} (MC) para RAH. Los objetivos principales son los siguientes: 1. Diseño de métodos y herramientas TIH para mejorar las transcripciones automáticas. 2. Evaluar los métodos y herramientas TIH empleando tareas de transcripción realistas extraídas de grandes repositorios de vídeos educacionales. 3. Mejorar la fiabilidad del TIH mediante la mejora de las MC. Resumen: El {reconocimiento automático del habla} (RAH) es una tarea crucial en una amplia gama de aplicaciones importantes que no podrían realizarse mediante transcripción manual. El RAH puede proporcionar transcripciones rentables en escenarios de creciente impacto social como el de los {cursos abiertos en linea masivos} (MOOC), para el que la disponibilidad de transcripciones es crucial, incluso cuando no son completamente perfectas. Las transcripciones permiten la automatización de procesos como buscar, resumir, recomendar, traducir; hacen que los contenidos sean más accesibles para hablantes no nativos y usuarios con discapacidades, etc. Incluso se ha comprobado que mejora el rendimiento de los estudiantes que aprenden de videos con subtítulos incluso cuando estos no son completamente perfectos. Desafortunadamente, la tecnología RAH actual aún está lejos de la precisión necesaria. Las transcripciones imperfectas resultantes del RAH pueden ser supervisadas y corregidas manualmente, pero el esfuerzo puede ser incluso superior al de la transcripción manual. Con el fin de aliviar este problema, esta tesis presenta un novedoso sistema de {transcripción interactiva del habla} (TIH). Este método TIH consigue reducir el esfuerzo de semi-supervisión siempre que sea aceptable una pequeña cantidad de errores; además mejora a la par los modelos RAH subyacentes. Con objeto de transportar el marco propuesto para MOOCs, también se investigaron otros métodos de interacción inteligentes que involucran esfuerzo limitado por parte del usuario. Además, se introdujo un nuevo método que aprovecha las interacciones para mejorar aún más las partes no supervisadas (ASR con {búsqueda restringida}). La investigación en TIH llevada a cabo se desplegó en una plataforma web con el que fue posible producir un número masivo de transcripciones de videos de dos conocidos repositorios, videoLectures.net y poliMedia. Por último, el rendimiento de la TIH y los sistemas de RAH se puede aumentar directamente mediante la mejora de la estimación de la {medida de confianza} (MC) de las palabras transcritas. Por este motivo se desarrollaron dos contribuciones: un nuevo modelo discriminativo {logístico} (LR); y la adaptación al locutor de la MC para los casos en que es posible, como por ejemplo en MOOCs.<br>[CAT] Aquest treball hi contribueix al camp del {reconeixment automàtic de la parla} (RAP). I en especial, al de la {transcripció interactiva de la parla} i el de {mesures de confiança} (MC) per a RAP. Els objectius principals són els següents: 1. Dissenyar mètodes i eines per a TIP per tal de millorar les transcripcions automàtiques. 2. Avaluar els mètodes i eines TIP per a tasques de transcripció realistes extretes de grans repositoris de vídeos educacionals. 3. Millorar la fiabilitat del TIP, mitjançant la millora de les MC. Resum: El {reconeixment automàtic de la parla} (RAP) és una tasca crucial per una àmplia gamma d'aplicacions importants que no es poden dur a terme per mitjà de la transcripció manual. El RAP pot proporcionar transcripcions en escenaris de creixent impacte social com els {cursos online oberts massius} (MOOC). Les transcripcions permeten automatitzar tasques com ara cercar, resumir, recomanar, traduir; a més a més, fa accessibles els continguts als parlants no nadius i els usuaris amb discapacitat, etc. Fins i tot, pot millorar el rendiment acadèmic de estudiants que aprenen de xerrades amb subtítols, encara que aquests subtítols no siguen perfectes. Malauradament, la tecnologia RAP actual encara està lluny de la precisió necessària. Les transcripcions imperfectes resultants de RAP poden ser supervisades i corregides manualment, però aquest l'esforç pot acabar sent superior a la transcripció manual. Per tal de resoldre aquest problema, en aquest treball es presenta un sistema nou per a {transcripció interactiva de la parla} (TIP). Aquest sistema TIP va ser reeixit en la reducció de l'esforç per quan es pot permetre una certa quantitat d'errors; així com també en en la millora dels models RAP subjacents. Per tal d'adequar el marc proposat per a MOOCs, també es van investigar altres mètodes d'interacció intel·ligents amb esforç d''usuari limitat. A més a més, es va introduir un nou mètode que aprofita les interaccions per tal de millorar encara més les parts no supervisades (RAP amb {cerca restringida}). La investigació en TIP duta a terme es va desplegar en una plataforma web amb la qual va ser possible produir un nombre massiu de transcripcions semi-supervisades de xerrades de repositoris ben coneguts, videoLectures.net i poliMedia. Finalment, el rendiment de la TIP i els sistemes de RAP es pot augmentar directament mitjançant la millora de l'estimació de la {Confiança Mesura} (MC) de les paraules transcrites. Per tant, es van desenvolupar dues contribucions: un nou model discriminatiu logístic (LR); i l'adaptació al locutor de la MC per casos en que és possible, per exemple amb MOOCs.<br>Sánchez Cortina, I. (2016). Confidence Measures for Automatic and Interactive Speech Recognition [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/61473<br>TESIS
APA, Harvard, Vancouver, ISO, and other styles
7

Booth, Martin. "Combining games and speech recognition in a multilingual educational environment / M. Booth." Thesis, North-West University, 2014. http://hdl.handle.net/10394/10607.

Full text
Abstract:
Playing has been part of people's lives since the beginning of time. However, play does not take place in silence (isolated from speech and sound). The games people play allow them to interact and to learn through experiences. Speech often forms an integral part of playing games. Video games also allow players to interact with a virtual world and learn through those experiences. Speech input has previously been explored as a way of interacting with a game, as talking is a natural way of communicating. By talking to a game, the experiences created during gameplay become more valuable, which in turn facilitates effective learning. In order to enable a game to “hear", some issues need to be considered. A game, that will serve as a platform for speech input, has to be developed. If the game will contain learning elements, expert knowledge regarding the learning content needs to be obtained. The game needs to communicate with a speech recognition system, which will recognise players' speech inputs. To understand the role of speech recognition in a game, players need to be tested while playing the game. The players' experiences and opinions can then be fed back into the development of speech recognition in educational games. This process was followed with six Financial Management students on the NWU Vaal Triangle campus. The students played FinMan, a game which teaches the fundamental concepts of the “Time value of money" principle. They played the game with the keyboard and mouse, as well as via speech commands. The students shared their experiences through a focus group discussion and by completing a questionnaire. Quantitative data was collected to back the students' experiences. The results show that, although the recognition accuracies and response times are important issues, speech recognition can play an essential part in educational games. By freeing learners to focus on the game content, speech recognition can make games more accessible and engaging, and consequently lead to more effective learning experiences.<br>MSc (Computer Science), North-West University, Vaal Triangle Campus, 2014
APA, Harvard, Vancouver, ISO, and other styles
8

Khaldieh, Salim Ahmad. "The role of phonological encoding (speech recoding) and visual processes in word recognition of American learners of Arabic as a foreign language /." The Ohio State University, 1990. http://rave.ohiolink.edu/etdc/view?acc_num=osu1487685204966592.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Cesene, Daniel Fredrick. "The Completeness of the Electronic Medical Record with the Implementation of Speech Recognition Technology." Youngstown State University / OhioLINK, 2014. http://rave.ohiolink.edu/etdc/view?acc_num=ysu1401735616.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Cox, Troy L. "Investigating Prompt Difficulty in an Automatically Scored Speaking Performance Assessment." BYU ScholarsArchive, 2013. https://scholarsarchive.byu.edu/etd/3929.

Full text
Abstract:
Speaking assessments for second language learners have traditionally been expensive to administer because of the cost of rating the speech samples. To reduce the cost, many researchers are investigating the potential of using automatic speech recognition (ASR) as a means to score examinee responses to open-ended prompts. This study examined the potential of using ASR timing fluency features to predict speech ratings and the effect of prompt difficulty in that process. A speaking test with ten prompts representing five different intended difficulty levels was administered to 201 subjects. The speech samples obtained were then (a) rated by human raters holistically, (b) rated by human raters analytically at the item level, and (c) scored automatically using PRAAT to calculate ten different ASR timing fluency features. The ratings and scores of the speech samples were analyzed with Rasch measurement to evaluate the functionality of the scales and the separation reliability of the examinees, raters, and items. There were three ASR timed fluency features that best predicted human speaking ratings: speech rate, mean syllables per run, and number of silent pauses. However, only 31% of the score variance was predicted by these features. The significance in this finding is that those fluency features alone likely provide insufficient information to predict human rated speaking ability accurately. Furthermore, neither the item difficulties calculated by the ASR nor those rated analytically by the human raters aligned with the intended item difficulty levels. The misalignment of the human raters with the intended difficulties led to a further analysis that found that it was problematic for raters to use a holistic scale at the item level. However, modifying the holistic scale to a scale that examined if the response to the prompt was at-level resulted in a significant correlation (r = .98, p < .01) between the item difficulties calculated analytically by the human raters and the intended difficulties. This result supports the hypothesis that item prompts are important when it comes to obtaining quality speech samples. As test developers seek to use ASR to score speaking assessments, caution is warranted to ensure that score differences are due to examinee ability and not the prompt composition of the test.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Books on the topic "Speech recognition in education"

1

Zechner, Klaus. Toward an understanding of the role of speech recognition in nonnative speech assessment. Educational Testing Service, 2007.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
2

Yu, Dong, and Li Deng. Automatic Speech Recognition. Springer London, 2015. http://dx.doi.org/10.1007/978-1-4471-5779-3.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Lee, Kai-Fu. Automatic Speech Recognition. Springer US, 1989. http://dx.doi.org/10.1007/978-1-4615-3650-5.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Bourlard, Hervé A., and Nelson Morgan. Connectionist Speech Recognition. Springer US, 1994. http://dx.doi.org/10.1007/978-1-4615-3210-1.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Markowitz, Judith A. Using speech recognition. Prentice Hall PTR, 1996.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
6

Woelfel, Matthias. Distant speech recognition. Wiley, 2009.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
7

Holmes, J. N. Speech synthesis and recognition. 2nd ed. Taylor & Francis, 2001.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
8

Laface, Pietro, and Renato De Mori, eds. Speech Recognition and Understanding. Springer Berlin Heidelberg, 1992. http://dx.doi.org/10.1007/978-3-642-76626-8.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Neustein, Amy, ed. Advances in Speech Recognition. Springer US, 2010. http://dx.doi.org/10.1007/978-1-4419-5951-5.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Ayuso, Antonio J. Rubio, and Juan M. López Soler, eds. Speech Recognition and Coding. Springer Berlin Heidelberg, 1995. http://dx.doi.org/10.1007/978-3-642-57745-1.

Full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Book chapters on the topic "Speech recognition in education"

1

Williams, Susan M., Peter G. Fairweather, and Don Nix. "Speech Recognition to Support Early Literacy." In Interactive Literacy Education. Routledge, 2023. http://dx.doi.org/10.4324/9781003417965-5.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Eskenazi, Maxine, and Jonathan Brown. "8. Teaching the creation of software that uses speech recognition." In Teacher Education in CALL. John Benjamins Publishing Company, 2006. http://dx.doi.org/10.1075/lllt.14.13esk.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Chen, Howard Hao-Jan. "Developing a Speaking Practice Website by Using Automatic Speech Recognition Technology." In Emerging Technologies for Education. Springer International Publishing, 2017. http://dx.doi.org/10.1007/978-3-319-52836-6_71.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Iosifov, Ievgen, Olena Iosifova, Oleh Romanovskyi, Volodymyr Sokolov, and Ihor Sukailo. "Transferability Evaluation of Speech Emotion Recognition Between Different Languages." In Advances in Computer Science for Engineering and Education. Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-031-04812-8_35.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Bramley, Corinna, and Keith Morrison. "The ideal speech situation, communicative action, recognition, and student engagement." In Student Engagement, Higher Education, and Social Justice. Routledge, 2022. http://dx.doi.org/10.4324/9781003331292-4.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Scott-Baumann, Alison. "Communities of Inquiry." In SpringerBriefs in Education. Springer Nature Singapore, 2023. http://dx.doi.org/10.1007/978-981-99-3475-1_3.

Full text
Abstract:
AbstractBefore diving into Ricœur’s historical and conceptual experiences of free speech about Algeria, Nanterre and USA in Chaps. 4, 5 and 6, I provide in this chapter a clear working definition of ‘free speech’ as a negotiated process and explore how, by using the Communities of Inquiry approach, young people can learn the art of discussion as they progress through university. As an antidote to the chilling effects of the culture wars, this develops a politics of pedagogy that entails mutual recognition of each other’s arguments and helps us to share the risk with each other of causing offence. There are significant differences between this approach and the communicational ethicsof Habermas. This chapter and Chaps. 4 and 6 end with sample Communities of Inquiry.
APA, Harvard, Vancouver, ISO, and other styles
7

Suk, Yong Ho, and Seung Ho Choi. "A Cepstral PDF Normalization Method for Noise Robust Speech Recognition." In Advances in Computer Science, Environment, Ecoinformatics, and Education. Springer Berlin Heidelberg, 2011. http://dx.doi.org/10.1007/978-3-642-23324-1_7.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Štuikys, Vytautas, and Renata Burbaitė. "Speech Recognition Technology in K–12 STEM-Driven Computer Science Education." In Evolution of STEM-Driven Computer Science Education. Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-48235-9_10.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Lee, Seung Eun. "Sharing Computation Resources in Image and Speech Recognition for Embedded Systems." In Advances in Computer Science, Environment, Ecoinformatics, and Education. Springer Berlin Heidelberg, 2011. http://dx.doi.org/10.1007/978-3-642-23324-1_25.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Zeng, Taiyao. "Deep Learning in Automatic Speech Recognition (ASR): A Review." In Proceedings of the 2022 7th International Conference on Modern Management and Education Technology (MMET 2022). Atlantis Press SARL, 2022. http://dx.doi.org/10.2991/978-2-494069-51-0_23.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Speech recognition in education"

1

Gomes, Vanessa M., Ana Patrícia F. M. Mascarenhas, Ivanoé J. Rodowanski, et al. "Using Speech and Text in Emotions Recognition." In 2024 Brazilian Symposium on Robotics (SBR), and 2024 Workshop on Robotics in Education (WRE). IEEE, 2024. https://doi.org/10.1109/sbr/wre63066.2024.10838143.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Liu, Yudong, and Shuyun Pei. "Research on Spanish Classroom Speech Recognition System Based on Deep Learning Algorithm." In 2024 5th International Conference on Information Science and Education (ICISE-IE). IEEE, 2024. https://doi.org/10.1109/icise-ie64355.2024.11025398.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Yu, Wei. "Research on Intelligent System of English Communication Aid Software Based on Speech Recognition." In 2024 International Conference on Computers, Information Processing and Advanced Education (CIPAE). IEEE, 2024. https://doi.org/10.1109/cipae64326.2024.00156.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Chen, Xiaohua. "Research on Improving Speech Recognition System for Oral English Evaluation using Deep Learning Model." In 2024 International Conference on Computers, Information Processing and Advanced Education (CIPAE). IEEE, 2024. https://doi.org/10.1109/cipae64326.2024.00144.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Zhang, Nan, Shuang Zhang, Weiwei Wang, and Yumin Xu. "Implementation of Russian Short Command Speech Recognition Based on Dynamic Time Warping (DTW) Technology." In 2024 7th International Conference on Education, Network and Information Technology (ICENIT). IEEE, 2024. https://doi.org/10.1109/icenit61951.2024.00030.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Panda, Abhilasha, Abhijeet Anand, Sujit Bebortta, Subhranshu Sekhar Tripathy, and Tanmay Mukherjee. "Real-Time Hate Speech Recognition Along with Educational Feedback and Automatic Reporting." In 2024 IEEE 4th International Conference on Applied Electromagnetics, Signal Processing, & Communication (AESPC). IEEE, 2024. https://doi.org/10.1109/aespc63931.2024.10872383.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Yuan, Victor, Alan Miller, and Iain Oliver. "Extended Abstract—Immersive Education with Historical Characters: Conversational MetaHuman Based on Large Language Model, Speech Recognition and Generation." In 11th International Conference of the Immersive Learning Research Network. The Immersive Learning Research Network, 2025. https://doi.org/10.56198/c9b1h457.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Wang, Bingbing. "Design and Optimization of an Intelligent English Interaction System Based on Speech Recognition and Natural Language Processing Technology." In 2024 7th International Conference on Education, Network and Information Technology (ICENIT). IEEE, 2024. https://doi.org/10.1109/icenit61951.2024.00016.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Mingmin Gong and Qi Luo. "Speech emotion recognition in web based education." In 2007 IEEE International Conference on Grey Systems and Intelligent Services. IEEE, 2007. http://dx.doi.org/10.1109/gsis.2007.4443439.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Ibrahim, Umar Adam, Moussa Boukar Mahatma, and Muhammed Aliyu Suleiman. "Framework for Hausa Speech Recognition." In 2022 5th Information Technology for Education and Development (ITED). IEEE, 2022. http://dx.doi.org/10.1109/ited56637.2022.10051610.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Speech recognition in education"

1

Sierra Noakes, Sierra Noakes, Alison Shell, Alexis M. Murillo, et al. An Ethical and Equitable Vision of AI in Education: Learning Across 28 Exploratory Projects. Digital Promise, 2024. http://dx.doi.org/10.51388/20.500.12265/232.

Full text
Abstract:
This report shares the learnings across 28 exploratory projects from teams across K-12 school districts, nonprofits, and nonprofit and for-profit edtech companies, leveraging AI to support numerous goals across K-12 educational settings. Through this report, we aim to highlight the early successes of AI, surface the key barriers that call for cross-disciplinary and collective problem-solving, and consider the potential for each sector to drive forward an equitable future for AI in education. Preliminary findings from these projects show early evidence of AI’s effectiveness in various tasks, including translation, speech recognition, personalization, organizing and summarizing large qualitative datasets, and streamlining tasks to allow teachers more time with their students. However, these projects also experienced challenges with the current capabilities of AI, often leading to resource- and time-intensive processes, as well as difficulties around adoption and implementation. Additionally, many surfaced concerns around the ethical development and use of AI. Through this work, we have seen exciting ways that cross-sector collaborations are taking shape and gained a large sample of examples that emphasize the need for co-design to build meaningful AI-enabled tools. We call on education leaders, educators, students, product developers, nonprofits, and philanthropic organizations to step back from our day-to-day and imagine a revolutionized education system.
APA, Harvard, Vancouver, ISO, and other styles
2

Hoeferlin, David M., Brian M. Ore, Stephen A. Thorn, and David Snyder. Speech Processing and Recognition (SPaRe). Defense Technical Information Center, 2011. http://dx.doi.org/10.21236/ada540142.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Kubala, F., S. Austin, C. Barry, J. Makhoul, P. Placeway, and R. Schwartz. Byblos Speech Recognition Benchmark Results. Defense Technical Information Center, 1991. http://dx.doi.org/10.21236/ada459943.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Schwartz, Richard, and Owen Kimball. Toward Real-Time Continuous Speech Recognition. Defense Technical Information Center, 1989. http://dx.doi.org/10.21236/ada208196.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Liu, Fu-Hua, Pedro J. Moreno, Richard M. Stern, and Alejandro Acero. Signal Processing for Robust Speech Recognition. Defense Technical Information Center, 1994. http://dx.doi.org/10.21236/ada457798.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Schwartz, R., Y.-L. Chow, A. Derr, M.-W. Feng, and O. Kimball. Statistical Modeling for Continuous Speech Recognition. Defense Technical Information Center, 1988. http://dx.doi.org/10.21236/ada192054.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

STANDARD OBJECT SYSTEMS INC SHALIMAR FL. Auditory Modeling for Noisy Speech Recognition. Defense Technical Information Center, 2000. http://dx.doi.org/10.21236/ada373379.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Pfister, M. Software Package for Speaker Independent or Dependent Speech Recognition Using Standard Objects for Phonetic Speech Recognition. Defense Technical Information Center, 1998. http://dx.doi.org/10.21236/ada341198.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Ore, Brian M. Speech Recognition, Articulatory Feature Detection, and Speech Synthesis in Multiple Languages. Defense Technical Information Center, 2009. http://dx.doi.org/10.21236/ada519140.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Draelos, Timothy J., Stephen Heck, Jennifer Galasso, and Ronald Brogan. Seismic Phase Identification with Speech Recognition Algorithms. Office of Scientific and Technical Information (OSTI), 2018. http://dx.doi.org/10.2172/1474260.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!