To see the other types of publications on this topic, follow the link: Voice quality.

Journal articles on the topic 'Voice quality'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 journal articles for your research on the topic 'Voice quality.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.

1

Kreiman, Jody, Bruce R. Gerratt, Kristin Precoda, and Gerald S. Berke. "Individual Differences in Voice Quality Perception." Journal of Speech, Language, and Hearing Research 35, no. 3 (June 1992): 512–20. http://dx.doi.org/10.1044/jshr.3503.512.

Full text
Abstract:
Sixteen listeners (10 expert, 6 naive) judged the dissimilarity of pairs of voices drawn from pathological and normal populations. Separate nonmetric multidimensional scaling solutions were calculated for each listener and voice set. The correlations between individual listeners’ dissimilarity ratings were low However, scaling solutions indicated that each subject judged the voices in a reliable, meaningful way. Listeners differed more from one another in their judgments of the pathological voices (which varied widely on a number of acoustic parameters) than they did for the normal voices (which formed a much more homogeneous set acoustically). The acoustic features listeners used to judge dissimilarity were predictable from the characteristics of the stimulus sets’ only parameters that showed substantial variability were perceptually salient across listeners. These results are consistent with prototype models of voice perception They suggest that traditional means of assessing listener reliability n voice perception tasks may not be appropriate, and highlight the importance of using explicit comparisons between stimuli when studying voice quality perception
APA, Harvard, Vancouver, ISO, and other styles
2

Kreiman, Jody. "Information conveyed by voice quality." Journal of the Acoustical Society of America 155, no. 2 (February 1, 2024): 1264–71. http://dx.doi.org/10.1121/10.0024609.

Full text
Abstract:
The problem of characterizing voice quality has long caused debate and frustration. The richness of the available descriptive vocabulary is overwhelming, but the density and complexity of the information voices convey lead some to conclude that language can never adequately specify what we hear. Others argue that terminology lacks an empirical basis, so that language-based scales are inadequate a priori. Efforts to provide meaningful instrumental characterizations have also had limited success. Such measures may capture sound patterns but cannot at present explain what characteristics, intentions, or identity listeners attribute to the speaker based on those patterns. However, some terms continually reappear across studies. These terms align with acoustic dimensions accounting for variance across speakers and languages and correlate with size and arousal across species. This suggests that labels for quality rest on a bedrock of biology: We have evolved to perceive voices in terms of size/arousal, and these factors structure both voice acoustics and descriptive language. Such linkages could help integrate studies of signals and their meaning, producing a truly interdisciplinary approach to the study of voice.
APA, Harvard, Vancouver, ISO, and other styles
3

Kreiman, Jody, Bruce R. Gerratt, and Kristin Precoda. "Listener Experience and Perception of Voice Quality." Journal of Speech, Language, and Hearing Research 33, no. 1 (March 1990): 103–15. http://dx.doi.org/10.1044/jshr.3301.103.

Full text
Abstract:
Five speech-language clinicians and 5 naive listeners rated the similarity of pairs of normal and dysphonic voices. Multidimensional scaling was used to determine the voice characteristics that were perceptually important for each voice set and listener group. Solution spaces were compared to determine if clinical experience affects perceptual strategies. Naive and expert listeners attended to different aspects of voice quality when judging the similarity of voices, for both normal and pathological voices. All naive listeners used similar perceptual strategies; however, individual clinicians differed substantially in the parameters they considered important when judging similarity. These differences were large enough to suggest that care must be taken when using data averaged across clinicians, because averaging obscures important aspects of an individual’s perceptual behavior.
APA, Harvard, Vancouver, ISO, and other styles
4

Kreiman, Jody, Bruce R. Gerratt, Gail B. Kempster, Andrew Erman, and Gerald S. Berke. "Perceptual Evaluation of Voice Quality." Journal of Speech, Language, and Hearing Research 36, no. 1 (February 1993): 21–40. http://dx.doi.org/10.1044/jshr.3601.21.

Full text
Abstract:
The reliability of listeners’ ratings of voice quality is a central issue in voice research because of the clinical primacy of such ratings and because they are the standard against which other measures are evaluated. However, an extensive literature review indicates that both intrarater and interrater reliability fluctuate greatly from study to study. Further, our own data indicate that ratings of vocal roughness vary widely across individual clinicians, with a single voice often receiving nearly the full range of possible ratings. No model or theoretical framework currently exists to explain these variations, although such a model might guide development of efficient, valid, and standardized clinical protocols for voice evaluation. We propose a theoretical framework that attributes variability in ratings to several sources (including listeners’ backgrounds and biases, the task used to gather ratings, interactions between listeners and tasks, and random error). This framework may guide development of new clinical voice and speech evaluation protocols, ultimately leading to more reliable perceptual ratings and a better understanding of the perceptual qualities of pathological voices.
APA, Harvard, Vancouver, ISO, and other styles
5

Pessoa, Aline Neves, Beatriz Cavalcanti de Albuquerque Caiuby Novaes, Lilian Kuhn Pereira, and Zuleica Antonia Camargo. "Voice quality and voice dynamics data." Journal of Speech Sciences 1, no. 2 (February 3, 2021): 17–33. http://dx.doi.org/10.20396/joss.v1i2.15024.

Full text
Abstract:
Acoustic and perceptual auditory analysis procedures present themselves as clinical tools which give support to the understanding of the speech features of hearing impaired children (HIC). Voice quality stems from the overlapped action of the larynx, the supralaryngeal vocal tract and the level of muscular tension throughout the speech flow. Nonetheless, voice dynamics is characterized by frequency, duration and intensity variations. This research aimed at investigating acoustic and perceptive correlates of a HIC child’s voice and dynamic quality. The child, who has a cochlear implanted, had his speech samples collected during speech therapy sessions. The male subject (R), who uses a unilateral cochlear implant (UCI), had his speech production samples recorded when he was 5 (05 samples) and 6 years old (05 samples), and which were later tagged Cut A and Cut B respectively. The recorded corpus was acoustically analyzed through the use of the SGEXpressionEvaluator script (Barbosa, 2009) running on the free software Praat v10. The measures which were automatically extracted by the script correspond to the fundamental frequency –f0, first f0 derivative, intensity, spectral fall and long term spectrum. The perceptual auditory analysis of the voice quality was based on the VPAS-PB script (Camargo e Madureira, 2008). The perceptual auditory judgments and the acoustic measures were subjected to statistical analysis procedures. At first the, the data (perceptual and acoustic) were separately analyzed through a hierarchical and agglomerative cluster analysis. Subsequently, they were examined together through the principal component analysis. Results revealed the existence of correspondence between the acoustic and perceptual auditory data. In the audio recorded data samples from Cut B (one year after the first one) greater variability tendencies in acoustic measures of f0 could be observed associated with laryngeal hyper function at the perceptual level plus silent pauses and the reduction of speech rate. From the integrated acoustic and perceptual analysis it was possible to keep a record of the child’s oral language development process. The data analysis in this study allowed the observation of several interaction levels between the vocal tract (lip movement extension adjusts, tongue and jaw, associated with velopharyngeal adjusts and muscular tension from the larynx), plus the inspection of speech dynamics elements (habitual pitch and speech rate) of a child’s speech who has a UCI implanted during a one-year-speech-therapy-process period. This source of information made the characterization of the child’s evolution possible, especially in terms of perceptual auditory analysis descriptions being phonetically motivated by the speech dynamics quality.
APA, Harvard, Vancouver, ISO, and other styles
6

Cuciniello, Marialucia, Terry Amorese, Claudia Greco, Zoraida Callejas Carrión, Carl Vogel, Gennaro Cordasco, and Anna Esposito. "A Synthetic Voice for an Assistive Conversational Agent: A Survey to Discover Italian Preferences regarding Synthetic Voice’s Gender and Quality Level." Human Behavior and Emerging Technologies 2023 (December 28, 2023): 1–15. http://dx.doi.org/10.1155/2023/8858268.

Full text
Abstract:
Based on a previous investigation, a quantitative study aimed to identify user’ preferences towards four synthetic voices of two different quality levels (classified through the sophistication of the synthesizer: low vs. high) is proposed. The voices administered to participants were developed considering two main aspects: the voice quality (high/low) and their gender (male/female). 182 unpaid participants were recruited for the study, divided in four groups according to their age, and therefore classified as adolescents, young adults, middle-aged, and seniors. To collect data regarding each voice, randomly audited by participants, the shortened version of the Virtual Agent Voice Acceptance Questionnaire (VAVAQ) was exploited. Outcomes of the previous study revealed that the voices of high quality, regardless of their gender, received a higher acclaim by all participants examined rather than the corresponding two voices assessed as lower quality. Conversely, findings of the current study suggest that the four new groups of participants involved agreed in showing their strong preference towards the high-quality voice gendered as female compared to all the other considered voices. Regarding the two voices gendered as male, the high-quality one was considered as more original and capable to arouse positive emotional states than the low-quality one. Moreover, the high-quality male voice was judged as more natural than the female low-quality one. Results provide some insights for future directions in the user experience and design field.
APA, Harvard, Vancouver, ISO, and other styles
7

Fujimura, Osamu, Kiyoshi Honda, Hideki Kawahara, Yasuyuki Konparu, Masanori Morise, and J. C. Williams. "Noh voice quality." Logopedics Phoniatrics Vocology 34, no. 4 (January 2009): 157–70. http://dx.doi.org/10.3109/14015430903002288.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Zhang, Yu, Rongjie Huang, Ruiqi Li, JinZheng He, Yan Xia, Feiyang Chen, Xinyu Duan, Baoxing Huai, and Zhou Zhao. "StyleSinger: Style Transfer for Out-of-Domain Singing Voice Synthesis." Proceedings of the AAAI Conference on Artificial Intelligence 38, no. 17 (March 24, 2024): 19597–605. http://dx.doi.org/10.1609/aaai.v38i17.29932.

Full text
Abstract:
Style transfer for out-of-domain (OOD) singing voice synthesis (SVS) focuses on generating high-quality singing voices with unseen styles (such as timbre, emotion, pronunciation, and articulation skills) derived from reference singing voice samples. However, the endeavor to model the intricate nuances of singing voice styles is an arduous task, as singing voices possess a remarkable degree of expressiveness. Moreover, existing SVS methods encounter a decline in the quality of synthesized singing voices in OOD scenarios, as they rest upon the assumption that the target vocal attributes are discernible during the training phase. To overcome these challenges, we propose StyleSinger, the first singing voice synthesis model for zero-shot style transfer of out-of-domain reference singing voice samples. StyleSinger incorporates two critical approaches for enhanced effectiveness: 1) the Residual Style Adaptor (RSA) which employs a residual quantization module to capture diverse style characteristics in singing voices, and 2) the Uncertainty Modeling Layer Normalization (UMLN) to perturb the style attributes within the content representation during the training phase and thus improve the model generalization. Our extensive evaluations in zero-shot style transfer undeniably establish that StyleSinger outperforms baseline models in both audio quality and similarity to the reference singing voice samples. Access to singing voice samples can be found at https://stylesinger.github.io/.
APA, Harvard, Vancouver, ISO, and other styles
9

Eskenazi, L., D. G. Childers, and D. M. Hicks. "Acoustic Correlates of Vocal Quality." Journal of Speech, Language, and Hearing Research 33, no. 2 (June 1990): 298–306. http://dx.doi.org/10.1044/jshr.3302.298.

Full text
Abstract:
We have investigated the relationship between various voice qualities and several acoustic measures made from the vowel /i/ phonated by subjects with normal voices and patients with vocal disorders. Among the patients (pathological voices), five qualities were investigated: overall severity, hoarseness, breathiness, roughness, and vocal fry. Six acoustic measures were examined. With one exception, all measures were extracted from the residue signal obtained by inverse filtering the speech signal using the linear predictive coding (LPC) technique. A formal listening test was implemented to rate each pathological voice for each vocal quality. A formal listening test also rated overall excellence of the normal voices. A scale of 1–7 was used. Multiple linear regression analysis between the results of the listening test and the various acoustic measures was used with the prediction sums of squares (PRESS) as the selection criteria. Useful prediction equations of order two or less were obtained relating certain acoustic measures and the ratings of pathological voices for each of the five qualities. The two most useful parameters for predicting vocal quality were the Pitch Amplitude (PA) and the Harmonics-to-Noise Ratio (HNR). No acoustic measure could rank the normal voices.
APA, Harvard, Vancouver, ISO, and other styles
10

Delgado Hernández, Jonathan, Nieves M. León Gómez, Alejandra Jiménez, Laura M. Izquierdo, and Ben Barsties v. Latoszek. "Validation of the Acoustic Voice Quality Index Version 03.01 and the Acoustic Breathiness Index in the Spanish language." Annals of Otology, Rhinology & Laryngology 127, no. 5 (February 28, 2018): 317–26. http://dx.doi.org/10.1177/0003489418761096.

Full text
Abstract:
Objective: The aim of this study was to validate the Acoustic Voice Quality Index 03.01 (AVQIv3) and the Acoustic Breathiness Index (ABI) in the Spanish language. Method: Concatenated voice samples of continuous speech (cs) and sustained vowel (sv) from 136 subjects with dysphonia and 47 vocally healthy subjects were perceptually judged for overall voice quality and breathiness severity. First, to reach a higher level of ecological validity, the proportions of cs and sv were equalized regarding the time length of 3 seconds sv part and voiced cs part, respectively. Second, concurrent validity and diagnostic accuracy were verified. Results: A moderate reliability of overall voice quality and breathiness severity from 5 experts was used. It was found that 33 syllables as standardization of the cs part, which represents 3 seconds of voiced cs, allows the equalization of both speech tasks. A strong correlation was revealed between AVQIv3 and overall voice quality and ABI and perceived breathiness severity. Additionally, the best diagnostic outcome was identified at a threshold of 2.28 and 3.40 for AVQIv3 and ABI, respectively. Conclusions: The AVQIv3 and ABI showed in the Spanish language valid and robust results to quantify abnormal voice qualities regarding overall voice quality and breathiness severity.
APA, Harvard, Vancouver, ISO, and other styles
11

Duarte-Borquez, Claudia, Maxine Van Doren, and Marc Garellek. "Utterance-Final Voice Quality in American English and Mexican Spanish Bilinguals." Languages 9, no. 3 (February 21, 2024): 70. http://dx.doi.org/10.3390/languages9030070.

Full text
Abstract:
We investigate utterance-final voice quality in bilinguals of English and Spanish, two languages which differ in the type of non-modal voice usually encountered at ends of utterances: American English often has phrase-final creak, whereas in Mexican Spanish, phrase-final voiced sounds are breathy or even devoiced. Twenty-one bilinguals from the San Diego-Tijuana border region were recorded (with electroglottography and audio) reading passages in English and Spanish. Ends of utterances were coded for their visual voice quality as “modal” (having no aspiration noise or voicing irregularity), “breathy” (having aspiration noise), “creaky” (having voicing irregularity), or “breathy-creaky” (having both aspiration noise and voicing irregularity). In utterance-final position, speakers showed more frequent use of both modal and creaky voice when speaking in English, and more frequent use of breathy and breathy-creaky voice when speaking in Spanish. We find no role of language dominance on the rates of these four voice qualities. The electroglottographic and acoustic analyses show that all voice qualities, even utterance-final creak, are produced with increased glottal spreading; the combination of distinct noise measures and amplitude of voicing can distinguish breathy, creaky, and breathy-creaky voice qualities from one another, and from modal voice.
APA, Harvard, Vancouver, ISO, and other styles
12

Bučević, Anđela, Ana Bonetti, and Luka Bonetti. "The voice quality of sports coaches." Logopedija 8, no. 1 (June 29, 2018): 1–5. http://dx.doi.org/10.31299/log.8.1.1.

Full text
Abstract:
The aim of this research paper was to examine the voice quality of sports coaches using the objective (acoustic) method. A total of 28 sports coaches (mean age 28.58, SD=5.08), from the City of Zagreb participated in this research. Recordings of the phonation of the vowel /a/ before and after one training session were obtained and analyzed using the PRAAT Program. Mean, minimal and maximal values of fundamental frequency, shimmer, jitter and harmonics-to-noise ratio were observed. The statistical analyses showed no statistically significant difference in acoustic voice quality of male and female coaches before and after the training session, or between male and female coaches. However, intra-individual differences among participants were observed, which may be significant in terms of their potential to affect the quality of their voices in the future.
APA, Harvard, Vancouver, ISO, and other styles
13

Steinhauer, Kimberly. "The Estill Voice Model: A paradigm for voice training and treatment." Journal of the Acoustical Society of America 154, no. 4_supplement (October 1, 2023): A352—A353. http://dx.doi.org/10.1121/10.0023770.

Full text
Abstract:
For over 40 years, the Estill Voice Model (EVM) has defined voice quality according to movement of anatomy and physiology. EVM addresses the daunting degrees of freedom issue in voice motor control by isolating Craft of voice production from Artistry and Performance Metaphysics. The EVM proposes an integrated implicit-explicit approach for voice motor learning that flows through all training and therapy protocols. Implicit instructions include auditory-perceptual prompts (e.g., quack like a duck to produce “twang”) and explicit prompts train physiologic conditions of the vocal anatomy correlated with the voice quality (e.g., narrow your aryepiglottic sphincter to produce “twang”). Estill exercises address power, source, and filter properties of voice production, and include narrowing the aryepiglottic sphincter for “ring” in opera and belt and for increased power in hypofunctional voices, and varying vocal fold mass for register shifts and optimizing contact for hyperfunctional voices. Patients learn to feel, see, and hear the voice via multiple feedback modes including hand gestures, magnitude estimation of bodily kinesthetic effort, visual acoustic cues in real-time spectral analysis programs. This presentation will highlight objective measurement science and clinical evidence for using Estill exercises in treatment for all voices, from the novice speaker to the expert performer.
APA, Harvard, Vancouver, ISO, and other styles
14

Linh, Vuong Thuy, Nguyen Van Vu, and Le Ngoc Giang. "Voice Signal Quality Assessment Based on Signal Quality Standards and Analysis." International Journal of Research Publication and Reviews 4, no. 6 (June 2023): 958–63. http://dx.doi.org/10.55248/gengpi.4.623.44854.

Full text
APA, Harvard, Vancouver, ISO, and other styles
15

Ninh Khánh, Duy. "Evaluation of speaker-dependent and average-voice Vietnamese statistical speech synthesis systems." Journal of Science and Technology Issue on Information and Communications Technology 17, no. 12.1 (December 31, 2019): 11. http://dx.doi.org/10.31130/jst-ud2019-035e.

Full text
Abstract:
This paper describes the development and evaluation of a Vietnamese statistical speech synthesis system using the average voice approach. Although speaker-dependent systems have been applied extensively, no average voice based system has been developed for Vietnamese so far. We have collected speech data from several Vietnamese native speakers and employed state-of-the-art speech analysis, model training and speaker adaptation techniques to develop the system. Besides, we have performed perceptual experiments to compare the quality of speaker-adapted (SA) voices built on the average voice model and speaker-dependent (SD) voices built on SD models, and to confirm the effects of contextual features including word boundary (WB) and part-of-speech (POS) on the quality of synthetic speech. Evaluation results show that SA voices have significantly higher naturalness than SD voices when the same limited contextual feature set excluding WB and POS is used. In addition, SA voices trained with limited contextual features excluding WB and POS still have better quality than SD voices trained with full contextual features including WB and POS. These results show the robustness of the average voice method over the speaker-dependent approach for Vietnamese statistical speech synthesis.
APA, Harvard, Vancouver, ISO, and other styles
16

Watterson, Thomas, Stephen C. McFarlane, and Kari L. Diamond. "Phoneme Effects on Vocal Effort and Vocal Quality." American Journal of Speech-Language Pathology 2, no. 2 (May 1993): 74–78. http://dx.doi.org/10.1044/1058-0360.0202.74.

Full text
Abstract:
This study demonstrated that "vocal effort" is a legitimate laryngeal perception that can be detected by many subjects who have voice disorders and by some normal control subjects. Further, subjects with voice disorders found that the degree of vocal effort in speech was greatest for voiceless obstruent consonants, followed by voiced obstruents; sonorants and nasals required the least effort. A panel of listeners, however, could not detect differences in roughness, breathiness, intonation, or overall vocal quality as the perceived vocal effort varied.
APA, Harvard, Vancouver, ISO, and other styles
17

Craig, Scotty D., Erin K. Chiou, and Noah L. Schroeder. "The Impact of Virtual Human Voice on Learner Trust." Proceedings of the Human Factors and Ergonomics Society Annual Meeting 63, no. 1 (November 2019): 2272–76. http://dx.doi.org/10.1177/1071181319631517.

Full text
Abstract:
The current study investigates if a virtual human’s voice can impact the user’s trust in interacting with the virtual human in a learning setting. It was hypothesized that trust is a malleable factor impacted by the quality of the virtual human’s voice. A randomized alternative treatments design with a pretest placed participants in either a low-quality Text-to-Speech (TTS) engine female voice (Microsoft speech engine), a high-quality TTS engine female voice (Neospeech voice engine), or a human voice (native female English speaker) condition. All three treatments were paired with the same female virtual human. Assessments for the study included a self-report pretest on knowledge of meteorology, which occurred before viewing the instructional video, and a measure of system trust. The current study found that voice type impacts a user’s trust ratings, with the human voice resulting in higher ratings compared to the two synthetic voices.
APA, Harvard, Vancouver, ISO, and other styles
18

Cistola, Giorgia, Alex Peiró-Lilja, Guillermo Cámbara, Ineke van der Meulen, and Mireia Farrús. "Influence of TTS Systems Performance on Reaction Times in People with Aphasia." Applied Sciences 11, no. 23 (November 29, 2021): 11320. http://dx.doi.org/10.3390/app112311320.

Full text
Abstract:
Text-to-speech (TTS) systems provide fundamental reading support for people with aphasia and reading difficulties. However, artificial voices are more difficult to process than natural voices. The current study is an extended analysis of the results of a clinical experiment investigating which, among three artificial voices and a digitised human voice, is more suitable for people with aphasia and reading impairments. Such results show that the voice synthesised with Ogmios TTS, a concatenative speech synthesis system, caused significantly slower reaction times than the other three voices used in the experiment. The present study explores whether and what voice quality metrics are linked to delayed reaction times. For this purpose, the voices were analysed using an automatic assessment of intelligibility, naturalness, and jitter and shimmer voice quality parameters. This analysis revealed that Ogmios TTS, in general, performed worse than the other voices in all parameters. These observations could explain the significantly delayed reaction times in people with aphasia and reading impairments when listening to Ogmios TTS and could open up consideration about which TTS to choose for compensative devices for these patients based on the voice analysis of these parameters.
APA, Harvard, Vancouver, ISO, and other styles
19

Chasaide, Ailbhe Ni, and Christer Gobl. "Toward a voice source characterization of voice quality." Journal of the Acoustical Society of America 88, S1 (November 1990): S152. http://dx.doi.org/10.1121/1.2028681.

Full text
APA, Harvard, Vancouver, ISO, and other styles
20

Mekiš, Jana, Primož Strojan, Dušan Mekiš, and Irena Hočevar Boltežar. "Change in Voice Quality after Radiotherapy for Early Glottic Cancer." Cancers 14, no. 12 (June 17, 2022): 2993. http://dx.doi.org/10.3390/cancers14122993.

Full text
Abstract:
Our aim was to track the changes in voice quality for two years after radiotherapy (RT) for early glottic cancer. A videoendostroboscopy, subjective patient and phoniatrician voice assessments, a Voice Handicap Index questionnaire, and objective acoustic measurements (F0, jitter, shimmer, maximal phonation time) were performed on 50 patients with T1 glottic carcinomas at 3, 12, and 24 months post-RT. The results were compared between the subsequent assessments, and between the assessments at 3 months and 24 months post-RT. The stroboscopy showed a gradual progression of fibrosis of the vocal folds with a significant difference apparent when the assessments at 3 months and 24 months were compared (p < 0.001). Almost all of the subjective assessments of voice quality showed an improvement during the first 2 years, but significant differences were noted at 24 months. Jitter and shimmer deteriorated in the first year after RT with a significant deterioration noticed between the sixth and twelfth months (p = 0.048 and p = 0.002, respectively). Two years after RT, only 8/50 (16%) patients had normal voices. The main reasons for a decreased voice quality after RT for early glottic cancer were post-RT changes in the larynx. Despite a significant improvement in the voice after RT shown in a few of the evaluation methods, only a minority of the patients had a normal voice two years post-RT.
APA, Harvard, Vancouver, ISO, and other styles
21

Podesva, Robert J., and Patrick Callier. "Voice Quality and Identity." Annual Review of Applied Linguistics 35 (March 2015): 173–94. http://dx.doi.org/10.1017/s0267190514000270.

Full text
Abstract:
ABSTRACTVariation in voice quality has long been recognized to have functions beyond the grammatically distinctive or phonetically useful roles it plays in many languages, indexing information about the speaker, participating in the construction of stance in interaction, or serving to identify the speaker as a unique individual. Though the links between voice quality and identity have been studied in phonetics, sociolinguistics, linguistic anthropology, forensic linguistics, and speech technology, considerable work remains to be done to problematize the ways in which the voice is taken as covering privileged, immediate meanings about the speaker's body and to break apart the ideologies that construct it as an inalienable, unitary, and invariant facet of a speaker's identity. We point out promising directions in recent research on the voice and bring up ideas for where this important area of research should be taken.
APA, Harvard, Vancouver, ISO, and other styles
22

Bele, Irene Velsvik. "Dimensionality in Voice Quality." Journal of Voice 21, no. 3 (May 2007): 257–72. http://dx.doi.org/10.1016/j.jvoice.2005.12.001.

Full text
APA, Harvard, Vancouver, ISO, and other styles
23

Dejonckere, P. H., and J. Lebacq. "Plasticity of Voice Quality." Journal of Voice 15, no. 2 (June 2001): 251–56. http://dx.doi.org/10.1016/s0892-1997(01)00025-x.

Full text
APA, Harvard, Vancouver, ISO, and other styles
24

Ferrell, Tara, Aaron Crowson, and Christopher Mayhorn. "How We Perceive and Trust Advice from Virtual Humans: The Influence of Voice Quality." Proceedings of the Human Factors and Ergonomics Society Annual Meeting 66, no. 1 (September 2022): 1189–93. http://dx.doi.org/10.1177/1071181322661440.

Full text
Abstract:
Due to the increase in virtual humans as a pedagogical agent, this study investigates how perceptions of virtual humans are affected by voice quality in understanding a budgeting scenario where trust is essential to learning and application. Eighty-six participants were randomly assigned to three conditions where voice quality for a virtual human varied (human voice, low quality text-to-speech, high quality text-to-speech) when narrating a financial literacy course. Measures of trust and course comprehension were collected. Results of the learning assessment suggest no difference in comprehension based on voice quality of a virtual human. No differences were observed between the voice quality groups in participants' perception of trust, the abilities to facilitate learning, credibility of the agent, human-likeness of the agent, or how engaging the agent was. This is possibly due to the lower age demographic who have become increasingly exposed to virtual human voices through popular platforms such as Tik-Tok.).
APA, Harvard, Vancouver, ISO, and other styles
25

Grillo, Elizabeth U., and Jenna N. Brosious. "Results of a Voice-Related Survey of Physical Education Student Teachers." Communication Disorders Quarterly 40, no. 2 (May 22, 2018): 99–108. http://dx.doi.org/10.1177/1525740118774207.

Full text
Abstract:
The current study investigated physical education (PE) student teachers’ understanding of the vocal demands of their future profession, interest in participating in a voice-training program, and the current impact of the voice on quality-of-life by the Voice Handicap Index (VHI). Seventy-four PE student teachers completed a voice-related survey and the VHI. Forty-three percent of participants indicated that teaching will negatively affect the voice and 29% of participants reported that they may develop a voice problem because of teaching; however, only 17% of participants indicated that a voice-training program was needed to learn healthy and effective voice use for teaching. Based on the results, participants knew that teaching may negatively impact the voice, but they were not convinced that a voice-training program was necessary. In addition, the student teachers’ voices were not negatively impacting quality-of-life, as determined by the VHI.
APA, Harvard, Vancouver, ISO, and other styles
26

Uloza, Virgilijus, Kipras Pribuišis, Nora Ulozaite-Staniene, Tadas Petrauskas, Robertas Damaševičius, and Rytis Maskeliūnas. "Accuracy Analysis of the Multiparametric Acoustic Voice Indices, the VWI, AVQI, ABI, and DSI Measures, in Differentiating between Normal and Dysphonic Voices." Journal of Clinical Medicine 13, no. 1 (December 23, 2023): 99. http://dx.doi.org/10.3390/jcm13010099.

Full text
Abstract:
The study aimed to investigate and compare the accuracy and robustness of the multiparametric acoustic voice indices (MAVIs), namely the Dysphonia Severity Index (DSI), Acoustic Voice Quality Index (AVQI), Acoustic Breathiness Index (ABI), and Voice Wellness Index (VWI) measures in differentiating normal and dysphonic voices. The study group consisted of 129 adult individuals including 49 with normal voices and 80 patients with pathological voices. The diagnostic accuracy of the investigated MAVI in differentiating between normal and pathological voices was assessed using receiver operating characteristics (ROC). Moderate to strong positive linear correlations were observed between different MAVIs. The ROC statistical analysis revealed that all used measurements manifested in a high level of accuracy (area under the curve (AUC) of 0.80 and greater) and an acceptable level of sensitivity and specificity in discriminating between normal and pathological voices. However, with AUC 0.99, the VWI demonstrated the highest diagnostic accuracy. The highest Youden index equaled 0.93, revealing that a VWI cut-off of 4.45 corresponds with highly acceptable sensitivity (97.50%) and specificity (95.92%). In conclusion, the VWI was found to be beneficial in describing differences in voice quality status and discriminating between normal and dysphonic voices based on clinical diagnosis, i.e., dysphonia type, implying the VWI’s reliable voice screening potential.
APA, Harvard, Vancouver, ISO, and other styles
27

Yau-ni Wan, Jenny. "Construing negotiation." Language and Dialogue 7, no. 2 (October 16, 2017): 137–62. http://dx.doi.org/10.1075/ld.7.2.01yau.

Full text
Abstract:
Abstract The call centre conversation is a telephonic exchange of voices between the customer and the customer service representative (CSR). Both lexicogrammatical and prosodic features are used to construe emotional and attitudinal recognition. Studying these features can investigate how the call centre discourse is construed, and how the interpersonal meaning takes shape through the text. The spoken data are constructed by Filipino CSRs and American English-speaking customers. The findings show that participants tend to make specific paralinguistic voice quality choices to express their emotions in dialogue. This article first discusses the voice quality framework for its semiotic features in relation to interpersonal meaning, reviews previous voice quality studies and later delineates how voice quality relates to interpersonal meaning in the calls.
APA, Harvard, Vancouver, ISO, and other styles
28

Arias, Mariano Rosique, JosÉ Luis RamÓN, Matilde Campos, and Juan JimÉNez Cervantes. "Acoustic analysis of the voice in phonatory fistuloplasty after total laryngectomy." Otolaryngology–Head and Neck Surgery 122, no. 5 (May 2000): 743–47. http://dx.doi.org/10.1067/mhn.2000.98359.

Full text
Abstract:
A comparative study of the voice with sustained phonation of the vowel /a/ was made in 3 groups of male patients: (1) 20 patients receiving total laryngectomy for epidermoid carcinoma of the larynx who had acquired good voice quality after a phonatory fistuloplasty with a Herrmann voice prosthesis; (2) 20 patients undergoing total laryngectomy for epidermoid carcinoma of the larynx who had learned esophageal speech; and (3) 20 subjects with normal voices. Statistical analysis yielded significant differences in fundamental voice frequency between the 3 groups, with the patients with phonatory prostheses revealing the closest to a normal voice. For other parameters used, such as jitter, shimmer, and harmonics/noise ratio, voice quality with a phonatory prosthesis was similar to that obtained with esophageal speech.
APA, Harvard, Vancouver, ISO, and other styles
29

Arias, Mariano Rosique, José Luis Ramón, Matilde Campos, and Juan Jiménez Cervantes. "Acoustic Analysis of the Voice in Phonatory Fistuloplasty after Total Laryngectomy." Otolaryngology–Head and Neck Surgery 122, no. 5 (May 2000): 743–47. http://dx.doi.org/10.1016/s0194-5998(00)70208-7.

Full text
Abstract:
A comparative study of the voice with sustained phonation of the vowel /a/ was made in 3 groups of male patients: (1) 20 patients receiving total laryngectomy for epidermoid carcinoma of the larynx who had acquired good voice quality after a phonatory fistuloplasty with a Herrmann voice prosthesis; (2) 20 patients undergoing total laryngectomy for epidermoid carcinoma of the larynx who had learned esophageal speech; and (3) 20 subjects with normal voices. Statistical analysis yielded significant differences in fundamental voice frequency between the 3 groups, with the patients with phonatory prostheses revealing the closest to a normal voice. For other parameters used, such as jitter, shimmer, and harmonics/noise ratio, voice quality with a phonatory prosthesis was similar to that obtained with esophageal speech.
APA, Harvard, Vancouver, ISO, and other styles
30

Wolfe, Virginia I., David P. Martin, and Chester I. Palmer. "Perception of Dysphonic Voice Quality by Naive Listeners." Journal of Speech, Language, and Hearing Research 43, no. 3 (June 2000): 697–705. http://dx.doi.org/10.1044/jslhr.4303.697.

Full text
Abstract:
For clinical assessment as well as student training, there is a need for information pertaining to the perceptual dimensions of dysphonic voice. To this end, 24 naive listeners judged the similarity of 10 female and 10 male vowel samples, selected from within a narrow range of fundamental frequencies. Most of the perceptual variance for both sets of voices was associated with "degree of abnormality" as reflected by perceptual ratings as well as combined acoustic measures, based upon filtered and unfiltered signals. A second perceptual dimension for female voices was associated with high frequency noise as reflected by two acoustic measures: breathiness index (BRI) and a high-frequency power ratio. A second perceptual dimension for male voices was associated with a breathy-overtight continuum as reflected by period deviation (PDdev) and perceptual ratings of breathiness. Results are discussed in terms of perceptual training and the clinical assessment of pathological voices.
APA, Harvard, Vancouver, ISO, and other styles
31

Uloza, Virgilijus, Nora Ulozaitė-Stanienė, Tadas Petrauskas, Kipras Pribuišis, Tomas Blažauskas, Robertas Damaševičius, and Rytis Maskeliūnas. "Reliability of Universal-Platform-Based Voice Screen Application in AVQI Measurements Captured with Different Smartphones." Journal of Clinical Medicine 12, no. 12 (June 18, 2023): 4119. http://dx.doi.org/10.3390/jcm12124119.

Full text
Abstract:
The aim of the study was to develop a universal-platform-based (UPB) application suitable for different smartphones for estimation of the Acoustic Voice Quality Index (AVQI) and evaluate its reliability in AVQI measurements and normal and pathological voice differentiation. Our study group consisted of 135 adult individuals, including 49 with normal voices and 86 patients with pathological voices. The developed UPB “Voice Screen” application installed on five iOS and Android smartphones was used for AVQI estimation. The AVQI measures calculated from voice recordings obtained from a reference studio microphone were compared with AVQI results obtained using smartphones. The diagnostic accuracy of differentiating normal and pathological voices was evaluated by applying receiver-operating characteristics. One-way ANOVA analysis did not detect statistically significant differences between mean AVQI scores revealed using a studio microphone and different smartphones (F = 0.759; p = 0.58). Almost perfect direct linear correlations (r = 0.991–0.987) were observed between the AVQI results obtained with a studio microphone and different smartphones. An acceptable level of precision of the AVQI in discriminating between normal and pathological voices was yielded, with areas under the curve (AUC) displaying 0.834–0.862. There were no statistically significant differences between the AUCs (p > 0.05) obtained from studio and smartphones’ microphones. The significant difference revealed between the AUCs was only 0.028. The UPB “Voice Screen” application represented an accurate and robust tool for voice quality measurements and normal vs. pathological voice screening purposes, demonstrating the potential to be used by patients and clinicians for voice assessment, employing both iOS and Android smartphones.
APA, Harvard, Vancouver, ISO, and other styles
32

Grillo, Elizabeth U. "A Nonrandomized Trial for Student Teachers of an In-Person and Telepractice Global Voice Prevention and Therapy Model With Estill Voice Training Assessed by the VoiceEvalU8 App." American Journal of Speech-Language Pathology 30, no. 2 (March 26, 2021): 566–83. http://dx.doi.org/10.1044/2020_ajslp-20-00200.

Full text
Abstract:
Purpose This study investigated the effects of the in-person and telepractice Global Voice Prevention and Therapy Model (GVPTM) treatment conditions and a control condition with vocally healthy student teachers. Method In this single-blinded, nonrandomized trial, 82 participants completed all aspects of the study. Estill Voice Training was used as the stimulability component of the GVPTM to train multiple new voices meeting all the vocal needs of the student teachers. Outcomes were assessed using acoustic, perceptual, and aerodynamic measures captured by the VoiceEvalU8 app at pre and post in fall and during student teaching in spring. Results Significant improvements were achieved for several acoustic and perceptual measures in the treatment conditions, but not in the control condition. The in-person and telepractice conditions produced similar results. The all-voiced phrase and connected speech were more successful in demonstrating voice change for some of the perturbation measures as compared to sustained /a/. Conclusions The treatment conditions were successful in improving the participants' voices for fundamental frequency and some acoustic perturbation measures while maintaining the improvements during student teaching. In addition, the treatment conditions were successful in decreasing the negative impact of voice-related quality of life and vocal fatigue during student teaching. Future research should address the effectiveness of the various components of the GVPTM, the application of the GVPTM with patients with voice disorders, the relevance of defining auditory–perceptual terms by the anatomy and physiology of the voice production system (i.e., Estill Voice Training), and the continued use of the VoiceEvalU8 app for clinical voice investigations. Supplemental Material https://doi.org/10.23641/asha.13626824
APA, Harvard, Vancouver, ISO, and other styles
33

Coto-Jiménez, Marvin. "Discriminative Multi-Stream Postfilters Based on Deep Learning for Enhancing Statistical Parametric Speech Synthesis." Biomimetics 6, no. 1 (February 7, 2021): 12. http://dx.doi.org/10.3390/biomimetics6010012.

Full text
Abstract:
Statistical parametric speech synthesis based on Hidden Markov Models has been an important technique for the production of artificial voices, due to its ability to produce results with high intelligibility and sophisticated features such as voice conversion and accent modification with a small footprint, particularly for low-resource languages where deep learning-based techniques remain unexplored. Despite the progress, the quality of the results, mainly based on Hidden Markov Models (HMM) does not reach those of the predominant approaches, based on unit selection of speech segments of deep learning. One of the proposals to improve the quality of HMM-based speech has been incorporating postfiltering stages, which pretend to increase the quality while preserving the advantages of the process. In this paper, we present a new approach to postfiltering synthesized voices with the application of discriminative postfilters, with several long short-term memory (LSTM) deep neural networks. Our motivation stems from modeling specific mapping from synthesized to natural speech on those segments corresponding to voiced or unvoiced sounds, due to the different qualities of those sounds and how HMM-based voices can present distinct degradation on each one. The paper analyses the discriminative postfilters obtained using five voices, evaluated using three objective measures, Mel cepstral distance and subjective tests. The results indicate the advantages of the discriminative postilters in comparison with the HTS voice and the non-discriminative postfilters.
APA, Harvard, Vancouver, ISO, and other styles
34

Yousuf, Assila, and David Solomon George. "A hybrid CNN-LSTM model with adaptive instance normalization for one shot singing voice conversion." AIMS Electronics and Electrical Engineering 8, no. 3 (2024): 282–300. http://dx.doi.org/10.3934/electreng.2024013.

Full text
Abstract:
<abstract><p>Singing voice conversion methods encounter challenges in achieving a delicate balance between synthesis quality and singer similarity. Traditional voice conversion techniques primarily emphasize singer similarity, often leading to robotic-sounding singing voices. Deep learning-based singing voice conversion techniques, however, focus on disentangling singer-dependent and singer-independent features. While this approach can enhance the quality of synthesized singing voices, many voice conversion systems still grapple with the issue of singer-dependent feature leakage into content embeddings. In the proposed singing voice conversion technique, an encoder decoder framework was implemented using a hybrid model of convolutional neural network (CNN) accompanied by long short term memory (LSTM). This paper investigated the use of activation guidance and adaptive instance normalization techniques for one shot singing voice conversion. The instance normalization (IN) layers within the auto-encoder effectively separated singer and content representations. During conversion, singer representations were transferred using adaptive instance normalization (AdaIN) layers. This singing voice system with the help of activation function prevented the transfer of singer information while conveying the singing content. Additionally, the fusion of LSTM with CNN can enhance voice conversion models by capturing both local and contextual features. The one-shot capability simplified the architecture, utilizing a single encoder and decoder. Impressively, the proposed hybrid CNN-LSTM model achieved remarkable performance without compromising either quality or similarity. The objective and subjective evaluation assessments showed that the proposed hybrid CNN-LSTM model outperformed the baseline architectures. Evaluation results showed a mean opinion score (MOS) of 2.93 for naturalness and 3.35 for melodic similarity. These hybrid CNN-LSTM techniques allowed it to perform high-quality voice conversion with minimal training data, making it a promising solution for various applications.</p></abstract>
APA, Harvard, Vancouver, ISO, and other styles
35

Ehrlich, Benjamin, Liyu Lin, and Jack Jiang. "Concatenation of the Moving Window Technique for Auditory-Perceptual Analysis of Voice Quality." American Journal of Speech-Language Pathology 27, no. 4 (November 21, 2018): 1426–33. http://dx.doi.org/10.1044/2018_ajslp-17-0103.

Full text
Abstract:
Purpose The purpose of this study is to develop a program to concatenate acoustic vowel segments that were selected with the moving window technique, a previously developed technique used to segment and select the least perturbed segment from a sustained vowel segment. The concatenated acoustic segments were compared with the nonconcatenated, short, individual acoustic segments for their ability to differentiate normal and pathological voices. The concatenation process sometimes created a clicking noise or beat, which was also analyzed to determine any confounding effects. Method A program was developed to concatenate the moving window segments. Listeners with no previous rating experience were trained and, then, rated 20 normal and 20 pathological voice segments, both concatenated (2 s) and short (0.2 s) for a total of 80 segments. Listeners evaluated these segments on both the Grade, Roughness, Breathiness, Asthenia, and Strain scale (GRBAS; 8 listeners) and the Consensus Auditory-Perceptual Evaluation of Voice (Kempster, Gerratt, Abbott, Barkmeier-Kraemer, & Hillman, 2009) scale (7 listeners). The sensitivity and specificity of these ratings were analyzed using a receiver-operating characteristic curve. To evaluate if there were increases in particular criteria due to the beat, differences between beat and nonbeat ratings were compared using a 2-tailed analysis of variance. Results Concatenated segments had a higher sensitivity and specificity for distinguishing pathological and normal voices than short segments. Compared with nonbeat segments, the beat had statistically similar increases for all criteria across Consensus Auditory-Perceptual Evaluation of Voice and GRBAS scales, except pitch and loudness. Conclusions The concatenated moving window method showed improved sensitivity and specificity for detecting voice disorders using auditory-perceptual analysis, compared with the short moving window segment. It is a helpful tool for perceptual analytic protocols, allowing for voice evaluation using standardized and automated voice-segmenting procedures. Supplemental Material https://doi.org/10.23641/asha.7100951
APA, Harvard, Vancouver, ISO, and other styles
36

Ma, Estella P.-M., and Edwin M.-L. Yiu. "Voice Activity and Participation Profile." Journal of Speech, Language, and Hearing Research 44, no. 3 (June 2001): 511–24. http://dx.doi.org/10.1044/1092-4388(2001/040).

Full text
Abstract:
Traditional clinical voice evaluation focuses primarily on the severity of voice impairment, with little emphasis on the impact of voice disorders on the individual’s quality of life. This study reports the development of a 28-item assessment tool that evaluates the perception of voice problem, activity limitation, and participation restriction using the International Classification of Impairments, Disabilities and Handicaps-2 Beta-1 concept (World Health Organization, 1997). The questionnaire was administered to 40 subjects with dysphonia and 40 control subjects with normal voices. Results showed that the dysphonic group reported significantly more severe voice problems, limitation in daily voice activities, and restricted participation in these activities than the control group. The study also showed that the perception of a voice problem by the dysphonic subjects correlated positively with the perception of limitation in voice activities and restricted participation. However, the self-perceived voice problem had little correlation with the degree of voice-quality impairment measured acoustically and perceptually by speech pathologists. The data also showed that the aggregate scores of activity limitation and participation restriction were positively correlated, and the extent of activity limitation and participation restriction was similar in all except the job area. These findings highlight the importance of identifying and quantifying the impact of dysphonia on the individual’s quality of life in the clinical management of voice disorders.
APA, Harvard, Vancouver, ISO, and other styles
37

Wang, Xiaochen, and Tao Wang. "Voice Recognition and Evaluation of Vocal Music Based on Neural Network." Computational Intelligence and Neuroscience 2022 (May 20, 2022): 1–9. http://dx.doi.org/10.1155/2022/3466987.

Full text
Abstract:
Artistic voice is the artistic life of professional voice users. In the process of selecting and cultivating artistic performing talents, the evaluation of voice even occupies a very important position. Therefore, an appropriate evaluation of the artistic voice is crucial. With the development of art education, how to scientifically evaluate artistic voice training methods and fairly select artistic voice talents is an urgent need for objective evaluation of artistic voice. The current evaluation methods for artistic voices are time-consuming, laborious, and highly subjective. In the objective evaluation of artistic voice, the selection of evaluation acoustic parameters is very important. Attempt to extract the average energy, average frequency error, and average range error of singing voice by using speech analysis technology as the objective evaluation acoustic parameters, use neural network method to objectively evaluate the singing quality of artistic voice, and compare with the subjective evaluation of senior professional teachers. In this paper, voice analysis technology is used to extract the first formant, third formant, fundamental frequency, sound range, fundamental frequency perturbation, first formant perturbation, third formant perturbation, and average energy of singing acoustic parameters. By using BP neural network methods, the quality of singing was evaluated objectively and compared with the subjective evaluation of senior vocal professional teachers. The results show that the BP neural network method can accurately and objectively evaluate the quality of singing voice by using the evaluation parameters, which is helpful in scientifically guiding the selection and training of artistic voice talents.
APA, Harvard, Vancouver, ISO, and other styles
38

Xanthate Duggirala, Suvarnalata, Michael Schwartze, Therese Van Amelsvoort, David E. J. Linden, Ana Pinheiro, and Sonja Kotz. "M53. EMOTIONAL SELF-VOICE PROCESSING AND ITS RELATIONSHIP WITH HALLUCINATORY PRONENESS." Schizophrenia Bulletin 46, Supplement_1 (April 2020): S154. http://dx.doi.org/10.1093/schbul/sbaa030.365.

Full text
Abstract:
Abstract Background Sensory brain areas typically reduce their activity when we speak, allowing us to differentiate our own from someone else’s speech. Similarly, the amplitude of the N100 component of the EEG event-related potential in response to own speech is smaller than for passive listening to own or someone else’s speech. This amplitude suppression effect seems to be altered in voice hearers, which in turn could result in source misattribution (e.g., self-produced voice attributed to an external source). Emotion in speech can have a comparable effect, altering not only self-voice processing but also differentiation of the quality of auditory hallucinations in clinical and non-clinical voice hearers. For example, unlike in non-clinical voice hearers, auditory hallucinations in clinical voice hearers are usually derogatory in content and negatively affect daily functioning. Recent research strongly suggests that clinical and non-clinical voice hearers lie on a continuum ranging from low to high hallucinatory proneness. Based on this notion, the present study used EEG to investigate the effects of manipulations of self-voice quality in self-generated and passively listened-to self-voice as a function of hallucinatory proneness (HP) in healthy young adults. This is the first EEG study that examined the interplay of sensory suppression, emotion, and HP in a non-clinical population. Methods Participants varying in HP (according to the Launay Slade Hallucination Scale) participated in a standardized button-press task to elicit their own voice (compared to passively listening to it) in which the self-voice changed stepwise from fully neutral to fully emotional. The experimental task comprised three conditions: motor-to-auditory (MA), where the button-press generated the voice, auditory only (AO), where the voice was presented without the button press, and motor only (MO-a control condition to remove the motor related artifacts from the MA condition), where the button press did not generate the voice. Neutral and angry self-voice (single syllable ‘ah’ and ‘oh’ vocalizations of 500 ms duration) were recorded for each participant before the EEG acquisition. These voices were morphed to generate a neutral to angry continuum consisting of five stimuli ranging from fully neutral to fully angry: 100% neutral, 60-40% neutral-angry, 50-50% neutral-angry and 40–60% neutral-angry and 100% angry. Results Preliminary results with 17 participants show a significant effect of emotional self-voice quality on N1 suppression effect, with a larger suppression effect for the 100% angry as compared to 100% neutral self-voice. On the other hand, 60-40% neutral-angry, 50-50% neutral-angry and 40–60% neutral-angry self-voice show an enhancement effect. Furthermore, the results show a significant interaction of HP and voice quality on N1 suppression effect such that high HP showed no N1 suppression effect for the 100% neutral self-voice and an enhanced N1 effect when emotional quality of the self-voice increased. Discussion These data suggest that participants perceive the manipulations in the self-voice quality such that they recognize their own fully neutral and angry voice depicted by N100 suppression effect. Similarly, an N100 enhancement effect for 50-50% neutral-angry voice suggest that it is perceived as the most uncertain or peculiar of all the stimuli. Further, low and high HP show difference in N100 suppression effect for different voices, suggesting that HP may alter self-voice processing and these alterations are enhanced for emotional self-voice. This further supports the fact that abnormal perceptual experiences in voice hearers are higher when auditory hallucinations are emotional in nature.
APA, Harvard, Vancouver, ISO, and other styles
39

Henton, Caroline. "Where is Female Synthetic Speech?" Journal of the International Phonetic Association 29, no. 1 (June 1999): 51–61. http://dx.doi.org/10.1017/s0025100300006411.

Full text
Abstract:
There is widespread, immediate and enduring demand for high quality, natural, intelligible synthetic female voices in the expanding speech technology industry. Yet synthetic female voices are scarce, both in parametric text-to-speech (TTS) systems and in concatenative ones. Current female synthetic speech largely lacks naturalness, pleasantness and tolerability. Some acoustic specifications of female voices that are relevant to synthesis are discussed in detail. Recent research pertaining to female voice quality is reported and a ranking of these various considerations is proposed. This paper reviews the present situation and considers why there is a paucity of female voice synthesis.
APA, Harvard, Vancouver, ISO, and other styles
40

Zraick, Richard I., Julie M. Liss, Michael F. Dorman, James L. Case, Leonard L. LaPointe, and Stephen P. Beals. "Multidimensional Scaling of Nasal Voice Quality." Journal of Speech, Language, and Hearing Research 43, no. 4 (August 2000): 989–96. http://dx.doi.org/10.1044/jslhr.4304.989.

Full text
Abstract:
Listeners judged the dissimilarity of pairs of synthesized nasal voices that varied on 3 dimensions. Separate nonmetric multidimensional scaling (MDS) solutions were calculated for each listener and the group. Similar 3-dimensional solutions were derived for the group and each of the listeners, with the group MDS solution accounting for 83% of the total variance in listeners' judgments. Dimension 1 ("Nasality") accounted for 54% of the variance, Dimension 2 ("Loudness") for 18% of the variance, and Dimension 3 ("Pitch") for 11% of the variance. The 3 dimensions were significantly and positively correlated with objective measures of nasalization, intensity, and fundamental frequency. The results of this experiment are discussed in relation to other MDS studies of voice perception, and there is a discussion of methodological issues for future research.
APA, Harvard, Vancouver, ISO, and other styles
41

Ugulino, Ana Celiane, Gisele Oliveira, and Mara Behlau. "Perceived dysphonia by the clinician's and patient's viewpoint." Jornal da Sociedade Brasileira de Fonoaudiologia 24, no. 2 (2012): 113–18. http://dx.doi.org/10.1590/s2179-64912012000200004.

Full text
Abstract:
PURPOSE: To verify the relationship between the clinician's vocal evaluation and vocal self-assessment and voice-related quality of life. METHODS: Participants were 96 individuals: 48 with vocal complaints and voice deviation (VCG), mean age of 51 years, with diagnosis and indication of voice therapy; and 48 with no vocal complaints and healthy voices (NVCG), mean age of 46 years. All participants answered the Voice-Related Quality of Life (V-RQOL) questionnaire, performed a vocal self-assessment and were submitted to auditory-perceptual analysis of voice. RESULTS: Mean V-RQOL scores were different between groups for all domains. Self-assessment results also showed differences between groups, which was not the case in the auditory-perceptual analysis of sustained vowel and connected speech, showing that the patient's perception was worse than the clinician's. There was correlation between the V-RQOL domains (Socio-emotional and Physical: 76.8%; Socio-emotional and Total: 90.8%; Physical and Total: 95.8%), as well as between the Socio-emotional (-52.9%), Physical (-43.1%) and Total (-52.2%) domains and the self-assessment. However, no correlation was found between auditory-perceptual analysis and self-assessment measures, except for a weak correlation between vocal self-assessment and auditory-perceptual analysis of the sustained vowel (33.3%). CONCLUSION: The clinician's perception does correspond to the individual's self-perception of his/her vocal quality and the impact of a voice deviation on his/her quality of life, but not directly. The individual's perception about his/her vocal quality and voice-related quality of life complements the clinician's perception regarding the overall degree of the voice deviation.
APA, Harvard, Vancouver, ISO, and other styles
42

Loakes, Debbie, and Adele Gregory. "Voice quality in Australian English." JASA Express Letters 2, no. 8 (August 2022): 085201. http://dx.doi.org/10.1121/10.0012994.

Full text
Abstract:
This study is an acoustic investigation of voice quality in Australian English. The speech of 33 Indigenous Australians (Aboriginal English speakers) is compared to that of 28 Anglo Australians [Mainstream Australian English (MAE) speakers] from two rural locations in Victoria. Analysis of F0 and H1*-H2* reveals that pitch and voice quality differ significantly for male speakers according to dialect and for female speakers according to location. This study highlights previously undescribed phonetic and sociophonetic variability in voice quality in Australian English.
APA, Harvard, Vancouver, ISO, and other styles
43

Tveterås, Gry. "Perceptual analysis of voice quality." Scandinavian Journal of Logopedics and Phoniatrics 17, no. 3-4 (January 1992): 145–49. http://dx.doi.org/10.3109/14015439209098732.

Full text
APA, Harvard, Vancouver, ISO, and other styles
44

Chen, Pin-Hsuan, and Che-Nan Yang. "Voice Quality Measurement in perfSONAR." Proceedings of the Asia-Pacific Advanced Network 30 (December 30, 2010): 104. http://dx.doi.org/10.7125/apan.30.15.

Full text
APA, Harvard, Vancouver, ISO, and other styles
45

Kreiman, Jody, Bruce R. Gerratt, and Theodore S. Bell. "Variability of voice quality ratings." Journal of the Acoustical Society of America 100, no. 4 (October 1996): 2828. http://dx.doi.org/10.1121/1.416658.

Full text
APA, Harvard, Vancouver, ISO, and other styles
46

Kreiman, Jody, and Bruce R. Gerratt. "Voice quality ratings: Theoretical considerations." Journal of the Acoustical Society of America 90, no. 4 (October 1991): 2328. http://dx.doi.org/10.1121/1.402232.

Full text
APA, Harvard, Vancouver, ISO, and other styles
47

Becker, Kara, Sameer ud Dowla Khan, and Lal Zimman. "Voice quality variation and gender." Journal of the Acoustical Society of America 136, no. 4 (October 2014): 2295. http://dx.doi.org/10.1121/1.4900303.

Full text
APA, Harvard, Vancouver, ISO, and other styles
48

Hillel, Alexander T., Selmin Karatayli-Ozgursoy, James R. Benke, Simon Best, Paulette Pacheco-Lopez, Kristine Teets, Heather Starmer, and Lee M. Akst. "Voice Quality in Laryngotracheal Stenosis." Annals of Otology, Rhinology & Laryngology 124, no. 5 (December 17, 2014): 413–18. http://dx.doi.org/10.1177/0003489414564249.

Full text
APA, Harvard, Vancouver, ISO, and other styles
49

Garellek, Marc. "Voice quality strengthening and glottalization." Journal of Phonetics 45 (July 2014): 106–13. http://dx.doi.org/10.1016/j.wocn.2014.04.001.

Full text
APA, Harvard, Vancouver, ISO, and other styles
50

Pershall, Kim E., and Daniel R. Boone. "Supraglottic contribution to voice quality." Journal of Voice 1, no. 2 (January 1987): 186–90. http://dx.doi.org/10.1016/s0892-1997(87)80044-9.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography