To see the other types of publications on this topic, follow the link: And impressive and expressive speech.

Dissertations / Theses on the topic 'And impressive and expressive speech'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'And impressive and expressive speech.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Jauk, Igor. "Unsupervised learning for expressive speech synthesis." Doctoral thesis, Universitat Politècnica de Catalunya, 2017. http://hdl.handle.net/10803/460814.

Full text
Abstract:
Nowadays, especially with the upswing of neural networks, speech synthesis is almost totally data driven. The goal of this thesis is to provide methods for automatic and unsupervised learning from data for expressive speech synthesis. In comparison to "ordinary" synthesis systems, it is more difficult to find reliable expressive training data, despite huge availability on sources like Internet. The main difficulty consists in the highly speaker- and situation-dependent nature of expressiveness, causing many and acoustically substantial variations. The consequences are, first, it is very difficult to define labels which reliably identify expressive speech with all nuances. The typical definition of 6 basic emotions, or alike, is a simplification which will have inexcusable consequences dealing with data outside the lab. Second, even if a label set is defined, apart of the enormous manual effort, it is difficult to gain sufficient training data for the models respecting all the nuances and variations. The goal of this thesis is to study automatic training methods for expressive speech synthesis avoiding labeling and to develop applications from these proposals. The focus lies on the acoustic and the semantic domains. For the part of the acoustic domain, the goal is to find suitable acoustic features to represent expressive speech, especially for the multi-speaker domain, as getting closer to real-life uncontrolled data. For this, the perspective will slide away from traditional, mainly prosody-based, features towards features gained with factor analysis, trying to identify the principal components of the expressiveness, namely using i-vectors. Results show that a combination of traditional and i-vector based features performs better in unsupervised clustering of expressive speech than traditional features and even better than large state-of-the-art sets in the multi-speaker domain. Once the feature set is defined, it is used for unsupervised clustering of an audiobook, where from each cluster a voice is trained. Then, the method is evaluated in an audiobook-editing application, where users can use the synthetic voices to create their own dialogues. The obtained results validate the proposal. In this editing application users choose synthetic voices and assign them to the sentences considering the speaking characters and the expressiveness. Involving the semantic domain, this assignment can be achieved automatically, at least partly. Words and sentences are represented numerically in trainable semantic vector spaces, called embeddings, and these can be used to predict the expressiveness to some extent. This method not only permits fully automatic reading of larger text passages, considering the local context, but can also be used as a semantic search engine for training data. Both applications are evaluated in a perceptual test showing the potential of the proposed method. Finally, accounting for the new tendencies in the speech synthesis world, deep neural network based expressive speech synthesis is designed and tested. Emotionally motivated semantic representations of text, sentiment embeddings, trained on the positiveness and the negativeness of movie reviews, are used as an additional input to the system. The neural network now learns not only from segmental and contextual information, but also from the sentiment embeddings, affecting especially prosody. The system is evaluated in two perceptual experiments which show preferences for the inclusion of sentiment embeddings as an additional input.<br>Hoy en día, especialmente con el auge de las redes neuronales, la síntesis de habla se basa casi totalmente en datos. El objetivo de esta tesis es proveer métodos de entrenamiento automático y no supervisado a partir de datos para la síntesis de habla expresiva. En comparación con sistemas de síntesis "neutrales", resulta más difícil encontrar datos de entrenamiento fiables para la síntesis expresiva, a pesar de la gran disponibilidad de recursos como internet. La dificultad principal se origina en la naturaleza del habla expresiva, altamente dependiente del hablante y la situación, resultando en muchas variaciones acústicas. Las consecuencias son, primero, que es muy difícil definir etiquetas que identifiquen fiablemente todos los detalles del habla expresiva. La definición típica de 6 emociones básicas es una simplificación que tendrá consecuencias inexcusables cuando se trata con datos fuera del laboratorio. Segundo, incluso si se llegara a definir un conjunto de etiquetas, aparte del enorme esfuerzo manual que supondría, sería muy difícil conseguir suficientes datos de entrenamiento para cada variante respetando todos sus matices. El objetivo de esta tesis es estudiar métodos de entrenamiento automático para la síntesis de habla expresiva evitando etiquetas y desarrollar aplicaciones a base de estas propuestas. El enfoque abarca los dominios acústico y semántico. Con respecto al dominio acústico, el objetivo es encontrar características acústicas aptas para representar habla expresiva, especialmente en el dominio multi-locutor, acercándose a datos reales e incontrolados. Para esto, la perspectiva se apartará de las características tradicionales, principalmente basadas en la prosodia, hacia características ganadas a partir del análisis de factores, intentando identificar los componentes principales de la expresividad, concretamente los i-vectors. Los resultados demuestran que una combinación de características tradicionales y de las basadas en los i-vectors rinde mejor en la tarea del "clustering" no supervisado del habla expresiva que solo las características tradicionales e incluso mejor que amplios conjuntos de características del estado del arte en el dominio multi-locutor. Una vez definido, el conjunto de características se utiliza para el "clustering" no supervisado de un audiolibro, entrenando de cada "cluster" una voz. El método se ha evaluado en una aplicación de edición de audiolibro, donde los usuarios utilizaban las voces sintéticas para crear sus propios diálogos. Los resultados obtenidos validan la propuesta. En la aplicación de edición, los usuarios eligen voces sintéticas y las asignan a frases considerando los personajes y la expresividad. Implicando el dominio semántico, esta asignación podría realizarse automáticamente. En esta parte de la tesis, palabras y frases se representan numéricamente en espacios vectoriales entrenables, llamados embeddings, y pueden utilizarse para predecir la expresividad. Este método no solo permite una lectura automática de pasajes de texto, tomando en cuenta el contexto local, sino que también puede utilizarse como una herramienta de búsqueda semántica para datos de entrenamiento. Ambas aplicaciones se han evaluado en un experimento perceptual demostrando el potencial de la metodología propuesta. Finalmente, siguiendo las nuevas tendencias en el mundo de la síntesis de habla basada en redes neuronales, se ha desarrollado y evaluado un sistema de síntesis de voz expresiva utilizando esta tecnología. Representaciones semánticas de texto, motivadas emocionalmente, llamadas "sentiment embeddings", entrenadas con reseñas de cine, se utilizan como input adicional en el sistema. La red neuronal ahora aprende no solamente de la información segmental y contextual, sino también de esta representación del sentimiento, afectando especialmente la prosodia. El sistema se ha evaluado en dos experimentos perceptuales, demostrando la preferencia del sistema que incluye esta nueva represent
APA, Harvard, Vancouver, ISO, and other styles
2

Shaw, Felix. "Expressive modulation of neutral visual speech." Thesis, University of East Anglia, 2015. https://ueaeprints.uea.ac.uk/56812/.

Full text
Abstract:
The need for animated graphical models of the human face is commonplace in the movies, video games and television industries, appearing in everything from low budget advertisements and free mobile apps, to Hollywood blockbusters costing hundreds of millions of dollars. Generative statistical models of animation attempt to address some of the drawbacks of industry standard practices such as labour intensity and creative inflexibility. This work describes one such method for transforming speech animation curves between different expressive styles. Beginning with the assumption that expressive speech animation is a mix of two components, a high-frequency speech component (the content) and a much lower-frequency expressive component (the style), we use Independent Component Analysis (ICA) to identify and manipulate these components independently of one another. Next we learn how the energy for different speaking styles is distributed in terms of the low-dimensional independent components model. Transforming the speaking style involves projecting new animation curves into the lowdimensional ICA space, redistributing the energy in the independent components, and finally reconstructing the animation curves by inverting the projection. We show that a single ICA model can be used for separating multiple expressive styles into their component parts. Subjective evaluations show that viewers can reliably identify the expressive style generated using our approach, and that they have difficulty in identifying transformed animated expressive speech from the equivalent ground-truth.
APA, Harvard, Vancouver, ISO, and other styles
3

Qader, Raheel. "Pronunciation and disfluency modeling for expressive speech synthesis." Thesis, Rennes 1, 2017. http://www.theses.fr/2017REN1S076/document.

Full text
Abstract:
Dans la première partie de cette thèse, nous présentons une nouvelle méthode de production de variantes de prononciations qui adapte des prononciations standards, c'est-à-dire issues d'un dictionnaire, à un style spontané. Cette méthode utilise une vaste gamme d'informations linguistiques, articulatoires et acoustiques, ainsi qu'un cadre probabiliste d'apprentissage automatique, à savoir les champs aléatoires conditionnels (CAC) et les modèles de langage. Nos expériences poussées sur le corpus Buckeye démontrent l'efficacité de l'approche à travers des évaluations objectives et perceptives. Des tests d'écoutes sur de la parole synthétisée montrent que les prononciations adaptées sont jugées plus spontanées que les prononciations standards, et même que celle réalisées par les locuteurs du corpus étudié. Par ailleurs, nous montrons que notre méthode peut être étendue à d'autres tâches d'adaptation, par exemple pour résoudre des problèmes d'incohérences entre les différentes séquences de phonèmes manipulées par un système de synthèse. La seconde partie de la thèse explore une nouvelle approche de production automatique de disfluences dans les énoncés en entrée d'un système de synthèse de la parole. L'approche proposée offre l'avantage de considérer plusieurs types de disfluences, à savoir des pauses, des répétitions et des révisions. Pour cela, nous présentons une formalisation novatrice du processus de production de disfluences à travers un mécanisme de composition de ces disfluences. Nous présentons une première implémentation de notre processus, elle aussi fondée sur des CAC et des modèles de langage, puis conduisons des évaluations objectives et perceptives. Celles-ci nous permettent de conclure à la bonne fonctionnalité de notre proposition et d'en discuter les pistes principales d'amélioration<br>In numerous domains, the usage of synthetic speech is conditioned upon the ability of speech synthesis systems to generate natural and expressive speech. In this frame, we address the problem of expressivity in TTS by incorporating two phenomena with a high impact on speech: pronunciation variants and speech disfluencies. In the first part of this thesis, we present a new pronunciation variant generation method which works by adapting standard i.e., dictionary-based, pronunciations to a spontaneous style. Its strength and originality lie in exploiting a wide range of linguistic, articulatory and acoustic features and to use a probabilistic machine learning framework, namely conditional random fields (CRFs) and language models. Extensive experiments on the Buckeye corpus demonstrate the effectiveness of this approach through objective and subjective evaluations. Listening tests on synthetic speech show that adapted pronunciations are judged as more spontaneous than standard ones, as well as those realized by real speakers. Furthermore, we show that the method can be extended to other adaptation tasks, for instance, to solve the problem of inconsistency between phoneme sequences handled in TTS systems. The second part of this thesis explores a novel approach to automatic generation of speech disfluencies for TTS. Speech disfluencies are one of the most pervasive phenomena in spontaneous speech, therefore being able to automatically generate them is crucial to have more expressive synthetic speech. The proposed approach provides the advantage of generating several types of disfluencies: pauses, repetitions and revisions. To achieve this task, we formalize the problem as a theoretical process, where transformation functions are iteratively composed. We present a first implementation of the proposed process using CRFs and language models, before conducting objective and perceptual evaluations. These experiments lead to the conclusion that our proposition is effective to generate disfluencies, and highlights perspectives for future improvements
APA, Harvard, Vancouver, ISO, and other styles
4

Somasundaram, Arunachalam. "A facial animation model for expressive audio-visual speech." Columbus, Ohio : Ohio State University, 2006. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=osu1148973645.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Chandra, Nishant. "HYBRID CONCATENATED-FORMANT EXPRESSIVE SPEECH SYNTHESIZER FOR KINESENSIC VOICES." MSSTATE, 2009. http://sun.library.msstate.edu/ETD-db/theses/available/etd-11092009-165602/.

Full text
Abstract:
Traditional and commercial speech synthesizers are incapable of synthesizing speech with proper emotion or prosody. Conveying prosody in artificially synthesized speech is difficult because of extreme variability in human speech. An arbitrary natural language sentence can have different meanings, depending upon the speaker, speaking style, context, and many other factors. Most concatenated speech synthesizers use phonemes, which are phonetic units defined by the International Phonetic Alphabet (IPA). The 50 phonemes in English are standardized and unique units of sound, but not expression. An earlier work proposed the analogy between speech and music ? ?speech is music, music is speech.? The speech data obtained from the master practitioners, who are trained in kinesensic voice, is marked on a five level intonation scale, which is similar to the music scale. From this speech data, 1324 unique expressive units, called expressemes®, are identified. The expressemes consist of melody and rhythm, which, in digital signal processing, is analogous to pitch, duration and energy of the signal. The expressemes have less acoustic and phonetic variability than phonemes, so they better convey the prosody. The goal is to develop a speech synthesizer which exploits the prosodic content of expressemes in order to synthesize expressive speech, with a small speech database. To create a reasonably small database that captures multiple expressions is a challenge because there may not be a complete set of speech segments available to create an emotion. Methods are suggested whereby acoustic mathematical modeling is used to create missing prosodic speech segments from the base prosody unit. New concatenated-formant hybrid speech synthesizer architecture is developed for this purpose. A pitch-synchronous time-varying frequency-warped wavelet transform based prosody manipulation algorithm is developed for transformation between prosodies. A time-varying frequency-warping transform is developed to smoothly concatenate the temporal and spectral parameters of adjacent expressemes to create intelligible speech. Additionally, issues specific to expressive speech synthesis using expressemes are resolved for example, Ergodic Hidden Markov Model based expresseme segmentation, model creation for F0 and segment duration, and target and join cost calculation. The performance of the hybrid synthesizer is measured against a commercially available synthesizer using objective and perceptual evaluations. Subjects consistently rated the hybrid synthesizer better in five different perceptual tests. 70% of speakers rated the hybrid synthesis as more expressive, and 72% preferred it over the commercial synthesizer. The hybrid synthesizer also got a comparable mean opinion score.
APA, Harvard, Vancouver, ISO, and other styles
6

RODRIGUES, PAULA SALGADO LUCENA. "EXPRESSIVE TALKING HEADS: EXPRESSIVE TALKING HEADS: A STUDY ON SPEECH AND FACIAL EXPRESSION IN VIRTUAL CHARACTERS." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2002. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=6525@1.

Full text
Abstract:
CONSELHO NACIONAL DE DESENVOLVIMENTO CIENTÍFICO E TECNOLÓGICO<br>FUNDAÇÃO PADRE LEONEL FRANCA<br>A face humana é interessante e desafiadora acima de tudo pela sua familiaridade. Essencialmente, é a parte do corpo utilizada para reconhecer indivíduos. Assim como a face, a fala é um importante instrumento na forma de comunicação do ser humano. Através da fala é possível externar pensamentos e, muitas vezes, ela indica o estado de ânimo em que uma pessoa se encontra. Juntos, fala e face são os principais elementos de interatividade entre os seres humanos. Contudo, reproduzir com naturalidade e fidelidade as peculiaridades destes dois elementos no universo computacional não é uma tarefa simples, constituindo-se em tópicos de pesquisa em diversas áreas, em particular na animação facial. Entre os diversos tipos de sistemas de animação facial, destacam-se como diretamente relacionados a este trabalho aqueles que envolvem a sincronização da fala de um personagem com a animação da sua face. Sistemas desse tipo são conhecidos como talking head ou talking face. Para o desenvolvimento de um sistema talking head, é necessário identificar as possíveis abordagens para a modelagem dos dois elementos básicos: fala e face. Os modelos utilizados irão influenciar não apenas a maneira como a animação é conduzida, mas a própria forma de interatividade do sistema. Uma contribuição importante deste trabalho é o estudo das possíveis abordagens e a proposta de uma taxonomia para a classificação de sistemas talking head. A partir da taxonomia proposta e fazendo uso de uma determinada abordagem para cada parâmetro analisado, foi desenvolvida uma aplicação que recebe como entrada um texto contendo a fala e anotações de expressividade e gera como saída, em tempo real, a animação de um personagem virtual enunciando o texto de entrada com o áudio e os movimentos faciais sincronizados. O sistema desenvolvido, denominado Expressive Talking Heads, explora a naturalidade da animação facial e ao mesmo tempo busca oferecer ao usuário uma interface com interatividade em tempo real. O Expressive Talking Heads pode ser executado tanto no modo isolado (stand alone) como acoplado a navegadores para a web, tendo sido projetado e desenvolvido com a preocupação de oferecer uma solução independente da plataforma e do sistema operacional utilizados.<br>The human face is interesting and challenging mainly because of its familiarity. Essentially, it is the part of the human body that is used to recognize individuals. As well as the face, the speech is an important instrument for human communication, allowing the exteriorization of thoughts and the definition of emotions. Together, speech and face are the main elements of interactivity among human beings. However, the natural and faithful reproductions of the pecularities of these elements in the computational universe is not a simple task, constituting topics of the research in the diverse areas, particularly in facial animation. Among the diverse types of facial animation systems developed, those that involve the facial animation of the virtual character combined with speech synchronization are distinguished as directly related to this work. These kinds of systems are known as talking head or talking face. Fot the development of a talking head system, it is necessary to identify the possible approaches for the speech and face modeling. The models used will influence not only the way that the animation is performed, but it will also affect the system´s interactivity. An important contribution of the present master thesis is the study of several possible approaches for the main elements and the proposal of taxonomy for the classification of the talking head systems. From the proposed taxonomy and making use of one approach for each analyzed paramenter, an application was developed that receives as input a text composed by the character´s speech and genus, language and emotion parameters, and it generates as output, in real time, the animation of a virtual character uttering the input text with speech synchronization and expressiveness. The system developed, called Expressive Talking Heads, explores the naturalness of facial animation and it seeks to offer the user a real- time interactivity interface. The Expressive Talking Heads system can run as a stand-alone applicattion or connected to web browsers. It was designed and developed to provide a platform and operating system independent solution.
APA, Harvard, Vancouver, ISO, and other styles
7

Costa, Paula Dornhofer Paro 1978. "Two-dimensional expressive speech animation = Animação 2D de fala expressiva." [s.n.], 2015. http://repositorio.unicamp.br/jspui/handle/REPOSIP/260903.

Full text
Abstract:
Orientador: José Mario De Martino<br>Tese (doutorado) - Universidade Estadual de Campinas, Faculdade de Engenharia Elétrica e de Computação<br>Made available in DSpace on 2018-08-26T21:43:57Z (GMT). No. of bitstreams: 1 Costa_PaulaDornhoferParo_D.pdf: 15894797 bytes, checksum: 194a20ae502dfc7198a008d576e23e4c (MD5) Previous issue date: 2015<br>Resumo: O desenvolvimento da tecnologia de animação facial busca atender uma demanda crescente por aplicações envolvendo assistentes, vendedores, tutores e apresentadores de notícias virtuais; personagens realistas de videogames, agentes sociais e ferramentas para experimentos científicos em psicologia e ciências comportamentais. Um aspecto relevante e desafiador no desenvolvimento de cabeças falantes, ou "talking heads", é a reprodução realista dos movimentos articulatórios da fala combinados aos elementos de comunicação não-verbal e de expressão de emoções. Este trabalho presenta uma metodologia de síntese de animação facial baseada em imagens, ou animação facial 2D, que permite a reprodução de uma ampla gama de estados emocionais de fala expressiva, além de suportar a modulação de movimentos da cabeça e o controle de elementos faciais tais como o piscar de olhos e o arqueamento de sobrancelhas. A síntese da animação utiliza uma base de imagens-protótipo que são processadas para obtenção dos quadros-chave da animação. Os pesos utilizados para a combinação das imagens-protótipo são derivados de um modelo estatístico de aparência e formas, construído a partir de um conjunto de imagens de treinamento extraídas de um corpus audiovisual de uma face real. A síntese das poses-chave é guiada pela transcrição fonética temporizada da fala a ser animada e pela informação do estado emocional almejado. As poses-chave representam visemas dependentes de contexto fonético que implicitamente modelam os efeitos da coarticulação na fala visual. A transição entre poses-chave adjacentes é realizada por um algoritmo de metamorfose não-linear entre imagens. As animações sintetizadas aplicando-se a metodologia proposta foram avaliadas por meio de avaliação perceptual de reconhecimento de emoções. Dentre as contribuições deste trabalho encontra-se a construção de uma base de dados de vídeo e captura de movimento para fala expressiva em português do Brasil<br>Abstract: The facial animation technology experiences an increasing demand for applications involving virtual assistants, sellers, tutors and newscasters; lifelike game characters, social agents, and tools for scientific experiments in psychology and behavioral sciences. A relevant and challenging aspect of the development of talking heads is the realistic reproduction of the speech articulatory movements combined with the elements of non-verbal communication and the expression of emotions. This work presents an image-based, or 2D, facial animation synthesis methodology that allows the reproduction of a wide range of expressive speech emotional states and also supports the modulation of head movements and the control of face elements, like the blinking of the eyes and the raising of the eyebrows. The synthesis of the animation uses a database of prototype images which are combined to produce animation keyframes. The weights used for combining the prototype images are derived from a statistical active appearance model (AAM), which is built from a set of sample images extracted from an audio-visual corpus of a real face. The generation of the animation keyframes is driven by the timed phonetic transcription of the speech to be animated and the desired emotional state. The keyposes consist of expressive context-dependent visemes that implicitly model the speech coarticulation effects. The transition between adjacent keyposes is performed through a non-linear image morphing algorithm. To evaluate the synthesized animations, a perceptual evaluation based on the recognition of emotions was performed. Among the contributions of the work is also the building of a database of expressive speech video and motion capture data for Brazilian Portuguese<br>Doutorado<br>Engenharia de Computação<br>Doutora em Engenharia Elétrica
APA, Harvard, Vancouver, ISO, and other styles
8

Midford, Nicole Anne. "Expressive Communication and Socialization Skills of Five-Year Olds with Slow Expressive Language Development." PDXScholar, 1993. https://pdxscholar.library.pdx.edu/open_access_etds/4612.

Full text
Abstract:
Beginning at birth, a child's receptive and expressive language skills are developing in stages. Likewise, the child's socialization skills are progressing in stages. However, it does not seem that communication and socialization are developing independently of each other. Rather, it seems that their development is interrelated.Children learn to speak in a social context, and social situations are necessary for the development of a variety of language structure~ On the same note, in order for those language structures to develop normally, it is necessary for the child to participate in different social situations. Social interactionists have theorized for some time that human language develops out of the social-communicative functions that language serves in human relations. Vygotsky (1962) theorized that language development, social development, and cognitive development all overlap. He stated that a child's social means of thought is language and referred to this as "verbal thought." This verbal thought process serves a major social function. It is through this verbal thought process that children have the ability to be socialized by others and to socialize with others: If, in fact,Expressive language skills and socialization skills do develop together, it would then seem logical that the child who is late to begin talking would also experience initial deficits in the development of socialization Subsequently, it would seem that the late-talking child (L T) who has persistent deficits in language would, in turn, maintain chronic deficits in socialization. Results of a study which set out to investigate the differences between two and three-year old subjects with a history of LT and their normal language peers indicated that subjects with a history of LT are, in fact, at risk for persistent delays in both expressive language and socialization (Paul, Spangle Looney, and Dahm, 1991). The purpose of this study was to compare the language and socialization skills of a group of five-year olds with a history of LT to a group of normal subjects of the same age. If significant differences were found between the two groups in either area, the scores of the subjects with a history of LT at age two would be correlated with their scores at age five to investigate whether a significant relationship existed between their scores at both ages. It was hypothesized that the subjects with a history of LTwould be at risk for longterm delays in both language and socialization. More specifically, the group of subjects with a history of LT, as a whole, would show significant delays in the areas of expressive language and socialization as compared to the normal controls. It was further hypothesized that the subjects with a history of LTs' scores at the age of two would reliably predict their scores at five, given a significant deficit in either area. The Vineland Adaptive Behavior Scales VABS (Sparrow, Balla, & Cicchetti, 1984) was the test instrument used to gather the data at both age levels, five years and two years. Parents of 25 subjects with a history of L T and 25 normal subjects were interviewed by a trained graduate researcher on their child's communication, daily living and socialization skills using the VABS. Results of an ANOVA and Tukey multiple comparisons indicated that the subjects with a history of LT, as a whole, scored significantly lower than the normal subjects in the areas of expressive communication and socialization at age five. Since a proportion of the test items in the socialization domain of the VABS require the child to verbalize, an item analysis between the verbal and the nonverbal test items was performed to determine the influence of the verbal test items on the subjects with a history of LTs' socialization scores. Results of the item-analysis indicated that the subjects with a history of L T's poor performance on the socialization scale was due to their deficits in social skills not their deficits in expressive language. Lastly, a Pearson Product Moment Correlational Test was conducted to investigate the relationship between the subjects with a history of LTs' scores at age two on the communication and the socialization scales and their scores at age five on the same scales. Results indicated that the subjects with a history of LTs' scores on both the socialization scale and the communication scale at age two correlated significantly with their scores on the socialization scale at age five. Therefore, the subjects with a history of LTs' socialization and communication scores at age two are good predictors of their adaptive social skills at the age of five.
APA, Harvard, Vancouver, ISO, and other styles
9

Unkefer, Carol Lynn. "Familiality of Early Expressive Language Delay: A Sibling Study." PDXScholar, 1994. https://pdxscholar.library.pdx.edu/open_access_etds/5162.

Full text
Abstract:
Researchers are seeking more information . on how and why language disorders tend to run in families, particularly siblings of language disordered children. The purpose of this study was to examine the prevalence of language and related disorders in the siblings of two groups of children: those with slow expressive language development (SELD) and those with a normal language history. This study sought to answer the following questions: 1) Is there a significant difference in prevalence of language problems in two groups of children: those with SELD and those with a normal language history?, and 2) Is there a greater probability of language problems in the siblings of children in the SELD group who have receptive/expressive language disorders when compared to those SELD children with pure expressive language deficits or to those with a normal language history? Subjects were 45 7-and 8-year old children participating in a longitudinal study at Portland State University. The children were divided into two groups, normal and SELD, based on test scores administered at intake to the original study. In order to look at the effect of a receptive component on heritability of language disorders, the SELD group was subgrouped into pure expressive language disorders and receptive/expressive language disorders based on tests administered at intake. A family history questionnaire was the method of data collection, asking parents to report on ten areas of language and related disorders in the siblings of subjects. Results of one-sided z-tests and a chi-square test were computed and consistently found a highly significant difference between groups, with families of SELD subjects more likely to report a history of language problems over the normal group. These results are consistent with previous research in showing the heritability of language disorders. Results may also indicate that a receptive language component is associated with heritability of specific language disorders among children.
APA, Harvard, Vancouver, ISO, and other styles
10

Dunst, Carl J., A. Lynn Williams, Carol M. Trivette, Andrew Simkus, and Deborah W. Hamby. "Relationships Between Inferential Reading Language Strategies and Young Children’s Comprehension and Expressive Language Competencies." Digital Commons @ East Tennessee State University, 2012. https://dc.etsu.edu/etsu-works/2010.

Full text
Abstract:
The relationships between different types of adult-support inferential book reading strategies and young children’s language and literacy competence were examined in 18 studies that included 1134 study participants. van Kleeck’s (2006) descriptions of two levels of inferencing and different types of inferential strategies at each level were used to code and analyze the patterns of correlations between the book reading strategies and the child outcomes. Results showed that parents’ and teachers’ use of different types of inferencing strategies were related to variations in the child outcomes, and that the effects of inferencing were conditioned on the children’s ages. Implications for practice are described.
APA, Harvard, Vancouver, ISO, and other styles
11

Marriott, Jane Elizabeth Clare. "Campaign finance regulation and expressive rights : a comparative study." Thesis, Brunel University, 2000. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.324780.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Giacherro, Traci Lee. "Effects of Receptive Language Deficits on Persisting Expressive Language Delays." PDXScholar, 1995. https://pdxscholar.library.pdx.edu/open_access_etds/4949.

Full text
Abstract:
Predicting language outcomes in children who at age two are "late talkers" is a concern of Speech Language Pathologists. Currently, there is no conclusive data allowing specialists to predict which children will outgrow their delays and which children will not. The purpose of the present study is to analyze the effect of a receptive language delay on the outcome of the slow expressive language delayed child, and determine whether or not it is a viable predictor of poor outcomes. The subject information used in this project was compiled from the data collected and reported by Paul (1991) during the Portland Language Development Project (PLDP). Children in the PLDP first participated in the longitudinal study between the ages of twenty to thirtyfour months. They were categorized as being slow in expressive language development if they produced fewer that fifty intelligible words during this age range. They were then subgrouped into an expressive-receptive delayed group if they scored more than one standard deviation below the mean on the Reynell Developmental Language Scales. Of the twenty-five subjects with complete data over the five years of the study, nineteen were considered to be solely expressively delayed, while the remaining six were classified as having both an expressive and a receptive language delay. Lee's Developmental Sentence Scoring (DSS) (1974) was used to track the subject's expressive language abilities to the age of seven. DSS scores were analyzed yearly, using the Mann-Whitney nonparametric statistical test. This would determine whether the subjects considered to be both expressively and receptively delayed were exhibiting more difficulties in their expressive language abilities than those subjects with expressive delays alone. The results of the study indicated that significant differences did not exist between the two groups. Therefore, there was insufficient evidence to conclude that a receptive language delay at twenty to thirty-four months of age is a feasible predictor of lasting expressive language delays. This leads to the recommendation that additional research be conducted focusing on areas other than receptive language abilities as being predictors of poor expressive language outcomes.
APA, Harvard, Vancouver, ISO, and other styles
13

Warren, Deborah Kay. "Nonlinguistic Cognitive Performance and Expressive and Receptive Language Scores in Children with Expressive Language Delay." PDXScholar, 1994. https://pdxscholar.library.pdx.edu/open_access_etds/4884.

Full text
Abstract:
This study was part of the Portland Language Development Project. The purpose was to establish reliability for the Goodenough Draw-A-Man Test. Additionally, nonlinguistic cognitive performance scores were correlated with soores from expressive and receptive language test soores. Finally, scores of overall cognitive function and of nonlinguistic cognitive function in children with normally developing language (NL) and with expressive language delay (ELD) were compared. The original group size was 60 children, 30 with ELD at the age of 20 months, and 30 who were a matched control group. These subjects were reevaluated during Kindergarten. The Draw-A-Man Test was administered to assess the subjects' nonlinguistic cognitive functioning. The McCarthy Scales of Children's Abilities CMCSA) was administered to assess the subjects' overall cognitive functioning. A free speech sample was analyzed using the Developmental Sentence Scoring (DSS) criteria to assess expressive language skills, and the Vineland Adaptive Behavior Scales
APA, Harvard, Vancouver, ISO, and other styles
14

Hare-Blye, Cynthia Lee. "Gender Differences in Slow Expressive Language Development." PDXScholar, 1994. https://pdxscholar.library.pdx.edu/open_access_etds/4854.

Full text
Abstract:
The contemporary research suggests that some children who present with early language delays as toddlers outgrow their delays while others continue to develop long-term language difficulties. Several studies over the years have focused on factors that might aid in predicting the outcome of late talkers. This current study emphasized exploring gender as a possible predictive factor. The purpose of this study was to determine if significant differences exist in the rate of growth in language skills, as indexed by scores on the Developmental Sentence Scoring (DSS) procedure (Lee, 1974) of boys versus girls who are late to start talking as toddlers. The research hypothesis was that boys who present as LT toddlers would score significantly higher than LT girls at each age level tested. The DSS is a norm-referenced instrument that assesses age-appropriate morphological development and syntax. The LT subjects used were part of the Portland Language Development Project, a longitudinal study. Spontaneous speech samples were collected, transcribed, and analyzed using the DSS procedure once each year from the time they were approximately 3 years of age, until the age of 7. Late talking children in this present study were grouped by gender. A Chi Square test was used to determine if the proportion of males scoring above the 10th percentile on the DSS was significantly different than the proportion of females scoring above the 10th percentile at each age. Results from this analysis indicated that at the age of 3 years, more boys than girls scored above the 10th percentile on the DSS. There were no significant differences found at the ages of 4, 5, 6, and 7. At-test was used to compare average DSS scores between the two genders for each year of the study. This test revealed a significant difference between the LT girls' and LT boys' scores at the age of 3 years. No significant differences were found for the subsequent years. However, difference between boys' and girls' scores at age 7 approached significance, with boys again scoring higher.
APA, Harvard, Vancouver, ISO, and other styles
15

Niese, Hannah L. "Jump Start Vocabulary: Teaching Shape Bias to Increase Expressive Vocabulary." Bowling Green State University / OhioLINK, 2017. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1489924354210921.

Full text
APA, Harvard, Vancouver, ISO, and other styles
16

Houle, Jana A. "Effects of VCIU on the volitional speech of an individual with severe expressive aphasia /." abstract and full text PDF (UNR users only), 2004. http://0-gateway.proquest.com.innopac.library.unr.edu/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:1425296.

Full text
APA, Harvard, Vancouver, ISO, and other styles
17

Elwood, Terril Joy. "Maternal linguistic input to normal and expressive language delayed toddlers." PDXScholar, 1989. https://pdxscholar.library.pdx.edu/open_access_etds/3864.

Full text
Abstract:
Research suggests that the linguistic environment of the expressive language delayed child is different from that of his peers. Does this difference actually exist and if so, what are its characteristics? The purpose of this study was to describe the linguistic characteristics of mothers' input to children with normal language acquisition and those of mothers of expressively delayed toddlers; and to identify any differences between these groups. Though considerable research exists in this area, few studies have dealt specifically with large groups of expressively delayed toddlers.
APA, Harvard, Vancouver, ISO, and other styles
18

Intachakra, Songthama. "Linguistic politeness in British English and Thai : a comparative analysis of three expressive speech acts." Thesis, Queen Mary, University of London, 2001. http://qmro.qmul.ac.uk/xmlui/handle/123456789/28852.

Full text
Abstract:
This study attempts to further our understanding of linguistic politeness by focusing on both a Western and a non-Western language. It is based on two sets of data (one spontaneous and the other elicited) and provides a comparative analysis of three expressive speech acts produced by native speakers of British English and Thai. At face value, compliments, apologies and thanks may seem to have little referential meaning, yet these speech acts can be crucially important in originating, maintaining or even terminating social relationships. The data reveal a tendency for the two groups of speakers to use the three politeness devices in a different manner, reflecting cross-cultural differences in social norms and value systems. This project follows earlier studies of similar nature, in particular those carried out on different varieties of English. The findings are interpreted within pragmatic and sociolinguistic theoretical frameworks, and are discussed in the following format: linguistic structures of the speech acts, their functions, the topics of compliments, apologies and thanks, interpersonal and contextual factors influencing the production of these expressives, and the responses given to them. The analysis has implications for language specialists and lay people alike, in that it brings together a number of important insights with regard to these speech features that may result in miscommunication if and when British and Thai speakers converse in intercultural situations.
APA, Harvard, Vancouver, ISO, and other styles
19

Caggiano, Mariateresa <1995&gt. "Defining the Elegiac Genre in the Anglo-Saxon World: an Analysis of Expressive Speech Acts." Master's Degree Thesis, Università Ca' Foscari Venezia, 2020. http://hdl.handle.net/10579/18358.

Full text
Abstract:
Defining the genre 'elegy' in the Anglo-Saxon cultural horizon and at the same time delimiting the boundaries within which some Old English lyrical compositions might be grouped is a far from trivial operation. Indeed, contrary to the classical tradition where the elegiac meter – a dactylic hexameter verse followed by a dactylic pentameter- represents the distinctive element of the genre, the Anglo-Saxon poetry lacks a characterising meter for the elegies. Since the early nineteenth century scholars have been applying the word “elegy” to nine Old English poems found in the Exeter Book manuscript that have traditionally been treated as genre for their general sense of lament. The current study is part of this field of research and proposes a new attempt to delimit the elegy genre through a pragmatic analysis. In details, the “speech act theory" will be applied since it provides an interesting and productive new way of looking at the elegiac discourse. The analysis of speech acts will be conducted on two levels: at a micro-structural level, namely by focussing on individual illocutionary acts expressed in single sentences, and at a macro-structural level, by focussing on the so-called macro-speech acts, where it seems possible to recognise a global discourse illocutionary point together with a discourse illocutionary force. The identification of an expressive illocutionary force, in particular that of lamenting, at the macro level of speech will represent a possible demonstration of the elegiac nature of these compositions.
APA, Harvard, Vancouver, ISO, and other styles
20

Henry, James Allen. "A comparison of the expressive speech of profoundly hearing-impaired children : "hearing aids on" versus "hearing aids off"." PDXScholar, 1987. https://pdxscholar.library.pdx.edu/open_access_etds/3714.

Full text
Abstract:
This investigation was conducted to determine whether the removal of hearing aids from these children for eighteen hours (+ 1/2 hour and including sleep time) would result in reduced speech intelligibility as perceived by a panel of listening judges who were unfamiliar with the speech of the deaf.
APA, Harvard, Vancouver, ISO, and other styles
21

Snell, Katherine C. "Receptive and Expressive Language Outcomes of Children with Cochlear Implants and CHARGE Syndrome." University of Cincinnati / OhioLINK, 2008. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1212092204.

Full text
APA, Harvard, Vancouver, ISO, and other styles
22

Montaño, Aparicio Raúl. "Prosodic and Voice Quality Cross-Language Analysis of Storytelling Expressive Categories Oriented to Text-To-Speech Synthesis." Doctoral thesis, Universitat Ramon Llull, 2016. http://hdl.handle.net/10803/390960.

Full text
Abstract:
Durant segles, la interpretació oral de contes i històries ha sigut una tradició mundial lligada a l’entreteniment, la educació, i la perpetuació de la cultura. En les últimes dècades, alguns treballs s’han centrat en analitzar aquest estil de parla ric en matisos expressius caracteritzats per determinats patrons acústics. En relació a això, també hi ha hagut un interès creixent en desenvolupar aplicacions de contar contes, com ara les de contacontes interactius. Aquesta tesi està orientada a millorar aspectes claus d’aquest tipus d’aplicacions: millorar la naturalitat de la parla sintètica expressiva a partir d’analitzar la parla de contacontes en detall, a més a més de proporcionar un millor llenguatge no verbal a un avatar parlant mitjançant la sincronització de la parla i els gestos. Per aconseguir aquests objectius és necessari comprendre les característiques acústiques d’aquest estil de parla i la interacció de la parla i els gestos. Pel que fa a característiques acústiques de la parla de contacontes, la literatura relacionada ha treballat en termes de prosòdia, mentre que només ha estat suggerit que la qualitat de la veu pot jugar un paper important per modelar les subtileses d’aquest estil. En aquesta tesi, el paper tant de la prosòdia com de la qualitat de la veu en l’estil indirecte de la parla de contacontes en diferents idiomes és analitzat per identificar les principal categories expressives que la composen i els paràmetres acústics que les caracteritzen. Per fer-ho, es proposa una metodologia d’anotació per aquest estil de parla a nivell de oració basada en modes de discurs dels contes (mode narratiu, descriptiu, i diàleg), introduint a més sub-modes narratius. Considerant aquesta metodologia d’anotació, l’estil indirecte d’una història orientada a una audiència jove (cobrint versions en castellà, anglès, francès, i alemany) és analitzat en termes de prosòdia i qualitat de la veu mitjançant anàlisis estadístics i discriminants, després de classificar els àudios de les oracions de la història en les seves categories expressives. Els resultats confirmen l’existència de les categories de contes amb diferències expressives subtils en tots els idiomes més enllà dels estils personals dels narradors. En aquest sentit, es presenten evidències que suggereixen que les categories expressives dels contes es transmeten amb matisos expressius més subtils que en les emocions bàsiques, després de comparar els resultats obtinguts amb aquells de parla emocional. Els anàlisis també mostren que la prosòdia i la qualitat de la veu contribueixen pràcticament de la mateixa manera a l’hora de discriminar entre les categories expressives dels contes, les quals son expressades amb patrons acústics similars en tots els idiomes analitzats. Cal destacar també la gran relació observada en la selecció de categoria per cada oració que han fet servir els diferents narradors encara quan, que sapiguem, no se’ls hi va donar cap indicació. Per poder traslladar totes aquestes categories a un sistema de text a parla basat en corpus, caldria enregistrar un corpus per cada categoria. No obstant, crear diferents corpus ad-hoc esdevé un tasca molt laboriosa. En la tesi, s’introdueix una alternativa basada en una metodologia d’anàlisi orientada a síntesi dissenyada per derivar models de regles des de un petit però representatiu conjunt d’oracions, que poden poder ser utilitzats per generar parla amb estil de contacontes a partir de parla neutra. Els experiments sobre suspens creixent com a prova de concepte mostren la viabilitat de la proposta en termes de naturalitat i similitud respecte un narrador de contes real. Finalment, pel que fa a interacció entre parla i gestos, es realitza un anàlisi de sincronia i èmfasi orientat a controlar un avatar de contacontes en 3D. Al tal efecte, es defineixen indicadors de força tant per els gestos com per la parla. Després de validar-los amb tests perceptius, una regla d’intensitat s’obté de la seva correlació. A més a més, una regla de sincronia es deriva per determinar correspondències temporals entre els gestos i la parla. Aquests anàlisis s’han dut a terme sobre interpretacions neutres i agressives per part d’un actor per cobrir un gran rang de nivells d’èmfasi, com a primer pas per avaluar la integració d’un avatar parlant després del sistema de text a parla.<br>Durante siglos, la interpretación oral de cuentos e historias ha sido una tradición mundial ligada al entretenimiento, la educación, y la perpetuación de la cultura. En las últimas décadas, algunos trabajos se han centrado en analizar este estilo de habla rico en matices expresivos caracterizados por determinados patrones acústicos. En relación a esto, también ha habido un interés creciente en desarrollar aplicaciones de contar cuentos, como las de cuentacuentos interactivos. Esta tesis está orientada a mejorar aspectos claves de este tipo de aplicaciones: mejorar la naturalidad del habla sintética expresiva a partir de analizar el habla de cuentacuentos en detalle, además de proporcionar un mejor lenguaje no verbal a un avatar parlante mediante la sincronización del habla y los gestos. Para conseguir estos objetivos es necesario comprender las características acústicas de este estilo de habla y la interacción del habla y los gestos. En cuanto a características acústicas del habla de narradores de cuentos, la literatura relacionada ha trabajado en términos de prosodia, mientras que sólo ha sido sugerido que la calidad de la voz puede jugar un papel importante para modelar las sutilezas de este estilo. En esta tesis, el papel tanto de la prosodia como de la calidad de la voz en el estilo indirecto del habla de cuentacuentos en diferentes idiomas es analizado para identificar las principales categorías expresivas que componen este estilo de habla y los parámetros acústicos que las caracterizan. Para ello, se propone una metodología de anotación a nivel de oración basada en modos de discurso de los cuentos (modo narrativo, descriptivo, y diálogo), introduciendo además sub-modos narrativos. Considerando esta metodología de anotación, el estilo indirecto de una historia orientada a una audiencia joven (cubriendo versiones en castellano, inglés, francés, y alemán) es analizado en términos de prosodia y calidad de la voz mediante análisis estadísticos y discriminantes, después de clasificar los audios de las oraciones de la historia en sus categorías expresivas. Los resultados confirman la existencia de las categorías de cuentos con diferencias expresivas sutiles en todos los idiomas más allá de los estilos personales de los narradores. En este sentido, se presentan evidencias que sugieren que las categorías expresivas de los cuentos se transmiten con matices expresivos más sutiles que en las emociones básicas, tras comparar los resultados obtenidos con aquellos de habla emocional. Los análisis también muestran que la prosodia y la calidad de la voz contribuyen prácticamente de la misma manera a la hora de discriminar entre las categorías expresivas de los cuentos, las cuales son expresadas con patrones acústicos similares en todos los idiomas analizados. Cabe destacar también la gran relación observada en la selección de categoría para cada oración que han utilizado los diferentes narradores aun cuando, que sepamos, no se les dio ninguna indicación. Para poder trasladar todas estas categorías a un sistema de texto a habla basado en corpus, habría que grabar un corpus para cada categoría. Sin embargo, crear diferentes corpus ad-hoc es una tarea muy laboriosa. En la tesis, se introduce una alternativa basada en una metodología de análisis orientada a síntesis diseñada para derivar modelos de reglas desde un pequeño pero representativo conjunto de oraciones, que pueden ser utilizados para generar habla de cuentacuentos a partir de neutra. Los experimentos sobre suspense creciente como prueba de concepto muestran la viabilidad de la propuesta en términos de naturalidad y similitud respecto a un narrador de cuentos real. Finalmente, en cuanto a interacción entre habla y gestos, se realiza un análisis de sincronía y énfasis orientado a controlar un avatar cuentacuentos en 3D. Al tal efecto, se definen indicadores de fuerza tanto para gestos como para habla. Después de validarlos con tests perceptivos, una regla de intensidad se obtiene de su correlación. Además, una regla de sincronía se deriva para determinar correspondencias temporales entre los gestos y el habla. Estos análisis se han llevado a cabo sobre interpretaciones neutras y agresivas por parte de un actor para cubrir un gran rango de niveles de énfasis, como primer paso para evaluar la integración de un avatar parlante después del sistema de texto a habla.<br>For ages, the oral interpretation of tales and stories has been a worldwide tradition tied to entertainment, education, and perpetuation of culture. During the last decades, some works have focused on the analysis of this particular speaking style rich in subtle expressive nuances represented by specific acoustic cues. In line with this fact, there has also been a growing interest in the development of storytelling applications, such as those related to interactive storytelling. This thesis deals with one of the key aspects of audiovisual storytellers: improving the naturalness of the expressive synthetic speech by analysing the storytelling speech in detail, together with providing better non-verbal language to a speaking avatar by synchronizing that speech with its gestures. To that effect, it is necessary to understand in detail the acoustic characteristics of this particular speaking style and the interaction between speech and gestures. Regarding the acoustic characteristics of storytelling speech, the related literature has dealt with the acoustic analysis of storytelling speech in terms of prosody, being only suggested that voice quality may play an important role for the modelling of its subtleties. In this thesis, the role of both prosody and voice quality in indirect storytelling speech is analysed across languages to identify the main expressive categories it is composed of together with the acoustic parameters that characterize them. To do so, an analysis methodology is proposed to annotate this particular speaking style at the sentence level based on storytelling discourse modes (narrative, descriptive, and dialogue), besides introducing narrative sub-modes. Considering this annotation methodology, the indirect speech of a story oriented to a young audience (covering the Spanish, English, French, and German versions) is analysed in terms of prosody and voice quality through statistical and discriminant analyses, after classifying the sentence-level utterances of the story in their corresponding expressive categories. The results confirm the existence of storytelling categories containing subtle expressive nuances across the considered languages beyond narrators' personal styles. In this sense, evidences are presented suggesting that such storytelling expressive categories are conveyed with subtler speech nuances than basic emotions by comparing their acoustic patterns to the ones obtained from emotional speech data. The analyses also show that both prosody and voice quality contribute almost equally to the discrimination among storytelling expressive categories, being conveyed with similar acoustic patterns across languages. It is also worth noting the strong relationship observed in the selection of the expressive category per utterance across the narrators even when, up to our knowledge, no previous indications were given to them. In order to translate all these expressive categories to a corpus-based Text-To-Speech system, the recording of a speech corpus for each category would be required. However, building ad-hoc speech corpora for each and every specific expressive style becomes a very daunting task. In this work, we introduce an alternative based on an analysis-oriented-to-synthesis methodology designed to derive rule-based models from a small but representative set of utterances, which can be used to generate storytelling speech from neutral speech. The experiments conducted on increasing suspense as a proof of concept show the viability of the proposal in terms of naturalness and storytelling resemblance. Finally, in what concerns the interaction between speech and gestures, an analysis is performed in terms of time and emphasis oriented to drive a 3D storytelling avatar. To that effect, strength indicators are defined for speech and gestures. After validating them through perceptual tests, an intensity rule is obtained from their correlation. Moreover, a synchrony rule is derived to determine temporal correspondences between speech and gestures. These analyses have been conducted on aggressive and neutral performances to cover a broad range of emphatic levels as a first step to evaluate the integration of a speaking avatar after the expressive Text-To-Speech system.
APA, Harvard, Vancouver, ISO, and other styles
23

Belfiore, Kathleen. "Intervention History of Children with Slow Expressive Language Development." PDXScholar, 1996. https://pdxscholar.library.pdx.edu/open_access_etds/4944.

Full text
Abstract:
Children who are identified with slow expressive language development (SELD) around the age of two are producing less than fifty intelligible words or no two word phrases. Current research suggests that some children with SELD outgrow their delay while others continue to develop long term language difficulties. The literature shows varied findings of short term recovery but long term deficits, and shifts in the specific expressive language deficits ~s the child with SELD matures and encounters increased language demands. Suggestions are found for a mix of monitoring and early intervention, in step with signs of readiness and dynamic assessments, to facilitate improved performance and hasten development, particularly in the areas of metalinguistics and narratives. This study attempted to support the recommendation of early intervention, particularly for those children with an initial greater severity levels of expressive communication delay at the age of two. The 24 male and seven female SELD subjects were part of the Portland Language Development Project, a longitudinal study. Intake was at two years, and placement in the Intervention (Rx) or No Intervention (No Rx) group was a result of follow-up information gathered from parents regarding enrollment in any early intervention services before the age of four: Using mean Developmental Sentence Scores (DSS) for four outcome points, 1-tests determined that no significant differences existed in the improvement of language production between the Rx and No Rx groups. Secondly, 1-tests showed no significant differences in the two group's initial severity levels, using the Expressive Communication sub-domain of the Vine~and Adaptive Behavior Scale (V ABS), as the measure of severity at intake. A non-significant trend of consistently higher actual mean DSS scores across all outcome points, and an actual lower mean Expressive Communication score on the V ABS at intake was noted for the Rx group. A significant difference was found in the mean intake ages of the two groups, with older toddlers falling into the Rx group. Research and clinical_ implications are discussed, including attention to the length, type and content of very early intervention services, effective initial and follow up assessments, and factors that favor recommending early intervention
APA, Harvard, Vancouver, ISO, and other styles
24

Miller, Sherri Lynn. "Percentage of phonological process usage in expressive language delayed children." PDXScholar, 1991. https://pdxscholar.library.pdx.edu/open_access_etds/4204.

Full text
Abstract:
Language delay and phonological delay have been shown to coexist. Because they so often co-occur, it is possible that they may interact, sharing a relationship during the child's development. A group of children who were "late talkers" as toddlers, achieved normal development in their syntactic ability by the preschool period. Because their language abilities are known to have increased rapidly, data on their phonological development could provide information on the relationship between phonological and syntactic development. The purpose of this study was to compare the percentage of phonological process usage of the eight most commonly used simplification processes in four-year-old expressive language delayed (ELD) children, children with a history of slow expressive language development (HX), and normally developing (ND) children. The questions this study sought to answer were: do ELD children exhibit a higher percentage of phonological process usage than ND children, and are HX children significantly different in their percentage of phonological process usage than ND and/or ELD children.
APA, Harvard, Vancouver, ISO, and other styles
25

Mac, Dang Khoa. "Génération de parole expressive dans le cas des langues à tons." Phd thesis, Université de Grenoble, 2012. http://tel.archives-ouvertes.fr/tel-00859201.

Full text
Abstract:
De plus en plus, l'interaction entre personne et machine se rapproche du naturel afin de ressembler à l'interaction entre humains, incluant l'expressivité (en particulier les émotions et les attitudes). Dans la communication parlée, les attitudes, et plus généralement les affects sociaux, sont véhiculés principalement par la prosodie. Pour les langues tonales, la prosodie est utilisée aussi pour coder l'information sémantique dans les variations de tons. Ce travail de thèse présente une étude des affects sociaux du vietnamien, une langue à tons et une langue peu dotée, afin d'appliquer les résultats obtenus à un système de synthèse de haute qualité capable de produire la parole " expressive " pour le vietnamien. Le premier travail de cette thèse consiste en la construction du premier corpus audio-visuel des attitudes vietnamiennes, qui contient seize attitudes. Ce corpus est ensuite utilisé pour étudier la perception audio-visuelle et interculturelle des attitudes vietnamiennes. Pour cela, une série de tests perceptifs a été effectuée avec des auditeurs natifs et non-natifs (des auditeurs francophones pour les non-natifs). Les résultats de ces tests montrent que les facteurs influant sur la perception des attitudes sont l'expression de l'attitude elle-même et la modalité de présentation (audio, visuelle et audio-visuelle). Ces résultats nous ont ainsi permis de trouver des affects sociaux communs ou interculturels entre le vietnamien et le français. Puis, un autre test de perception a été réalisé sur des phrases avec tons afin d'explorer l'effet du système tonal du vietnamien sur la perception des attitudes. Les résultats montrent que les juges non-natifs peuvent traiter et séparer les indices tonals locaux et les traits saillants prosodiques de portée globale. Après une présentation de nos études sur les affects sociaux en vietnamien, nous décrivons notre modélisation de la prosodie des attitudes en vue de la synthèse de la parole expressive en vietnamien. En nous basant sur le modèle de superposition des contours fonctionnels, nous proposons une méthode pour modéliser et générer de la prosodie expressive en vietnamien. Cette méthode est ensuite appliquée pour générer de la parole expressive en vietnamien, puis évaluée par des tests de perception sur les énoncés synthétiques. Les résultats de perception valident bien la performance de notre modèle et confirment que l'approche de superposition de contours fonctionnels peut être utilisée pour modéliser une prosodie complexe comme dans le cas de la parole expressive d'une langue à tons.
APA, Harvard, Vancouver, ISO, and other styles
26

Lohr-Flanders, Marla. "The effect of otitis media on articulation in expressive language-delayed children." PDXScholar, 1992. https://pdxscholar.library.pdx.edu/open_access_etds/4365.

Full text
Abstract:
Researchers have long been concerned with the effects of otitis media on speech and language acquisition because of the high correlation of a mild to moderate hearing loss during the time period that fluid (effusion) may be in the middle ear. Middle-ear effusion would prevent many of the auditory messages from accurately reaching the nervous system (Zinkus, 1986). Deprived of the ability to discern the subtle acoustic differences that provide information for phonetic contrasts, a child's speech acquisition may differ from children who do not experience such losses.
APA, Harvard, Vancouver, ISO, and other styles
27

Oliveira, de Sousa Felipe. "Reason-giving as an act of recognition." Thesis, University of Edinburgh, 2017. http://hdl.handle.net/1842/25838.

Full text
Abstract:
This thesis defends the claim that reason-giving is a discrete type of speech act (of an expressive kind) that has a distinctive value. It further argues that this value is best understood in terms of recognition, rather than justification, and that it is intrinsic to reason-giving. Its main aim is to argue against the commonly-held view that the main, sometimes the only, value to reason-giving lies in its capacity to provide justification (and in the related claim that if reasons cannot justify, then reason-giving has no value). The argument presented is intended to support that recognition (of a certain type) is a value that reason-giving has independently from any other value that it might or might not have – including justification; and hence, that reason-giving has a certain distinctive value that is not predicated upon a capacity for actually achieving justification. In particular, this thesis argues, based on speech act theory and on the concept of recognition, that this value is best understood as consisting in the expression of a particular type of recognition for the other. To establish this claim, in chapter one, it begins by setting out the standard view: that the value of reason-giving lies in its capacity to justify, and analyses some of the moves that have been made in the literature when the connection between reason-giving and justification breaks down. In chapters two to four, it uses speech act theory to analyse the acts of arguing and reason-giving, and to argue that reason-giving is a discrete speech act that has features in common with but is not reducible to arguing. Finally, in chapter five, it defends the claim that reason-giving has an intrinsic value, and that this value is best understood as an expressive value: namely, the expression of recognition for the other as a rational being (which is a valuable feature of the other’s humanity); and that it has this value regardless of whether the reasons in question are “good” from a justificatory standpoint.
APA, Harvard, Vancouver, ISO, and other styles
28

Chipinka, Megan. "The effects of augmented input on receptive and expressive language for native augmentative and alternative communication (AAC) users during shared storybook readings." Thesis, California State University, Long Beach, 2017. http://pqdtopen.proquest.com/#viewpdf?dispub=10196441.

Full text
Abstract:
<p> The study was a pre-experimental pre- and post-treatment single case study which focused on evaluating the effects of augmentative and alternative communication (AAC) modeling during shared storybook readings. Aided AAC modeling interventions provide verbal and visual language models to support language comprehension and use for children with complex communication needs (CCN). The study measured four aspects of change before and after the AAC modeling phase including a) the number of communicative turns by the AAC user; b) the complexity and length of the initiations and responses made by the AAC user; c) the accuracy in responses to comprehension questions following the story; d) perceptions of the parent participant in regards to the intervention. The results indicated that when aided AAC modeling was implemented the child participant demonstrated an increase in: the number of communicative turns, accuracy in answering comprehension questions, comprehension of story grammar terminology, and production of story retells.</p>
APA, Harvard, Vancouver, ISO, and other styles
29

Spangle-Looney, Shawn. "Communication and socialization profiles in toddlers with expressive language delay." PDXScholar, 1988. https://pdxscholar.library.pdx.edu/open_access_etds/3843.

Full text
Abstract:
The purpose of this study was to compare expressive communication, receptive communication, and socialization achievement in 18- to 34-month-old ELD toddlers to the same skills in normally-speaking children. The questions this study sought to answer were, how do the three skill areas in ELD toddlers compare with the same skills in normal toddlers?, will ELD subjects evidence specific profiles of deficits involving not only expressive but receptive and social skills as well? and, within the ELD subjects will two subgroups emerge, one group having poor expressive skills only, and a second group having deficits in addition to expression.
APA, Harvard, Vancouver, ISO, and other styles
30

Rivera, Perez Jean F. "The Use of Text-to-Speech to Teach Vocabulary to English Language Learners." University of Cincinnati / OhioLINK, 2016. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1470753301.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Zuehlsdorff, Kathleen Marie. "A comparative study of the TEEM and the morphological aspects of the BLST and TOLD-P." PDXScholar, 1985. https://pdxscholar.library.pdx.edu/open_access_etds/3525.

Full text
Abstract:
The purpose of this investigation was to examine the construct validity of a new test, which purports to measure morphology, entitled Test for Examining Expressive Morphology (TEEM) (Shipley, Stone and Sue, 1983). Additional tests of expressive morphology, the Bankson Language Screening Test (BLST) (Bankson, 1977) and the Test of Language Development-Primary (TOLD-P) (Newcomer and Hammill, 1982) were utilized to determine the association of the TEEM with two highly-researched instruments.
APA, Harvard, Vancouver, ISO, and other styles
32

Boggs, Teresa. "Parent Seminar Providing Relevant Parent Education: Expressive Language Skills in Children with Autism." Digital Commons @ East Tennessee State University, 2015. https://dc.etsu.edu/etsu-works/1518.

Full text
APA, Harvard, Vancouver, ISO, and other styles
33

Baker, Kim Denise. "A comparison of expressive vocabulary produced by nonambulatory, speaking preschool children and ambulatory speaking preschool children." PDXScholar, 1992. https://pdxscholar.library.pdx.edu/open_access_etds/4250.

Full text
Abstract:
Children with severe physical disabilities often do not have the capabilities for oral communication. Professionals are frequently faced with selecting vocabulary for children who are unable to use vocal output because of severe motor impairments. A child who is nonambulatory may have additional reasons for communicating and sees the world from a different viewpoint than his ambulatory peers. Selecting appropriate words for an initial lexicon that are useful to nonspeaking disabled children that also meet normal language acquisition standards is a concern. This study specifically addresses this concern by looking at the vocabulary differences of ambulatory and nonambulatory preschool children. The purpose of this research project was to compare expressive vocabulary produced by nonambulatory, speaking children with the expressive vocabulary produced by ambulatory, speaking children. It is suggested that the vocabulary of nonambulatory, speaking children might be more appropriate for selecting a lexicon for AAC systems if indeed, they are different from words produced by ambulatory, speaking children.
APA, Harvard, Vancouver, ISO, and other styles
34

Sturmel, Nicolas. "Analyse de la qualité vocale appliquée à la parole expressive." Phd thesis, Université Paris Sud - Paris XI, 2011. http://tel.archives-ouvertes.fr/tel-00591638.

Full text
Abstract:
L'analyse des signaux de parole permet de comprendre le fonctionnement de l'appareil vocal, mais aussi de décrire de nouveaux paramètres permettant de qualifier et quantifier la perception de la voix. Dans le cas de la parole expressive, l'intérêt se porte sur des variations importantes de qualité vocales et sur leurs liens avec l'expressivité et l'intention du sujet. Afin de décrire ces liens, il convient de pouvoir estimer les paramètres du modèle de production mais aussi de décomposer le signal vocal en chacune des parties qui contribuent à ce modèle. Le travail réalisé au cours de cette thèse s'axe donc autour de la segmentation et la décomposition des signaux vocaux et de l'estimation des paramètres du modèle de production vocale : Tout d'abord, la décomposition multi-échelles des signaux vocaux est abordée. En reprenant la méthode LoMA qui trace des lignes suivant les amplitudes maximum sur les réponses temporelles au banc de filtre en ondelettes, il est possible d'y détecter un certain nombre de caractéristiques du signal vocal : les instants de fermeture glottique, l'énergie associée à chaque cycle ainsi que sa distribution spectrale, le quotient ouvert du cycle glottique (par l'observation du retard de phase du premier harmonique). Cette méthode est ensuite testée sur des signaux synthétiques et réels. Puis, la décomposition harmonique + bruit des signaux vocaux est abordée. Une méthode existante (PAPD - Périodic/APériodic Décomposition) est adaptée aux variations de fréquence fondamentale par le biais de la variation dynamique de la taille de la fenêtre d'analyse et est appelée PAP-A. Cette nouvelle méthode est ensuite testée sur une base de signaux synthétiques. La sensibilité à la précision d'estimation de la fréquence fondamentale est notamment abordée. Les résultats montrent des décompositions de meilleures qualité pour PAP-A par rapport à PAPD. Ensuite, le problème de la déconvolution source/filtre est abordé. La séparation source/filtre par ZZT (zéros de la transformée en Z) est comparée aux méthodes usuelles à base de prédiction linéaire. La ZZT est utilisée pour estimer les paramètres du modèle de la source glottique via une méthode simple mais robuste qui permet une estimation conjointe de deux paramètres du débit glottique : le quotient ouvert et l'asymétrie. La méthode ainsi développée est testée et combinée à l'estimation du quotient ouvert par ondelettes. Finalement, ces trois méthodes d'estimations sont appliquées à un grand nombre de fichiers d'une base de données comportant différents styles d'élocution. Les résultats de cette analyse sont discutés afin de caractériser le lien entre style, valeur des paramètres de la production vocale et qualité vocale. On constate notamment l'émergence très nette de groupes de styles.
APA, Harvard, Vancouver, ISO, and other styles
35

Wagener, Jeanie P. "The correlation between correct verbal and nonverbal responses on an intelligence test and expressive language test score." PDXScholar, 1991. https://pdxscholar.library.pdx.edu/open_access_etds/4241.

Full text
Abstract:
The purpose of this study was to examine the relationship between verbal and nonverbal indices of intelligence from an infant development scale given at approximately two years of age with scores on an expressive language test administered to the same children two years later. The questions this study sought to answer were: (1) is there a difference between the number of verbal and nonverbal items passed at 18-34 months by a group of normal children and a group of late talkers, (2) is there a significant relationship between correct verbal and nonverbal responses on an inf ant development test and scores on an expressive language test two years later in normal children and late talkers, (3) is there a significant relationship between correct verbal receptive items on an infant development test and scores on an expressive language test two years later in normal children and late talkers, and (4) is there a significant relationship between correct expressive responses on an infant development test and expressive scores two years later in the normal children and the group of later talkers.
APA, Harvard, Vancouver, ISO, and other styles
36

Doukhan, David. "Synthèse de parole expressive au delà du niveau de la phrase : le cas du conte pour enfant : conception et analyse de corpus de contes pour la synthèse de parole expressive." Thesis, Paris 11, 2013. http://www.theses.fr/2013PA112165/document.

Full text
Abstract:
L'objectif de la thèse est de proposer des méthodes permettant d'améliorer l'expressivité des systèmes de synthèse de la parole. Une des propositions centrales de ce travail est de définir, utiliser et mesurer l'impact de structures linguistiques opérant au delà du niveau de la phrase, par opposition aux approches opérant sur des phrases isolées de leur contexte. Le cadre de l'étude est restreint au cas de la lecture de contes pour enfants. Les contes ont la particularité d'avoir été l'objet d'un certain nombre d'études visant à en dégager une structure narrative et de faire intervenir une certain nombre de stéréotypes de personnages (héros, méchant, fée) dont le discours est souvent rapporté. Ces caractéristiques particulières sont exploitées pour modéliser les propriétés prosodiques des contes au delà du niveau de la phrase. La transmission orale des contes a souvent été associée à une pratique musicale (chants, instruments) et leur lecture reste associée à des propriétés mélodiques très riches, dont la reproduction reste un défi pour les synthétiseurs de parole modernes. Pour répondre à ces problématiques, un premier corpus de contes écrits est collecté et annoté avec des informations relatives à la structure narrative des contes, l'identification et l'attribution des citations directes, le référencement des mentions des personnages ainsi que des entités nommées et des énumérations étendues. Le corpus analysé est décrit en terme de couverture et d'accord inter-annotateurs. Il est utilisé pour modéliser des systèmes de segmentation des contes en épisode, de détection des citations directes, des actes de dialogue et des modes de communication. Un deuxième corpus de contes lus par un locuteur professionnel est présenté. La parole est alignée avec les transcriptions lexicale et phonétique, les annotations du corpus texte et des méta-informations décrivant les caractéristiques des personnages intervenant dans le conte. Les relations entre les annotations linguistiques et les propriétés prosodiques observées dans le corpus de parole sont décrites et modélisées. Finalement, un prototype de contrôle des paramètres expressifs du synthétiseur par sélection d'unités Acapela est réalisé. Le prototype génère des instructions prosodiques opérant au delà du niveau de la phrase, notamment en utilisant les informations liées à la structure du conte et à la distinction entre discours direct et discours rapporté. La validation du prototype de contrôle est réalisée dans le cadre d'une expérience perceptive, qui montre une amélioration significative de la qualité de la synthèse<br>The aim of this thesis is to propose ways to improve the expressiveness of speech synthesis systems. One of the central propositions of this work is to define, use and measure the impact of linguistic structures operating beyond the sentence level, as opposed to approaches operating on sentences out of their context. The scope of the study is restricted to the case of storytelling for children. The stories have the distinction of having been the subject of a number of studies in order to highlight a narrative structure and involve a number of stereotypical characters (hero, villain, fairy) whose speech is often reported. These special features are used to model the prosodic properties tales beyond the sentence level. The oral transmission of tales was often associated with musical practice (vocals, instruments) and their reading is associated with rich melodic properties including reproduction remains a challenge for modern speech synthesizers. To address these issues, a first corpus of written tales is collected and annotated with information about the narrative structure of stories, identification and allocation of direct quotations, referencing references to characters as well as named entities and enumerations areas. The corpus analyzed is described in terms of coverage and inter-annotator agreement. It is used to model systems segmentation tales episode, detection of direct quotes, dialogue acts and modes of communication. A second corpus of stories read by a professional speaker is presented. The word is aligned with the lexical and phonetic transcriptions, annotations of the corpus text and meta-information describing the characteristics of the characters involved in the story. The relationship between linguistic annotations and prosodic properties observed in the speech corpus are described and modeled. Finally, a prototype control expressive synthesizer parameters by Acapela unit selection is made. The prototype generates prosodic operating instructions beyond the sentence level, including using the information related to the structure of the story and the distinction between direct speech and reported speech. Prototype validation control is performed through a perceptual experience, which shows a significant improvement in the quality of the synthesis
APA, Harvard, Vancouver, ISO, and other styles
37

Culley, Amanda. "The Effects of Event Knowledge and Parent Input on the Language Skills of Children with and without Language Impairment." Ohio University Honors Tutorial College / OhioLINK, 2013. http://rave.ohiolink.edu/etdc/view?acc_num=ouhonors1366997797.

Full text
APA, Harvard, Vancouver, ISO, and other styles
38

Ryckbost, Lisa M. "PHONOLOGICAL AND LEXICAL INFLUENCES ON VOCABULARY DEVELOPMENT: PREDICTING THE AGES AT WHICH INDIVIDUAL WORDS ARE ACQUIRED." Bowling Green State University / OhioLINK, 2006. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1143483102.

Full text
APA, Harvard, Vancouver, ISO, and other styles
39

Clancy, Kathleen Ann. "Second Grade Academic Performance in Normal Children, Children with a History of, and Children with Expressive Language Delay." PDXScholar, 1994. https://pdxscholar.library.pdx.edu/open_access_etds/4740.

Full text
Abstract:
Interest in children who are diagnosed with expressive language delay has increased over the years. This has resulted in follow-up studies which have suggested that these children would have difficulties in academics during their elementary school years (Hall & Tomblin, 1978; Weiner, 1974) The current study sought to determine if children with a history of and children with continued expressive language delay would have problems with academics once they reached the second grade. The Peabody Individual Achievement Test (PIAT) was used to measure academic performance. It was chosen for it's reliable standardization and use of five different subtest areas to determine overall academic achievement. The purpose of this study was to determine if there are significant differences in academic performance on PIAT between three groups of second grade children with different language histories. The three groups are: 1) children with normal language history 2) children with a history of expressive language delay (HELD) who were identified as late to talk between 20 and 34 months of age, but who received a score at or above the tenth percentile in the second grade on the DSS (Developmental Sentence Scoring, Lee 1974), and 3) children with chronic expressive language delay (ELD) who were identified as late to talkers between 20 and 34 months of age, and received a score below the tenth percentile in the second grade on the DSS. Significant differences were found between the ELD group and the Normal group in the areas of Math and General Information as well as the Total Test Score. The ELD group also performed significantly lower than the HELD group in the areas of Math and the Total Test Score. There were no significant differences found between the HELD group and the Normals or between the ELD and HELD groups on the General Information subtest. These results were consistent with the most recent research article by Whitehurst and Fischel (1994) which looked at three longitudinal studies and found that by five years of age most children diagnosed with specific expressive language delay were performing within the normal range in ·various areas of language development.
APA, Harvard, Vancouver, ISO, and other styles
40

Andrews, David J. "A Comparative Study of Phonemic Segmentation Skills in First Grade Children with Normal, Disordered, and Slow Expressive Language Development." PDXScholar, 1994. https://pdxscholar.library.pdx.edu/open_access_etds/4750.

Full text
Abstract:
Children with slow expressive language development often catch up to their normally developing peers in expressive language, but may still exhibit difficulties with metalinguistic skills. Research shows that children who have difficulty with phonemic awareness also have difficulty with reading, which is important for success in school. Speech-language pathologists assist children who have difficulty with expressive oral language and facilitate language development in children who have difficulties with learning metalinguistic skills, such as phonemic awareness. The purpose of the present study was to compare the phoneme segmentation skills in three groups of children: (a) children with a history of oral expressive language delay (HELD) (n= 22) who were identified as toddlers with slow developing expressive language, but caught up to their normally developing peers by first grade; (b) children identified as toddlers with slow developing oral expressive language and by first grade still maintained the expressive language delays (ELD) (n= 7); and (c) children who were identified at age two as developing normal oral expressive language and maintained normal oral expressive language development (NL) (n= 23) in first grade. The children participated in a phonological segmentation test. The study answered four questions: Is there a significant difference among the three groups of children in the number of correct responses on a phonological segmentation test at ( 1 ) the one phoneme level, (2) the two phoneme level, (3) the three phoneme level, and ( 4) the total number of correct responses. Utilizing an ANOVA test, a significant difference was found among the groups at the two phoneme level, with a trend toward a significant difference at the one phoneme level. Other significant differences were not found. The difference at the two phoneme level was between the ELD group and the normal group, as well as between the ELD group and the HELD group.
APA, Harvard, Vancouver, ISO, and other styles
41

Evrard, Marc. "Synthèse de parole expressive à partir du texte : Des phonostyles au contrôle gestuel pour la synthèse paramétrique statistique." Thesis, Paris 11, 2015. http://www.theses.fr/2015PA112202.

Full text
Abstract:
L’objectif de cette thèse est l’étude et la conception d’une plateforme de synthèse de parole expressive.Le système de synthèse — LIPS3, développé dans le cadre de ce travail, incorpore deux éléments : un module linguistique et un module de synthèse paramétrique par apprentissage statistique (construit à l’aide de HTS et de STRAIGHT). Le système s’appuie sur un corpus monolocuteur conçu, enregistréet étiqueté à cette occasion.Une première étude, sur l’influence de la qualité de l’étiquetage du corpus d’apprentissage, indique que la synthèse paramétrique statistique est robuste aux erreurs de labels et d’alignement. Cela répond au problème de la variation des réalisations phonétiques en parole expressive.Une seconde étude, sur l’analyse acoustico-phonétique du corpus permet la caractérisation de l’espace expressif utilisé par la locutrice pour réaliser les consignes expressives qui lui ont été fournies. Les paramètres de source et les paramètres articulatoires sont analysés suivant les classes phonétiques, ce qui permet une caractérisation fine des phonostyles.Une troisième étude porte sur l’intonation et le rythme. Calliphony 2.0 est une interface de contrôlechironomique temps-réel permettant la modification de paramètres prosodiques (f0 et tempo) des signaux de synthèse sans perte de qualité, via une manipulation directe de ces paramètres. Une étude sur la stylisation de l’intonation et du rythme par contrôle gestuel montre que cette interface permet l’amélioration, non-seulement de la qualité expressive de la parole synthétisée, mais aussi de la qualité globale perçue en comparaison avec la modélisation statistique de la prosodie.Ces études montrent que la synthèse paramétrique, combinée à une interface chironomique, offre une solution performante pour la synthèse de la parole expressive, ainsi qu’un outil d’expérimentation puissant pour l’étude de la prosodie<br>The subject of this thesis was the study and conception of a platform for expressive speech synthesis.The LIPS3 Text-to-Speech system — developed in the context of this thesis — includes a linguistic module and a parametric statistical module (built upon HTS and STRAIGHT). The system was based on a new single-speaker corpus, designed, recorded and annotated.The first study analyzed the influence of the precision of the training corpus phonetic labeling on the synthesis quality. It showed that statistical parametric synthesis is robust to labeling and alignment errors. This addresses the issue of variation in phonetic realizations for expressive speech.The second study presents an acoustico-phonetic analysis of the corpus, characterizing the expressive space used by the speaker to instantiate the instructions that described the different expressive conditions. Voice source parameters and articulatory settings were analyzed according to their phonetic classes, which allowed for a fine phonostylistic characterization.The third study focused on intonation and rhythm. Calliphony 2.0 is a real-time chironomic interface that controls the f0 and rhythmic parameters of prosody, using drawing/writing hand gestures with a stylus and a graphic tablet. These hand-controlled modulations are used to enhance the TTS output, producing speech that is more realistic, without degradation as it is directly applied to the vocoder parameters. Intonation and rhythm stylization using this interface brings significant improvement to the prototypicality of expressivity, as well as to the general quality of synthetic speech.These studies show that parametric statistical synthesis, combined with a chironomic interface, offers an efficient solution for expressive speech synthesis, as well as a powerful tool for the study of prosody
APA, Harvard, Vancouver, ISO, and other styles
42

Simonson, Kathryn. "Impact of Age of Implantation on Receptive and Expressive Language Outcomes of Children with Developmental Disabilities and Cochlear Implants." University of Cincinnati / OhioLINK, 2010. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1277140787.

Full text
APA, Harvard, Vancouver, ISO, and other styles
43

Jones, Beth Ann. "Temperament Differences in Children with a History of Slow Expressive Language Development and Their Peers with Normal Language Development." PDXScholar, 1996. https://pdxscholar.library.pdx.edu/open_access_etds/5281.

Full text
Abstract:
Language is the way in which individuals are able to express ideas, feelings, needs, expectations, and form relationships with others in their surrounding environment. A disruption in language development may negatively impact a child's social development. Research shows that children with language delays or disorders tend to have increased social and behavioral difficulties (Cantwell & Baker, 1977; Caulfield, Fischel, DeBaryshe, & Whitehurst, 1989). However, research has not examined temperament differences in young children with language delays or disorders. The question this study sought to answer was: Is there a significant difference in the dimensions of temperament between children with a history of slow expressive language development and their peers with normal language development? The subjects in this study ranged in age from 64 to 74 months. They included 33 children with a history of slow expressive language development (SELD) and 27 children with normal language development. The temperamental characteristics of each of the subjects was assessed by ratings provided by their parents, utilizing the Parent Temperament Questionnaire for Children (Thomas, Chess, & Korn, 1977), a questionnaire to assess the way a child behaves during everyday situations. The question was analyzed by calculating the means and standard deviations for the nine temperament dimensions for· the two groups. To determine if there were significant differences among the two groups, two tailed t-tests were computed at the .05 level of significance. The Parent Temperament Questionnaire for Children (Thomas et al., 1977) showed the children with a history of slow expressive language development to have significantly different scores in the dimensions of approach-withdrawal and intensity of reaction than the subjects with normal language development at 5 years of age.
APA, Harvard, Vancouver, ISO, and other styles
44

Kouklia, Charlotte. "Dominance, hostilité et expressivité vocale dans le débat politique ˸ étude perceptive et acoustique du conseil municipal de Montreuil (93100)." Thesis, Sorbonne Paris Cité, 2019. http://www.theses.fr/2019USPCA013/document.

Full text
Abstract:
Cette thèse étudie les rapports entre l’expression d’attitudes agressives contrôlées et la perception de la dominance, à partir d’extraits de séances télévisées du conseil Municipal de Montreuil (93100) durant l’année 2013, période marquée par un climat politique vif et hostile. Un corpus a été constitué à partir d’extraits de parole spontanée de la Maire, Dominique Voynet, et de quatre de ses opposants. Les cinq locuteurs ont participé à l’enregistrement d’une relecture neutre du corpus de leurs propres extraits de parole (25 stimuli par locuteur) ainsi qu’à l’auto-évaluation perceptive de leurs stimuli (profils émotionnels), dont les résultats ont été comparés à l’évaluation perceptive des extraits par des auditeurs naïfs. Tous les locuteurs ont été perçus comme légèrement hostiles, légèrement mécontents et légèrement dominants malgré des profils émotionnels auto-évalués variables. Les extraits originaux et relus ont été comparés au niveau de leur structuration prosodico-syntaxique et de leurs caractéristiques temporelles et mélodiques. Les résultats montrent que 1) certains locuteurs semblent plus s’appuyer sur des paramètres mélodiques et d’autres sur des paramètres temporels ; 2) on peut néanmoins dégager les tendances générales concernant les corrélats dans la parole de l’hostilité et de la dominance dans notre corpus : a) des écarts entre structuration syntaxique et prosodique des extraits, b) la réduction ou l’absence d’allongements syllabiques finaux pré-pausaux, c) de fortes variations de plage de variation de F0 de part et d’autre des pauses silencieuses<br>This dissertation studies the relationship between the expression of controlled aggressive attitudes and the perception of dominance, based on extracts from televised sessions of the Municipal Council of Montreuil during 2013; a period marked by a lively and hostile political climate. We constituted a corpus of spontaneous speech extracts from the Mayor, Dominique Voynet, and four of her opponents. During subsequent recording sessions, the five speakers were asked to read transcriptions of their own speech extracts in a neutral tone (25 stimuli per speaker). They also participated in a self-evaluation questionnaire that focused on the perception of emotional profiles in their own stimuli. The results of the speakers’ perceptive evaluations were compared to those of a control group of naïve listeners. All speakers were perceived as slightly hostile, discontent and dominant, despite the variability observed in the self-evaluated profiles. The original and re-read extracts were compared in their prosodic- syntactic structure as well as their temporal and melodic characteristics. We show that: 1) some speakers seem to rely mostly on melodic parameters whereas others primarily use temporal parameters, 2) nevertheless, general trends emerge regarding the speech correlates of hostility and dominance in our corpus, notably: a) discrepancies between the syntactic and the prosodic structure of the extracts, b) reduction or absence of pre-pausal final syllabic lengthening, c) large variations in F0 range on both sides of silent pauses
APA, Harvard, Vancouver, ISO, and other styles
45

Sini, Aghilas. "Caractérisation et génération de l’expressivité en fonction des styles de parole pour la construction de livres audio." Thesis, Rennes 1, 2020. http://www.theses.fr/2020REN1S026.

Full text
Abstract:
Dans ces travaux de thèse nous abordons l'expressivité de la parole lue avec un type de données particulier qui sont les livres audio. Les livres audio sont des enregistrements audio d’œuvres littéraires fait par des professionnels (des acteurs, des chanteurs, des narrateurs professionnels) ou par des amateurs. Ces enregistrements peuvent être destinés à un public particulier (aveugles ou personnes mal voyantes). La disponibilité de ce genre de données en grande quantité avec une assez bonne qualité a attiré l'attention de la communauté scientifique en traitement automatique du langage et de la parole en général, ainsi que des chercheurs spécialisés dans la synthèse de parole expressive. Pour explorer ce vaste champ d'investigation qui est l'expressivité, nous proposons dans cette thèse d'étudier trois entités élémentaires de l'expressivité qui sont véhiculées par les livres audio: l'émotion, les variations liées aux changements discursifs et les propriétés du locuteur. Nous traitons ces patrons d'un point de vue prosodique. Les principales contributions de cette thèse sont la construction d'un corpus de livres audio comportant un nombre important d'enregistrements partiellement annotés par un expert, une étude quantitative caractérisant les émotions dans ce type de données, la construction de modèles basés sur des techniques d'apprentissage automatique pour l'annotation automatique de types de discours et enfin nous proposons une représentation vectorielle de l'identité prosodique d'un locuteur dans le cadre de la synthèse statistique paramétrique de la parole<br>In this thesis, we study the expressivity of read speech with a particular type of data, which are audiobooks. Audiobooks are audio recordings of literary works made by professionals (actors, singers, professional narrators) or by amateurs. These recordings may be intended for a particular audience (blind or visually impaired people). The availability of this kind of data in large quantities with a good enough quality has attracted the attention of the research community in automatic speech and language processing in general and of researchers specialized in expressive speech synthesis systems. We propose in this thesis to study three elementary entities of expressivity that are conveyed by audiobooks: emotion, variations related to discursive changes, and speaker properties. We treat these patterns from a prosodic point of view. The main contributions of this thesis are: the construction of a corpus of audiobooks with a large number of recordings partially annotated by an expert, a quantitative study characterizing the emotions in this type of data, the construction of a model based on automatic learning techniques for the automatic annotation of discourse types and finally we propose a vector representation of the prosodic identity of a speaker in the framework of parametric statistical speech synthesis
APA, Harvard, Vancouver, ISO, and other styles
46

Smith, Kristen A. "The Receptive and Expressive Language Outcomes of Children who have Received Cochlear Implants and have an Autism Spectrum Disorder." University of Cincinnati / OhioLINK, 2008. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1212095057.

Full text
APA, Harvard, Vancouver, ISO, and other styles
47

Siniša, Suzić. "Parametarska sinteza ekspresivnog govora." Phd thesis, Univerzitet u Novom Sadu, Fakultet tehničkih nauka u Novom Sadu, 2019. https://www.cris.uns.ac.rs/record.jsf?recordId=110631&source=NDLTD&language=en.

Full text
Abstract:
U disertaciji su opisani postupci sinteze ekspresivnog govorakorišćenjem parametarskih pristupa. Pokazano je da se korišćenjemdubokih neuronskih mreža dobijaju bolji rezultati nego korišćenjemskrivenix Markovljevih modela. Predložene su tri nove metode zasintezu ekspresivnog govora korišćenjem dubokih neuronskih mreža:metoda kodova stila, metoda dodatne obuke mreže i arhitekturazasnovana na deljenim skrivenim slojevima. Pokazano je da se najboljirezultati dobijaju korišćenjem metode kodova stila. Takođe jepredložana i nova metoda za transplantaciju emocija/stilovabazirana na deljenim skrivenim slojevima. Predložena metodaocenjena je bolje od referentne metode iz literature.<br>In this thesis methods for expressive speech synthesis using parametricapproaches are presented. It is shown that better results are achived withusage of deep neural networks compared to synthesis based on hiddenMarkov models. Three new methods for synthesis of expresive speech usingdeep neural networks are presented: style codes, model re-training andshared hidden layer architecture. It is shown that best results are achived byusing style code method. The new method for style transplantation based onshared hidden layer architecture is also proposed. It is shown that thismethod outperforms referent method from literature.
APA, Harvard, Vancouver, ISO, and other styles
48

Ding, Yu. "Modèle statistique de l'animation expressive de la parole et du rire pour un agent conversationnel animé." Thesis, Paris, ENST, 2014. http://www.theses.fr/2014ENST0050/document.

Full text
Abstract:
Notre objectif est de simuler des comportements multimodaux expressifs pour les agents conversationnels animés ACA. Ceux-ci sont des entités dotées de capacités affectives et communicationnelles; ils ont souvent une apparence humaine. Quand un ACA parle ou rit, il est capable de montrer de façon autonome des comportements multimodaux pour enrichir et compléter son discours prononcé et transmettre des informations qualitatives telles que ses émotions. Notre recherche utilise les modèles d’apprentissage à partir données. Un modèle de génération de comportements multimodaux pour un personnage virtuel parlant avec des émotions différentes a été proposé ainsi qu’un modèle de simulation du comportement de rire sur un ACA. Notre objectif est d'étudier et de développer des générateurs d'animation pour simuler la parole expressive et le rire d’un ACA. En partant de la relation liant prosodie de la parole et comportements multimodaux, notre générateur d'animation prend en entrée les signaux audio prononcés et fournit en sortie des comportements multimodaux. Notre travail vise à utiliser un modèle statistique pour saisir la relation entre les signaux donnés en entrée et les signaux de sortie; puis cette relation est transformée en modèle d’animation 3D. Durant l'étape d’apprentissage, le modèle statistique est entrainé à partir de paramètres communs qui sont composés de paramètres d'entrée et de sortie. La relation entre les signaux d'entrée et de sortie peut être capturée et caractérisée par les paramètres du modèle statistique. Dans l'étape de synthèse, le modèle entrainé est utilisé pour produire des signaux de sortie (expressions faciale, mouvement de tête et du torse) à partir des signaux d'entrée (F0, énergie de la parole ou pseudo-phonème du rire). La relation apprise durant la phase d'apprentissage peut être rendue dans les signaux de sortie. Notre module proposé est basé sur des variantes des modèles de Markov cachés (HMM), appelées HMM contextuels. Ce modèle est capable de capturer la relation entre les mouvements multimodaux et de la parole (ou rire); puis cette relation est rendue par l’animation de l’ACA<br>Our aim is to render expressive multimodal behaviors for Embodied conversational agents, ECAs. ECAs are entities endowed with communicative and emotional capabilities; they have human-like appearance. When an ECA is speaking or laughing, it is capable of displaying autonomously behaviors to enrich and complement the uttered speech and to convey qualitative information such as emotion. Our research lies in the data-driven approach. It focuses on generating the multimodal behaviors for a virtual character speaking with different emotions. It is also concerned with simulating laughing behavior on an ECA. Our aim is to study and to develop human-like animation generators for speaking and laughing ECA. On the basis of the relationship linking speech prosody and multimodal behaviors, our animation generator takes as input human uttered audio signals and output multimodal behaviors. Our work focuses on using statistical framework to capture the relationship between the input and the output signals; then this relationship is rendered into synthesized animation. In the training step, the statistical framework is trained based on joint features, which are composed of input and of output features. The relation between input and output signals can be captured and characterized by the parameters of the statistical framework. In the synthesis step, the trained framework is used to produce output signals (facial expression, head and torso movements) from input signals (F0, energy for speech or pseudo-phoneme of laughter). The relation captured in the training phase can be rendered into the output signals. Our proposed module is based on variants of Hidden Markov Model (HMM), called Contextual HMM. This model is capable of capturing the relationship between human motions and speech (or laughter); then such relationship is rendered into the synthesized animations
APA, Harvard, Vancouver, ISO, and other styles
49

Dahmani, Sara. "Synthèse audiovisuelle de la parole expressive : modélisation des émotions par apprentissage profond." Electronic Thesis or Diss., Université de Lorraine, 2020. http://www.theses.fr/2020LORR0137.

Full text
Abstract:
Les travaux de cette thèse portent sur la modélisation des émotions pour la synthèse audiovisuelle expressive de la parole à partir du texte. Aujourd’hui, les résultats des systèmes de synthèse de la parole à partir du texte sont de bonne qualité, toutefois la synthèse audiovisuelle reste encore une problématique ouverte et la synthèse expressive l’est encore d’avantage. Nous proposons dans le cadre de cette thèse une méthode de modélisation des émotions malléable et flexible, permettant de mélanger les émotions comme on mélange les teintes sur une palette de couleurs. Dans une première partie, nous présentons et étudions deux corpus expressifs que nous avons construits. La stratégie d’acquisition ainsi que le contenu expressif de ces corpus sont analysés pour valider leur utilisation à des fins de synthèse audiovisuelle de la parole. Dans une seconde partie, nous proposons deux architectures neuronales pour la synthèse de la parole. Nous avons utilisé ces deux architectures pour modéliser trois aspects de la parole : 1) les durées des sons, 2) la modalité acoustique et 3) la modalité visuelle. Dans un premier temps, nous avons adopté une architecture entièrement connectée. Cette dernière nous a permis d’étudier le comportement des réseaux de neurones face à différents descripteurs contextuels et linguistiques. Nous avons aussi pu analyser, via des mesures objectives, la capacité du réseau à modéliser les émotions. La deuxième architecture neuronale proposée est celle d’un auto-encodeur variationnel. Cette architecture est capable d’apprendre une représentation latente des émotions sans utiliser les étiquettes des émotions. Après analyse de l’espace latent des émotions, nous avons proposé une procédure de structuration de ce dernier pour pouvoir passer d’une représentation par catégorie vers une représentation continue des émotions. Nous avons pu valider, via des expériences perceptives, la capacité de notre système à générer des émotions, des nuances d’émotions et des mélanges d’émotions, et cela pour la synthèse audiovisuelle expressive de la parole à partir du texte<br>: The work of this thesis concerns the modeling of emotions for expressive audiovisual textto-speech synthesis. Today, the results of text-to-speech synthesis systems are of good quality, however audiovisual synthesis remains an open issue and expressive synthesis is even less studied. As part of this thesis, we present an emotions modeling method which is malleable and flexible, and allows us to mix emotions as we mix shades on a palette of colors. In the first part, we present and study two expressive corpora that we have built. The recording strategy and the expressive content of these corpora are analyzed to validate their use for the purpose of audiovisual speech synthesis. In the second part, we present two neural architectures for speech synthesis. We used these two architectures to model three aspects of speech : 1) the duration of sounds, 2) the acoustic modality and 3) the visual modality. First, we use a fully connected architecture. This architecture allowed us to study the behavior of neural networks when dealing with different contextual and linguistic descriptors. We were also able to analyze, with objective measures, the network’s ability to model emotions. The second neural architecture proposed is a variational auto-encoder. This architecture is able to learn a latent representation of emotions without using emotion labels. After analyzing the latent space of emotions, we presented a procedure for structuring it in order to move from a discrete representation of emotions to a continuous one. We were able to validate, through perceptual experiments, the ability of our system to generate emotions, nuances of emotions and mixtures of emotions, and this for expressive audiovisual text-to-speech synthesis
APA, Harvard, Vancouver, ISO, and other styles
50

Ding, Yu. "Modèle statistique de l'animation expressive de la parole et du rire pour un agent conversationnel animé." Electronic Thesis or Diss., Paris, ENST, 2014. http://www.theses.fr/2014ENST0050.

Full text
Abstract:
Notre objectif est de simuler des comportements multimodaux expressifs pour les agents conversationnels animés ACA. Ceux-ci sont des entités dotées de capacités affectives et communicationnelles; ils ont souvent une apparence humaine. Quand un ACA parle ou rit, il est capable de montrer de façon autonome des comportements multimodaux pour enrichir et compléter son discours prononcé et transmettre des informations qualitatives telles que ses émotions. Notre recherche utilise les modèles d’apprentissage à partir données. Un modèle de génération de comportements multimodaux pour un personnage virtuel parlant avec des émotions différentes a été proposé ainsi qu’un modèle de simulation du comportement de rire sur un ACA. Notre objectif est d'étudier et de développer des générateurs d'animation pour simuler la parole expressive et le rire d’un ACA. En partant de la relation liant prosodie de la parole et comportements multimodaux, notre générateur d'animation prend en entrée les signaux audio prononcés et fournit en sortie des comportements multimodaux. Notre travail vise à utiliser un modèle statistique pour saisir la relation entre les signaux donnés en entrée et les signaux de sortie; puis cette relation est transformée en modèle d’animation 3D. Durant l'étape d’apprentissage, le modèle statistique est entrainé à partir de paramètres communs qui sont composés de paramètres d'entrée et de sortie. La relation entre les signaux d'entrée et de sortie peut être capturée et caractérisée par les paramètres du modèle statistique. Dans l'étape de synthèse, le modèle entrainé est utilisé pour produire des signaux de sortie (expressions faciale, mouvement de tête et du torse) à partir des signaux d'entrée (F0, énergie de la parole ou pseudo-phonème du rire). La relation apprise durant la phase d'apprentissage peut être rendue dans les signaux de sortie. Notre module proposé est basé sur des variantes des modèles de Markov cachés (HMM), appelées HMM contextuels. Ce modèle est capable de capturer la relation entre les mouvements multimodaux et de la parole (ou rire); puis cette relation est rendue par l’animation de l’ACA<br>Our aim is to render expressive multimodal behaviors for Embodied conversational agents, ECAs. ECAs are entities endowed with communicative and emotional capabilities; they have human-like appearance. When an ECA is speaking or laughing, it is capable of displaying autonomously behaviors to enrich and complement the uttered speech and to convey qualitative information such as emotion. Our research lies in the data-driven approach. It focuses on generating the multimodal behaviors for a virtual character speaking with different emotions. It is also concerned with simulating laughing behavior on an ECA. Our aim is to study and to develop human-like animation generators for speaking and laughing ECA. On the basis of the relationship linking speech prosody and multimodal behaviors, our animation generator takes as input human uttered audio signals and output multimodal behaviors. Our work focuses on using statistical framework to capture the relationship between the input and the output signals; then this relationship is rendered into synthesized animation. In the training step, the statistical framework is trained based on joint features, which are composed of input and of output features. The relation between input and output signals can be captured and characterized by the parameters of the statistical framework. In the synthesis step, the trained framework is used to produce output signals (facial expression, head and torso movements) from input signals (F0, energy for speech or pseudo-phoneme of laughter). The relation captured in the training phase can be rendered into the output signals. Our proposed module is based on variants of Hidden Markov Model (HMM), called Contextual HMM. This model is capable of capturing the relationship between human motions and speech (or laughter); then such relationship is rendered into the synthesized animations
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!