Academic literature on the topic 'Visemes'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Visemes.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Visemes"

1

Chelali, Fatma Zohra, and Amar Djeradi. "Primary Research on Arabic Visemes, Analysis in Space and Frequency Domain." International Journal of Mobile Computing and Multimedia Communications 3, no. 4 (October 2011): 1–19. http://dx.doi.org/10.4018/jmcmc.2011100101.

Full text
Abstract:
Visemes are the unique facial positions required to produce phonemes, which are the smallest phonetic unit distinguished by the speakers of a particular language. Each language has multiple phonemes and visemes, and each viseme can have multiple phonemes. However, current literature on viseme research indicates that the mapping between phonemes and visemes is many-to-one: there are many phonemes which look alike visually, and hence they fall into the same visemic category. To evaluate the performance of the proposed method, the authors collected a large number of speech visual signal of five Algerian speakers male and female at different moments pronouncing 28 Arabic phonemes. For each frame the lip area is manually located with a rectangle of size proportional to 120*160 and centred on the mouth, and converted to gray scale. Finally, the mean and the standard deviation of the values of the pixels of the lip area are computed by using 20 images for each phoneme sequence to classify the visemes. The pitch analysis is investigated to show its variation for each viseme.
APA, Harvard, Vancouver, ISO, and other styles
2

Bear, Helen L., and Richard Harvey. "Alternative Visual Units for an Optimized Phoneme-Based Lipreading System." Applied Sciences 9, no. 18 (September 15, 2019): 3870. http://dx.doi.org/10.3390/app9183870.

Full text
Abstract:
Lipreading is understanding speech from observed lip movements. An observed series of lip motions is an ordered sequence of visual lip gestures. These gestures are commonly known, but as yet are not formally defined, as `visemes’. In this article, we describe a structured approach which allows us to create speaker-dependent visemes with a fixed number of visemes within each set. We create sets of visemes for sizes two to 45. Each set of visemes is based upon clustering phonemes, thus each set has a unique phoneme-to-viseme mapping. We first present an experiment using these maps and the Resource Management Audio-Visual (RMAV) dataset which shows the effect of changing the viseme map size in speaker-dependent machine lipreading and demonstrate that word recognition with phoneme classifiers is possible. Furthermore, we show that there are intermediate units between visemes and phonemes which are better still. Second, we present a novel two-pass training scheme for phoneme classifiers. This approach uses our new intermediary visual units from our first experiment in the first pass as classifiers; before using the phoneme-to-viseme maps, we retrain these into phoneme classifiers. This method significantly improves on previous lipreading results with RMAV speakers.
APA, Harvard, Vancouver, ISO, and other styles
3

Owens, Elmer, and Barbara Blazek. "Visemes Observed by Hearing-Impaired and Normal-Hearing Adult Viewers." Journal of Speech, Language, and Hearing Research 28, no. 3 (September 1985): 381–93. http://dx.doi.org/10.1044/jshr.2803.381.

Full text
Abstract:
A series of VCV nonsense syllables formed with 23 consonants and the vowels //, /i/, /u/, and // was presented on videotape without sound to 5 hearing-impaired adults and 5 adults with normal hearing. The two-fold purpose was (a) to determine whether the two groups would perform the same in their identification of visemes and (b) to observe whether the identification of visemes is influenced by vowel context. There were no differences between the two groups either with respect to the overall percentage of items correct or to the visemes identified. Noticeable differences occurred in viseme identification between the /u/ context and the other 3 vowel contexts; visemes with // differed slightly from those with // and /i/; and there were no differences in viseme identification for // and /i/ contexts. Findings were in general agreement with other studies with respect to the visemes identified, provided it is acknowledged that changes can occur depending on variables such as talkers, stimuli, recording and viewing conditions, training procedures, and statistical criteria. A composite grouping consists of /p,b,m/; /f,v/; /θ,ð/; /w,r/; /t∫,d,∫,/; and /t,d,s,k,n,g,l/.
APA, Harvard, Vancouver, ISO, and other styles
4

Fenghour, Souheil, Daqing Chen, Kun Guo, Bo Li, and Perry Xiao. "An Effective Conversion of Visemes to Words for High-Performance Automatic Lipreading." Sensors 21, no. 23 (November 26, 2021): 7890. http://dx.doi.org/10.3390/s21237890.

Full text
Abstract:
As an alternative approach, viseme-based lipreading systems have demonstrated promising performance results in decoding videos of people uttering entire sentences. However, the overall performance of such systems has been significantly affected by the efficiency of the conversion of visemes to words during the lipreading process. As shown in the literature, the issue has become a bottleneck of such systems where the system’s performance can decrease dramatically from a high classification accuracy of visemes (e.g., over 90%) to a comparatively very low classification accuracy of words (e.g., only just over 60%). The underlying cause of this phenomenon is that roughly half of the words in the English language are homophemes, i.e., a set of visemes can map to multiple words, e.g., “time” and “some”. In this paper, aiming to tackle this issue, a deep learning network model with an Attention based Gated Recurrent Unit is proposed for efficient viseme-to-word conversion and compared against three other approaches. The proposed approach features strong robustness, high efficiency, and short execution time. The approach has been verified with analysis and practical experiments of predicting sentences from benchmark LRS2 and LRS3 datasets. The main contributions of the paper are as follows: (1) A model is developed, which is effective in converting visemes to words, discriminating between homopheme words, and is robust to incorrectly classified visemes; (2) the model proposed uses a few parameters and, therefore, little overhead and time are required to train and execute; and (3) an improved performance in predicting spoken sentences from the LRS2 dataset with an attained word accuracy rate of 79.6%—an improvement of 15.0% compared with the state-of-the-art approaches.
APA, Harvard, Vancouver, ISO, and other styles
5

Preminger, Jill E., Hwei-Bing Lin, Michel Payen, and Harry Levitt. "Selective Visual Masking in Speechreading." Journal of Speech, Language, and Hearing Research 41, no. 3 (June 1998): 564–75. http://dx.doi.org/10.1044/jslhr.4103.564.

Full text
Abstract:
Using digital video technology, selective aspects of a face can be masked by identifying the pixels that represent it and then, by adjusting the gray levels, effectively eliminate that facial aspect. In groups of young adults with normal vision and hearing, consonant-viseme recognition was measured for closed sets of vowel-consonant-vowel disyllables. In the first experiment viseme recognition was measured while the tongue and teeth were masked and while the entire mouth was masked. The results showed that masking of the tongue and teeth had little effect on viseme recognition, and when the entire mouth was masked, participants continued to identify consonant visemes with 70% or greater accuracy in the /a/ and // vowel contexts. In the second experiment, viseme recognition was measured when the upper part of the face and the mouth were masked and when the lower part of the face and the mouth were masked. The results showed that when the mouth and the upper part of the face were masked, performance was poor, but information was available to identify the consonantviseme /f/. When the mouth and the lower part of the face were masked, viseme recognition was quite poor, but information was available to discriminate the consonant-viseme /p/ from other consonant visemes.
APA, Harvard, Vancouver, ISO, and other styles
6

De Martino, José Mario, Léo Pini Magalhães, and Fábio Violaro. "Facial animation based on context-dependent visemes." Computers & Graphics 30, no. 6 (December 2006): 971–80. http://dx.doi.org/10.1016/j.cag.2006.08.017.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Lalonde, Kaylah, and Grace A. Dwyer. "Visual phonemic knowledge and audiovisual speech-in-noise perception in school-age children." Journal of the Acoustical Society of America 153, no. 3_supplement (March 1, 2023): A337. http://dx.doi.org/10.1121/10.0019067.

Full text
Abstract:
Our mental representations of speech sounds include information about the visible articulatory gestures that accompany different speech sounds. We call this visual phonemic knowledge. This study examined development of school-age children’s visual phonemic knowledge and their ability to use visual phonemic knowledge to supplement audiovisual speech processing. Sixty-two children (5–16 years) and 18 adults (19–35 years) completed auditory-only, visual-only, and audiovisual tests of consonant-vowel syllable repetition. Auditory-only and audiovisual conditions were presented in steady-state, speech-spectrum noise at individually set SNRs. Consonant confusions were analyzed to define visemes (clusters of phonemes that are visually confusable with one another but visually distinct from other phonemes) evident in adults’ responses to visual-only consonants and to compute the proportion of errors in each participant and modality that were within adults' visemes. Children were less accurate than adults at visual-only consonant identification. However, children as young as 5 years of age demonstrated some visual phonemic knowledge. Comparison of error patterns across conditions indicated that children used visual phonemic knowledge during audiovisual speech-in-noise recognition. Details regarding the order of acquisition of viseme will be discussed.
APA, Harvard, Vancouver, ISO, and other styles
8

Lazalde, Oscar Martinez, Steve Maddock, and Michael Meredith. "A Constraint-Based Approach to Visual Speech for a Mexican-Spanish Talking Head." International Journal of Computer Games Technology 2008 (2008): 1–7. http://dx.doi.org/10.1155/2008/412056.

Full text
Abstract:
A common approach to produce visual speech is to interpolate the parameters describing a sequence of mouth shapes, known as visemes, where a viseme corresponds to a phoneme in an utterance. The interpolation process must consider the issue of context-dependent shape, or coarticulation, in order to produce realistic-looking speech. We describe an approach to such pose-based interpolation that deals with coarticulation using a constraint-based technique. This is demonstrated using a Mexican-Spanish talking head, which can vary its speed of talking and produce coarticulation effects.
APA, Harvard, Vancouver, ISO, and other styles
9

Thangthai, Ausdang, Ben Milner, and Sarah Taylor. "Synthesising visual speech using dynamic visemes and deep learning architectures." Computer Speech & Language 55 (May 2019): 101–19. http://dx.doi.org/10.1016/j.csl.2018.11.003.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Henton, Caroline. "Beyond visemes: Using disemes in synthetic speech with facial animation." Journal of the Acoustical Society of America 95, no. 5 (May 1994): 3010. http://dx.doi.org/10.1121/1.408830.

Full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Visemes"

1

Taylor, Sarah. "Discovering dynamic visemes." Thesis, University of East Anglia, 2013. https://ueaeprints.uea.ac.uk/47913/.

Full text
Abstract:
This thesis introduces a set of new, dynamic units of visual speech which are learnt using computer vision and machine learning techniques. Rather than clustering phoneme labels as is done traditionally, the visible articulators of a speaker are tracked and automatically segmented into short, visually intuitive speech gestures based on the dynamics of the articulators. The segmented gestures are clustered into dynamic visemes, such that movements relating to the same visual function appear within the same cluster. Speech animation can then be generated on any facial model by mapping a phoneme sequence to a sequence of dynamic visemes, and stitching together an example of each viseme in the sequence. Dynamic visemes model coarticulation and maintain the dynamics of the original speech, so simple blending at the concatenation boundaries ensures a smooth transition. The efficacy of dynamic visemes for computer animation is formally evaluated both objectively and subjectively, and compared with traditional phoneme to static lip-pose interpolation.
APA, Harvard, Vancouver, ISO, and other styles
2

Bear, Helen L. "Decoding visemes : improving machine lip-reading." Thesis, University of East Anglia, 2016. https://ueaeprints.uea.ac.uk/59384/.

Full text
Abstract:
This thesis is about improving machine lip-reading, that is, the classification of speech from only visual cues of a speaker. Machine lip-reading is a niche research problem in both areas of speech processing and computer vision. Current challenges for machine lip-reading fall into two groups: the content of the video, such as the rate at which a person is speaking or; the parameters of the video recording for example, the video resolution. We begin our work with a literature review to understand the restrictions current technology limits machine lip-reading recognition and conduct an experiment into resolution affects. We show that high definition video is not needed to successfully lip-read with a computer. The term 'viseme' is used in machine lip-reading to represent a visual cue or gesture which corresponds to a subgroup of phonemes where the phonemes are indistinguishable in the visual speech signal. Whilst a viseme is yet to be formally defined, we use the common working definition: 'a viseme is a group of phonemes with identical appearance on the lips'. A phoneme is the smallest acoustic unit a human can utter. Because there are more phonemes per viseme, mapping between the units creates a many-to-one relationship. Many mappings have been presented, and we conduct an experiment to determine which mapping produces the most accurate classification. Our results show Lee's [82] is best. Lee's classification also outperforms machine lip-reading systems which use the popular Fisher [48] phoneme-to-viseme map. Further to this, we propose three methods of deriving speaker-dependent phoneme-to-viseme maps and compare our new approaches to Lee's. Our results show the sensitivity of phoneme clustering and we use our new knowledge for our first suggested augmentation to the conventional lip-reading system. Speaker independence in machine lip-reading classification is another unsolved obstacle. It has been observed, in the visual domain, that classifiers need training on the test subject to achieve the best classification. Thus machine lip-reading is highly dependent upon the speaker. Speaker independence is the opposite of this, or in other words, is the classification of a speaker not present in the classifier's training data. We investigate the dependence of phoneme-to-viseme maps between speakers. Our results show there is not a high variability of visual cues, but there is high variability in trajectory between visual cues of an individual speaker with the same ground truth. This implies a dependency upon the number of visemes within each set for each individual. Finally, we investigate how many visemes is the optimum number within a set. We show the phoneme-to-viseme maps in literature rarely have enough visemes and the optimal number, which varies by speaker, ranges from 11 to 35. The last difficulty we address is decoding from visemes back to phonemes and into words. Traditionally this is completed using a language model. The language model unit is either: the same as the classifier, e.g. visemes or phonemes; or the language model unit is words. In a novel approach we use these optimum range viseme sets within hierarchical training of phoneme labelled classifiers. This new method of classifier training demonstrates significant increase in classification with a word language network.
APA, Harvard, Vancouver, ISO, and other styles
3

Ramage, Matthew David. "Disproving visemes as the basic visual unit of speech." Thesis, Curtin University, 2013. http://hdl.handle.net/20.500.11937/1618.

Full text
Abstract:
Visemes are groups of phonemes that are considered visually indistinguishable, but there are many doubts regarding their use in visual speech recognition. In this thesis, a visual speech recogniser is constructed to test the validity of visemes. Examining the phoneme output of the recogniser shows that it is not possible to construct a viseme grouping that exhibits the required phoneme confusion characteristics. This thesis proves that phonemes, not visemes, are the basic visual unit of speech.
APA, Harvard, Vancouver, ISO, and other styles
4

Thangthai, Ausdang. "Visual speech synthesis using dynamic visemes and deep learning architectures." Thesis, University of East Anglia, 2018. https://ueaeprints.uea.ac.uk/69371/.

Full text
Abstract:
The aim of this work is to improve the naturalness of visual speech synthesis produced automatically from a linguistic input over existing methods. Firstly, the most important contribution is on the investigation of the most suitable speech units for the visual speech synthesis. We propose the use of dynamic visemes instead of phonemes or static visemes and found that dynamic visemes can generate better visual speech than either phone or static viseme units. Moreover, best performance is obtained by a combined phoneme-dynamic viseme system. Secondly, we examine the most appropriate model between hidden Markov model (HMM) and different deep learning models that include feedforward and recurrent structures consisting of one-to-one, many-to-one and many-to-many architectures. Results suggested that that frame-by-frame synthesis from deep learning approach outperforms state-based synthesis from HMM approaches and an encoder-decoder many-to-many architecture is better than the one-to-one and many-to-one architectures. Thirdly, we explore the importance of contextual features that include information at varying linguistic levels, from frame level up to the utterance level. Our findings found that frame level information is the most valuable feature, as it is able to avoid discontinuities in the visual feature sequence and produces a smooth and realistic animation output. Fourthly, we found that the two most common objective measures of correlation and root mean square error are not able to indicate realism and naturalness of human perceived quality. We introduce an alternative objective measure and show that the global variance is a better indicator of human perception of quality. Finally, we propose a novel method to convert a given text input and phoneme transcription into a dynamic viseme transcription in the case when a reference dynamic viseme sequence is not available. Subjective preference tests confirmed that our proposed method is able to produce animation, that are statistically indistinguishable from animation produced using reference data.
APA, Harvard, Vancouver, ISO, and other styles
5

Zailskas, Vytautas. "Lietuvių šnekos vizemų vizualizavimas." Master's thesis, Lithuanian Academic Libraries Network (LABT), 2011. http://vddb.laba.lt/obj/LT-eLABa-0001:E.02~2011~D_20110615_134342-38011.

Full text
Abstract:
Magistro baigiamajame darbe analizuojama lietuvių šnekos vizemų vizualizacijos problema, tiriami lietuvių kalbos vizemų vizualizacijos požymiai, galimybės koduoti SAMPA kodais analizė, analizuojami programinės įrangos kūrimo metodai ir algoritmai. Sukurtas algoritmas sprendžiantis lietuvių šnekos vizemų vizualizacijos problemą. Pasirinktas kompiuterinės grafikos tipas, tinkantis iškeltiems tikslams įvykdyti – vektorinė grafika. Sukurti du vektorių transformacijos metodai, išanalizuoti jų skirtumai ir praktinio panaudojimo galimybės. Sukurta programinė įranga įgalinanti vartotoją kurti vizemas, jas transformuoti ir derinti jų vaizdavimo trukmę įvairiais koeficientais ir vykdyti animaciją pagal pasirinktą transformacijos metodą. Sudarytos penkios vizemos ir dviejų lietuviškų žodžių animacijos, kurių pagalba atliktas tyrimas parodantis darbo ir metodų realizavimo kokybę bei pritaikomumą.
In this final Master's thesis work features of Lithuanian speech visemes visualization are analyzed. Possibility of coding with SAMPA codes, software methods and algorithms are inspected. Type of computer graphics is picked, which is suitable for software objectives – vector graphics. Two transformation methods for vector graphics are created and their differences and practical usability are analyzed. Software for visemes creation, transformation and tuning of their duration and duration of transformation between visemes is created and described. The main purpose of this software is to animate Lithuanian speech by the method selected. Five visemes for two Lithuanian words animation is created. Using these visemes research has been done which is showing the quality of realization and adaptability of this software.
APA, Harvard, Vancouver, ISO, and other styles
6

Martinez, Lazalde Oscar Manuel. "Analyzing and evaluating the use of visemes in an interpolative synthesizer for visual speech." Thesis, University of Sheffield, 2010. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.531167.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Axelsson, Andreas, and Erik Björhäll. "Real Time Speech Driven Face Animation." Thesis, Linköping University, Department of Electrical Engineering, 2003. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-2015.

Full text
Abstract:

The goal of this project is to implement a system to analyse an audio signal containing speech, and produce a classifcation of lip shape categories (visemes) in order to synchronize the lips of a computer generated face with the speech.

The thesis describes the work to derive a method that maps speech to lip move- ments, on an animated face model, in real time. The method is implemented in C++ on the PC/Windows platform. The program reads speech from pre-recorded audio files and continuously performs spectral analysis of the speech. Neural networks are used to classify the speech into a sequence of phonemes, and the corresponding visemes are shown on the screen.

Some time delay between input speech and the visualization could not be avoided, but the overall visual impression is that sound and animation are synchronized.

APA, Harvard, Vancouver, ISO, and other styles
8

Akdemir, Eren. "Bimodal Automatic Speech Segmentation And Boundary Refinement Techniques." Phd thesis, METU, 2010. http://etd.lib.metu.edu.tr/upload/3/12611732/index.pdf.

Full text
Abstract:
Automatic segmentation of speech is compulsory for building large speech databases to be used in speech processing applications. This study proposes a bimodal automatic speech segmentation system that uses either articulator motion information (AMI) or visual information obtained by a camera in collaboration with auditory information. The presence of visual modality is shown to be very beneficial in speech recognition applications, improving the performance and noise robustness of those systems. In this dissertation a significant increase in the performance of the automatic speech segmentation system is achieved by using a bimodal approach. Automatic speech segmentation systems have a tradeoff between precision and resulting number of gross errors. Boundary refinement techniques are used in order to increase precision of these systems without decreasing the system performance. Two novel boundary refinement techniques are proposed in this thesis
a hidden Markov model (HMM) based fine tuning system and an inverse filtering based fine tuning system. The segment boundaries obtained by the bimodal speech segmentation system are improved further by using these techniques. To fulfill these goals, a complete two-stage automatic speech segmentation system is produced and tested in two different databases. A phonetically rich Turkish audiovisual speech database, that contains acoustic data and camera recordings of 1600 Turkish sentences uttered by a male speaker, is build from scratch in order to be used in the experiments. The visual features of the recordings are extracted and manual phonetic alignment of the database is done to be used as a ground truth for the performance tests of the automatic speech segmentation systems.
APA, Harvard, Vancouver, ISO, and other styles
9

Melenchón, Maldonado Javier. "Síntesis Audiovisual Realista Personalizable." Doctoral thesis, Universitat Ramon Llull, 2007. http://hdl.handle.net/10803/9133.

Full text
Abstract:
Es presenta un esquema únic per a la síntesi i anàlisi audiovisual personalitzable realista de seqüències audiovisuals de cares parlants i seqüències visuals de llengua de signes en àmbit domèstic. En el primer cas, amb animació totalment sincronitzada a través d'una font de text o veu; en el segon, utilitzant la tècnica de lletrejar paraules mitjançant la ma. Les seves possibilitats de personalització faciliten la creació de seqüències audiovisuals per part d'usuaris no experts. Les aplicacions possibles d'aquest esquema de síntesis comprenen des de la creació de personatges virtuals realistes per interacció natural o vídeo jocs fins vídeo conferència des de molt baix ample de banda i telefonia visual per a les persones amb problemes d'oïda, passant per oferir ajuda a la pronunciació i la comunicació a aquest mateix col·lectiu. El sistema permet processar seqüències llargues amb un consum de recursos molt reduït, sobre tot, en el referent a l'emmagatzematge, gràcies al desenvolupament d'un nou procediment de càlcul incremental per a la descomposició en valors singulars amb actualització de la informació mitja. Aquest procediment es complementa amb altres tres: el decremental, el de partició i el de composició.
Se presenta un esquema único para la síntesis y análisis audiovisual personalizable realista de secuencias audiovisuales de caras parlantes y secuencias visuales de lengua de signos en entorno doméstico. En el primer caso, con animación totalmente sincronizada a través de una fuente de texto o voz; en el segundo, utilizando la técnica de deletreo de palabras mediante la mano. Sus posibilidades de personalización facilitan la creación de secuencias audiovisuales por parte de usuarios no expertos. Las aplicaciones posibles de este esquema de síntesis comprenden desde la creación de personajes virtuales realistas para interacción natural o vídeo juegos hasta vídeo conferencia de muy bajo ancho de banda y telefonía visual para las personas con problemas de oído, pasando por ofrecer ayuda en la pronunciación y la comunicación a este mismo colectivo. El sistema permite procesar secuencias largas con un consumo de recursos muy reducido gracias al desarrollo de un nuevo procedimiento de cálculo incremental para la descomposición en valores singulares con actualización de la información media.
A shared framework for realistic and personalizable audiovisual synthesis and analysis of audiovisual sequences of talking heads and visual sequences of sign language is presented in a domestic environment. The former has full synchronized animation using a text or auditory source of information; the latter consists in finger spelling. Their personalization capabilities ease the creation of audiovisual sequences by non expert users. The applications range from realistic virtual avatars for natural interaction or videogames to low bandwidth videoconference and visual telephony for the hard of hearing, including help to speech therapists. Long sequences can be processed with reduced resources, specially storing ones. This is allowed thanks to the proposed scheme for the incremental singular value decomposition with mean preservation. This scheme is complemented with another three: the decremental, the split and the composed ones.
APA, Harvard, Vancouver, ISO, and other styles
10

Turkmani, Aseel. "Visual analysis of viseme dynamics." Thesis, University of Surrey, 2008. http://epubs.surrey.ac.uk/804944/.

Full text
Abstract:
Face-to-face dialogue is the most natural mode of communication between humans. The combination of human visual perception of expression and perception in changes in intonation provides semantic information that communicates idea, feelings and concepts. The realistic modelling of speech movements, through automatic facial animation, and maintaining audio-visual coherence is still a challenge in both the computer graphics and film industry.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Books on the topic "Visemes"

1

Velázquez, Cuauhtémoc J. Allende. Tornillo de banco. [Oaxaca, Mexico]: Instituto Politécnico Nacional, Centro Interdisciplinario y de Investigación para el Desarrollo Integral Regional, Unidad Oaxaca, 1988.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
2

Murvai, Olga. Viseltes szavak: Hoc est transsylvanicum : tanulmánygyűjtemény. Csíkszereda: Pallas-Akadémia, 2006.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
3

Brodin, Elin. De vises sten. [Oslo]: Dreyer, 1989.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
4

L' architecture du IIIe Reich: Origines intellectuelles et visées idéologiques. Berne: P. Lang, 1987.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
5

Legrand-Blain, Marie. Spiriferacea (Brachiopoda) viséens et serpukhoviens du Sahara algérien. Brest: Université de Bretagne occidentale, 1986.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
6

Brochu, André. La visée critique: Essais autobiographiques et littéraires. Montréal: Boréal, 1988.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
7

Demografia e viseve shqiptare në Greqinë e veriut. Tetovë: Luma, 2013.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
8

Malomfălean, Laurențiu. Nocturnalul "postmodern": Când visele imaginează hiperindivizi cu onirografii. Bucureşti: Tracus Arte, 2015.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
9

Per, Aslaksen, ed. De trykte illegale visene: Hvordan ble de produsert? Skien: Falken Forlag, 1988.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
10

Rowling, J. K. Harry Potter och de vises sten. 2nd ed. Stockholm, Sweden: Tiden, 2001.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Book chapters on the topic "Visemes"

1

Visser, Michiel, Mannes Poel, and Anton Nijholt. "Classifying Visemes for Automatic Lipreading." In Text, Speech and Dialogue, 349–52. Berlin, Heidelberg: Springer Berlin Heidelberg, 1999. http://dx.doi.org/10.1007/3-540-48239-3_65.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Shdaifat, Islam, Rolf-Rainer Grigat, and Stefan Lütgert. "Recognition of the German Visemes Using Multiple Feature Matching." In Lecture Notes in Computer Science, 437–42. Berlin, Heidelberg: Springer Berlin Heidelberg, 2001. http://dx.doi.org/10.1007/3-540-45404-7_58.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Pajorová, Eva, and Ladislav Hluchý. "Correct Speech Visemes as a Root of Total Communication Method for Deaf People." In Agent and Multi-Agent Systems. Technologies and Applications, 389–95. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-30947-2_43.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Leszczynski, Mariusz, and Władysław Skarbek. "Viseme Classification for Talking Head Application." In Computer Analysis of Images and Patterns, 773–80. Berlin, Heidelberg: Springer Berlin Heidelberg, 2005. http://dx.doi.org/10.1007/11556121_95.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Leszczynski, Mariusz, Władysław Skarbek, and Stanisław Badura. "Fast Viseme Recognition for Talking Head Application." In Lecture Notes in Computer Science, 516–23. Berlin, Heidelberg: Springer Berlin Heidelberg, 2005. http://dx.doi.org/10.1007/11559573_64.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Revéret, Lionel, and Christian Benoît. "A viseme-based approach to labiometrics for automatic lipreading." In Audio- and Video-based Biometric Person Authentication, 335–42. Berlin, Heidelberg: Springer Berlin Heidelberg, 1997. http://dx.doi.org/10.1007/bfb0016013.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Lee, Soonkyu, and Dongsuk Yook. "Viseme Recognition Experiment Using Context Dependent Hidden Markov Models." In Intelligent Data Engineering and Automated Learning — IDEAL 2002, 557–61. Berlin, Heidelberg: Springer Berlin Heidelberg, 2002. http://dx.doi.org/10.1007/3-540-45675-9_84.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Bastanfard, Azam, Mohammad Aghaahmadi, Alireza Abdi kelishami, Maryam Fazel, and Maedeh Moghadam. "Persian Viseme Classification for Developing Visual Speech Training Application." In Advances in Multimedia Information Processing - PCM 2009, 1080–85. Berlin, Heidelberg: Springer Berlin Heidelberg, 2009. http://dx.doi.org/10.1007/978-3-642-10467-1_104.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Koller, Oscar, Hermann Ney, and Richard Bowden. "Read My Lips: Continuous Signer Independent Weakly Supervised Viseme Recognition." In Computer Vision – ECCV 2014, 281–96. Cham: Springer International Publishing, 2014. http://dx.doi.org/10.1007/978-3-319-10590-1_19.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Fernandez-Lopez, Adriana, and Federico M. Sukno. "Optimizing Phoneme-to-Viseme Mapping for Continuous Lip-Reading in Spanish." In Communications in Computer and Information Science, 305–28. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-12209-6_15.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Visemes"

1

Bear, Helen L., and Richard Harvey. "Decoding visemes: Improving machine lip-reading." In 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2016. http://dx.doi.org/10.1109/icassp.2016.7472029.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Santhosh Kumar, S. "Encoding Malayalam visemes for facial image synthesis." In IET Conference on Wireless, Mobile and Multimedia Networks. IEE, 2008. http://dx.doi.org/10.1049/cp:20080195.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Hui Zhao and Chaojing Tang. "Visual speech synthesis based on Chinese dynamic visemes." In 2008 International Conference on Information and Automation (ICIA). IEEE, 2008. http://dx.doi.org/10.1109/icinfa.2008.4607983.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Costa, Paula Dornhofer Paro, and José Mario De Martino. "Compact 2D facial animation based on context-dependent visemes." In the ACM/SSPNET 2nd International Symposium. New York, New York, USA: ACM Press, 2010. http://dx.doi.org/10.1145/1924035.1924047.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Thangthai, Ausdang, Ben Milner, and Sarah Taylor. "Visual Speech Synthesis Using Dynamic Visemes, Contextual Features and DNNs." In Interspeech 2016. ISCA, 2016. http://dx.doi.org/10.21437/interspeech.2016-1084.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Okita, Shinsuke, Yasue Mitsukura, and Nozomu Hamada. "Lip reading system using novel Japanese visemes classification and hierarchical weighted discrimination." In 2013 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS). IEEE, 2013. http://dx.doi.org/10.1109/ispacs.2013.6704608.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Okita, Shinsuke, Yasue Mitsukura, and Nozomu Hamada. "Augmented classification of Japanese visemes and hierarchical weighted discrimination for visual speech recognition." In 2013 IEEE Conference on Systems, Process & Control (ICSPC). IEEE, 2013. http://dx.doi.org/10.1109/spc.2013.6735104.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Xie Lei, Jiang Dongmei, I. Ravyse, Zhao Rongchun, W. Verhelst, H. Sahli, and J. Conlenis. "Visualize speech: a continuous speech recognition system for facial animation using acoustic visemes." In Proceedings of 2003 International Conference on Neural Networks and Signal Processing. IEEE, 2003. http://dx.doi.org/10.1109/icnnsp.2003.1280738.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Dehshibi, Mohammad Mahdi, Meysam Alavi, and Jamshid Shanbehzadeh. "Kernel-based Persian viseme clustering." In 2013 13th International Conference on Hybrid Intelligent Systems (HIS). IEEE, 2013. http://dx.doi.org/10.1109/his.2013.6920468.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Roslan, Rosniza, Nursuriati Jamil, Noraini Seman, and Syafiqa Ain Alfida Abdul Rahim. "Face and mouth localization of viseme." In 2016 IEEE Industrial Electronics and Applications Conference (IEACon). IEEE, 2016. http://dx.doi.org/10.1109/ieacon.2016.8067385.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Visemes"

1

Adhikari, Kamal, Bharat Adhikari, Sue Cavill, Santosh Mehrotra, Vijeta Rao Bejjanki, and Matteus Van Der Velden. Monitoria de Campanhas de Saneamento: Metas, Relatórios e Informação, e Realismo. Institute of Development Studies, June 2022. http://dx.doi.org/10.19088/slh.2022.008.

Full text
Abstract:
Muitos governos da Ásia e de África definiram prazos ambiciosos para os seus países se tornarem livres de fecalismo a céu aberto (ODF). Alguns concluíram recentemente campanhas nacionais de saneamento; vários países têm campanhas em curso; enquanto outros estão no processo de conceptualização e planificação. Monitorar resultados e dar conta deles é um dos principais desafios que se colocam a essas campanhas. Este número de Fronteiras do Saneamento apresenta as lições aprendidas até agora para que estas possam ser usadas nas campanhas governamentais em curso e a realizar que visem acabar com o fecalismo a céu aberto e melhorar o acesso ao saneamento gerido com segurança.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography