Academic literature on the topic 'Perception/McGyrk effect'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Perception/McGyrk effect.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Perception/McGyrk effect"

1

Möttönen, Riikka, Kaisa Tiippana, Mikko Sams, and Hanna Puharinen. "Sound Location Can Influence Audiovisual Speech Perception When Spatial Attention Is Manipulated." Seeing and Perceiving 24, no. 1 (2011): 67–90. http://dx.doi.org/10.1163/187847511x557308.

Full text
Abstract:
AbstractAudiovisual speech perception has been considered to operate independent of sound location, since the McGurk effect (altered auditory speech perception caused by conflicting visual speech) has been shown to be unaffected by whether speech sounds are presented in the same or different location as a talking face. Here we show that sound location effects arise with manipulation of spatial attention. Sounds were presented from loudspeakers in five locations: the centre (location of the talking face) and 45°/90° to the left/right. Auditory spatial attention was focused on a location by presenting the majority (90%) of sounds from this location. In Experiment 1, the majority of sounds emanated from the centre, and the McGurk effect was enhanced there. In Experiment 2, the major location was 90° to the left, causing the McGurk effect to be stronger on the left and centre than on the right. Under control conditions, when sounds were presented with equal probability from all locations, the McGurk effect tended to be stronger for sounds emanating from the centre, but this tendency was not reliable. Additionally, reaction times were the shortest for a congruent audiovisual stimulus, and this was the case independent of location. Our main finding is that sound location can modulate audiovisual speech perception, and that spatial attention plays a role in this modulation.
APA, Harvard, Vancouver, ISO, and other styles
2

Magnotti, John F., Debshila Basu Mallick, and Michael S. Beauchamp. "Reducing Playback Rate of Audiovisual Speech Leads to a Surprising Decrease in the McGurk Effect." Multisensory Research 31, no. 1-2 (2018): 19–38. http://dx.doi.org/10.1163/22134808-00002586.

Full text
Abstract:
We report the unexpected finding that slowing video playback decreases perception of the McGurk effect. This reduction is counter-intuitive because the illusion depends on visual speech influencing the perception of auditory speech, and slowing speech should increase the amount of visual information available to observers. We recorded perceptual data from 110 subjects viewing audiovisual syllables (either McGurk or congruent control stimuli) played back at one of three rates: the rate used by the talker during recording (the natural rate), a slow rate (50% of natural), or a fast rate (200% of natural). We replicated previous studies showing dramatic variability in McGurk susceptibility at the natural rate, ranging from 0–100% across subjects and from 26–76% across the eight McGurk stimuli tested. Relative to the natural rate, slowed playback reduced the frequency of McGurk responses by 11% (79% of subjects showed a reduction) and reduced congruent accuracy by 3% (25% of subjects showed a reduction). Fast playback rate had little effect on McGurk responses or congruent accuracy. To determine whether our results are consistent with Bayesian integration, we constructed a Bayes-optimal model that incorporated two assumptions: individuals combine auditory and visual information according to their reliability, and changing playback rate affects sensory reliability. The model reproduced both our findings of large individual differences and the playback rate effect. This work illustrates that surprises remain in the McGurk effect and that Bayesian integration provides a useful framework for understanding audiovisual speech perception.
APA, Harvard, Vancouver, ISO, and other styles
3

Omata, Kei, and Ken Mogi. "Fusion and combination in audio-visual integration." Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 464, no. 2090 (November 27, 2007): 319–40. http://dx.doi.org/10.1098/rspa.2007.1910.

Full text
Abstract:
Language is essentially multi-modal in its sensory origin, the daily conversation depending heavily on the audio-visual (AV) information. Although the perception of spoken language is primarily dominated by audition, the perception of facial expression, particularly that of the mouth, helps us comprehend speech. The McGurk effect is a striking phenomenon where the perceived phoneme is affected by the simultaneous observation of lip movement, and probably reflects the underlying AV integration process. The elucidation of the principles involved in this unique perceptual anomaly poses an interesting problem. Here we study the nature of the McGurk effect by means of neural networks (self-organizing maps, SOM) designed to extract patterns inherent in audio and visual stimuli. It is shown that a McGurk effect-like classification of incoming information occurs without any additional constraint or procedure added to the network, suggesting that the anomaly is a consequence of the AV integration process. Within this framework, an explanation is given for the asymmetric effect of AV pairs in causing the McGurk effect (fusion or combination) based on the ‘distance’ relationship between audio or visual information within the SOM. Our result reveals some generic features of the cognitive process of phoneme perception, and AV sensory integration in general.
APA, Harvard, Vancouver, ISO, and other styles
4

Alsius, Agnès, Martin Paré, and Kevin G. Munhall. "Forty Years After Hearing Lips and Seeing Voices: the McGurk Effect Revisited." Multisensory Research 31, no. 1-2 (2018): 111–44. http://dx.doi.org/10.1163/22134808-00002565.

Full text
Abstract:
Since its discovery 40 years ago, the McGurk illusion has been usually cited as a prototypical paradigmatic case of multisensory binding in humans, and has been extensively used in speech perception studies as a proxy measure for audiovisual integration mechanisms. Despite the well-established practice of using the McGurk illusion as a tool for studying the mechanisms underlying audiovisual speech integration, the magnitude of the illusion varies enormously across studies. Furthermore, the processing of McGurk stimuli differs from congruent audiovisual processing at both phenomenological and neural levels. This questions the suitability of this illusion as a tool to quantify the necessary and sufficient conditions under which audiovisual integration occurs in natural conditions. In this paper, we review some of the practical and theoretical issues related to the use of the McGurk illusion as an experimental paradigm. We believe that, without a richer understanding of the mechanisms involved in the processing of the McGurk effect, experimenters should be really cautious when generalizing data generated by McGurk stimuli to matching audiovisual speech events.
APA, Harvard, Vancouver, ISO, and other styles
5

MacDonald, John. "Hearing Lips and Seeing Voices: the Origins and Development of the ‘McGurk Effect’ and Reflections on Audio–Visual Speech Perception Over the Last 40 Years." Multisensory Research 31, no. 1-2 (2018): 7–18. http://dx.doi.org/10.1163/22134808-00002548.

Full text
Abstract:
In 1976 Harry McGurk and I published a paper in Nature, entitled ‘Hearing Lips and Seeing Voices’. The paper described a new audio–visual illusion we had discovered that showed the perception of auditorily presented speech could be influenced by the simultaneous presentation of incongruent visual speech. This hitherto unknown effect has since had a profound impact on audiovisual speech perception research. The phenomenon has come to be known as the ‘McGurk effect’, and the original paper has been cited in excess of 4800 times. In this paper I describe the background to the discovery of the effect, the rationale for the generation of the initial stimuli, the construction of the exemplars used and the serendipitous nature of the finding. The paper will also cover the reaction (and non-reaction) to the Nature publication, the growth of research on, and utilizing the ‘McGurk effect’ and end with some reflections on the significance of the finding.
APA, Harvard, Vancouver, ISO, and other styles
6

Lüttke, Claudia S., Alexis Pérez-Bellido, and Floris P. de Lange. "Rapid recalibration of speech perception after experiencing the McGurk illusion." Royal Society Open Science 5, no. 3 (March 2018): 170909. http://dx.doi.org/10.1098/rsos.170909.

Full text
Abstract:
The human brain can quickly adapt to changes in the environment. One example is phonetic recalibration: a speech sound is interpreted differently depending on the visual speech and this interpretation persists in the absence of visual information. Here, we examined the mechanisms of phonetic recalibration. Participants categorized the auditory syllables /aba/ and /ada/, which were sometimes preceded by the so-called McGurk stimuli (in which an /aba/ sound, due to visual /aga/ input, is often perceived as ‘ada’). We found that only one trial of exposure to the McGurk illusion was sufficient to induce a recalibration effect, i.e. an auditory /aba/ stimulus was subsequently more often perceived as ‘ada’. Furthermore, phonetic recalibration took place only when auditory and visual inputs were integrated to ‘ada’ (McGurk illusion). Moreover, this recalibration depended on the sensory similarity between the preceding and current auditory stimulus. Finally, signal detection theoretical analysis showed that McGurk-induced phonetic recalibration resulted in both a criterion shift towards /ada/ and a reduced sensitivity to distinguish between /aba/ and /ada/ sounds. The current study shows that phonetic recalibration is dependent on the perceptual integration of audiovisual information and leads to a perceptual shift in phoneme categorization.
APA, Harvard, Vancouver, ISO, and other styles
7

Lindborg, Alma, and Tobias S. Andersen. "Bayesian binding and fusion models explain illusion and enhancement effects in audiovisual speech perception." PLOS ONE 16, no. 2 (February 19, 2021): e0246986. http://dx.doi.org/10.1371/journal.pone.0246986.

Full text
Abstract:
Speech is perceived with both the ears and the eyes. Adding congruent visual speech improves the perception of a faint auditory speech stimulus, whereas adding incongruent visual speech can alter the perception of the utterance. The latter phenomenon is the case of the McGurk illusion, where an auditory stimulus such as e.g. “ba” dubbed onto a visual stimulus such as “ga” produces the illusion of hearing “da”. Bayesian models of multisensory perception suggest that both the enhancement and the illusion case can be described as a two-step process of binding (informed by prior knowledge) and fusion (informed by the information reliability of each sensory cue). However, there is to date no study which has accounted for how they each contribute to audiovisual speech perception. In this study, we expose subjects to both congruent and incongruent audiovisual speech, manipulating the binding and the fusion stages simultaneously. This is done by varying both temporal offset (binding) and auditory and visual signal-to-noise ratio (fusion). We fit two Bayesian models to the behavioural data and show that they can both account for the enhancement effect in congruent audiovisual speech, as well as the McGurk illusion. This modelling approach allows us to disentangle the effects of binding and fusion on behavioural responses. Moreover, we find that these models have greater predictive power than a forced fusion model. This study provides a systematic and quantitative approach to measuring audiovisual integration in the perception of the McGurk illusion as well as congruent audiovisual speech, which we hope will inform future work on audiovisual speech perception.
APA, Harvard, Vancouver, ISO, and other styles
8

Lu, Hong, and Chaochao Pan. "The McGurk effect in self-recognition of people with schizophrenia." Social Behavior and Personality: an international journal 48, no. 6 (June 2, 2020): 1–8. http://dx.doi.org/10.2224/sbp.9219.

Full text
Abstract:
The McGurk effect is a robust illusion phenomenon in the perception of speech; however, there is little research on its demonstration in nonverbal domains. Thus, we tested for the McGurk effect in the context of self-recognition. We presented a group of people with schizophrenia and a control group of people without mental illnesses, with 2 videos accompanied by a soundtrack featuring different identity information. The first video had a matched face and voice; the other featured conflicting face–voice information. The participants judged if the voice in the video was their own or someone else's. The results show there was a robust McGurk effect in self-recognition, which was stronger among participants with schizophrenia because of the influence of self-disorder. Further, people with schizophrenia were less accurate in voice self-recognition when there was conflicting face–voice identity information. Thus, presenting audiovisual-consistent information is conducive to information processing for people with schizophrenia.
APA, Harvard, Vancouver, ISO, and other styles
9

Walker, Grant M., Patrick Sarahan Rollo, Nitin Tandon, and Gregory Hickok. "Effect of Bilateral Opercular Syndrome on Speech Perception." Neurobiology of Language 2, no. 3 (2021): 335–53. http://dx.doi.org/10.1162/nol_a_00037.

Full text
Abstract:
Abstract Speech perception ability and structural neuroimaging were investigated in two cases of bilateral opercular syndrome. Due to bilateral ablation of the motor control center for the lower face and surrounds, these rare cases provide an opportunity to evaluate the necessity of cortical motor representations for speech perception, a cornerstone of some neurocomputational theories of language processing. Speech perception, including audiovisual integration (i.e., the McGurk effect), was mostly unaffected in these cases, although verbal short-term memory impairment hindered performance on several tasks that are traditionally used to evaluate speech perception. The results suggest that the role of the cortical motor system in speech perception is context-dependent and supplementary, not inherent or necessary.
APA, Harvard, Vancouver, ISO, and other styles
10

Sams, M. "Audiovisual Speech Perception." Perception 26, no. 1_suppl (August 1997): 347. http://dx.doi.org/10.1068/v970029.

Full text
Abstract:
Persons with hearing loss use visual information from articulation to improve their speech perception. Even persons with normal hearing utilise visual information, especially when the stimulus-to-noise ratio is poor. A dramatic demonstration of the role of vision in speech perception is the audiovisual fusion called the ‘McGurk effect’. When the auditory syllable /pa/ is presented in synchrony with the face articulating the syllable /ka/, the subject usually perceives /ta/ or /ka/. The illusory perception is clearly auditory in nature. We recently studied the audiovisual fusion (acoustical /p/, visual /k/) for Finnish (1) syllables, and (2) words. Only 3% of the subjects perceived the syllables according to the acoustical input, ie in 97% of the subjects the perception was influenced by the visual information. For words the percentage of acoustical identifications was 10%. The results demonstrate a very strong influence of visual information of articulation in face-to-face speech perception. Word meaning and sentence context have a negligible influence on the fusion. We have also recorded neuromagnetic responses of the human cortex when the subjects both heard and saw speech. Some subjects showed a distinct response to a ‘McGurk’ stimulus. The response was rather late, emerging about 200 ms from the onset of the auditory stimulus. We suggest that the perisylvian cortex, close to the source area for the auditory 100 ms response (M100), may be activated by the discordant stimuli. The behavioural and neuromagnetic results suggest a precognitive audiovisual speech integration occurring at a relatively early processing level.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Perception/McGyrk effect"

1

Huyse, Aurélie. "Intégration audio-visuelle de la parole: le poids de la vision varie-t-il en fonction de l'âge et du développement langagier?" Doctoral thesis, Universite Libre de Bruxelles, 2012. http://hdl.handle.net/2013/ULB-DIPOT:oai:dipot.ulb.ac.be:2013/209690.

Full text
Abstract:
Pour percevoir la parole, le cerveau humain utilise les informations sensorielles provenant non seulement de la modalité auditive mais également de la modalité visuelle. En effet, de précédentes recherches ont mis en évidence l’importance de la lecture labiale dans la perception de la parole, en montrant sa capacité à améliorer et à modifier celle-ci. C’est ce que l’on appelle l’intégration audio-visuelle de la parole. L’objectif de cette thèse de doctorat était d’étudier la possibilité de faire varier ce processus d’intégration en fonction de différentes variables. Ce travail s’inscrit ainsi au cœur d’un débat régnant depuis plusieurs années, celui opposant l’hypothèse d’une intégration audio-visuelle universelle à l’hypothèse d’une intégration dépendante du contexte. C’est dans ce cadre que nous avons réalisé les cinq études constituant cette thèse, chacune d’elles investiguant l’impact d’une variable bien précise sur l’intégration bimodale :la qualité du signal visuel, l’âge des participants, le fait de porter un implant cochléaire, l’âge au moment de l’implantation cochléaire et le fait d’avoir des troubles spécifiques du langage.

Le paradigme expérimental utilisé consistait toujours en une tâche d’identification de syllabes présentées dans trois modalités :auditive seule, visuelle seule et audio-visuelle (congruente et incongruente). Les cinq études avaient également comme point commun la présentation de stimuli visuels dont la qualité était réduite, visant à empêcher une lecture labiale de bonne qualité. Le but de chacune de ces études était non seulement d’examiner si les performances variaient en fonction des variables investiguées mais également de déterminer si les différences provenaient bien du processus d’intégration lui-même et non uniquement de différences au niveau de la perception unimodale. Pour cela, les scores des participants ont été comparés à des scores prédits sur base d’un modèle prenant en compte les variations individuelles des poids auditifs et visuels, le weighted fuzzy-logical model of perception.

L’ensemble des résultats, discuté dans la dernière partie de ce travail, fait pencher la balance en faveur de l’hypothèse d’une intégration dépendante du contexte. Nous proposons alors une nouvelle architecture de fusion bimodale, prenant en compte ces dernières données. Enfin, les implications sont aussi d’ordre pratique, suggérant la nécessité d’incorporer des évaluations et rééducations à la fois auditives et visuelles dans le cadre des programmes de revalidation de personnes âgées, dysphasiques ou avec implant cochléaire./During face-to-face conversation, perception of auditory speech is influenced by the visual speech cues contained in lip movements. Indeed, previous research has highlighted the ability of lip-reading to enhance and even modify speech perception. This phenomenon is known as audio-visual integration. The aim of this doctoral thesis is to study the possibility of modifying this audio-visual integration according to several variables. This work lies into the scope of an important debate between invariant versus subject-dependent audio-visual integration in speech processing. Each study of this dissertation investigates the impact of a specific variable on bimodal integration: the quality of the visual input, age of participants, the use of a cochlear implant, age at cochlear implantation and the presence of specific language impairments.

The paradigm used always consisted of a syllable identification task, where syllables were presented in three modalities: auditory only, visual only and audio-visual (congruent and incongruent). There was also a condition where the quality of the visual input was reduced, in order to prevent a lip-reading of good quality. The aim of each of the five studies was not only to examine whether performances were modified according to the variable under study but also to ascertain that differences were indeed issued from the integration process itself. Thereby, our results were analyzed in the framework of model predictive of audio-visual speech performance (weighted fuzzy-logical model of perception) in order to disentangle unisensory effects from audio-visual integration effects.

Taken together, our data suggest that speech integration is not automatic but rather depends on the context. We propose a new architecture of bimodal fusions, taking these considerations into account. Finally, there are also practical implications suggesting the need to incorporate not only auditory but also visual exercise in the rehabilitation programs of older adults and children with cochlear implants or with specific language impairements.


Doctorat en Sciences Psychologiques et de l'éducation
info:eu-repo/semantics/nonPublished

APA, Harvard, Vancouver, ISO, and other styles
2

Colin, Cécile. "Etude comportementale et électrophysiologique des processus impliqués dans l'effet Mcgurk et dans l'effet de ventriloquie." Doctoral thesis, Universite Libre de Bruxelles, 2001. http://hdl.handle.net/2013/ULB-DIPOT:oai:dipot.ulb.ac.be:2013/211513.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Graham, Robert Edward. "MUSIC TO OUR EYES: ASSESSING THE ROLE OF EXPERIENCE FOR MULTISENSORY INTEGRATION IN MUSIC PERCEPTION." OpenSIUC, 2017. https://opensiuc.lib.siu.edu/dissertations/1491.

Full text
Abstract:
Based on research on the “McGurk Effect” (McGurk & McDonald, 1976) in speech perception, some researchers (e.g. Liberman & Mattingly, 1985) have argued that humans uniquely interpret auditory and visual (motor) speech signals as a single intended audiovisual articulatory gesture, and that such multisensory integration is innate and specific to language. Our goal for the present study was to determine if a McGurk-like Effect holds true for music perception as well, as a domain for which innateness and experience can be disentangled more easily than in language. We sought to investigate the effects of visual musical information on auditory music perception and judgment, the impact of music experience on such audiovisual integration, and the possible role of eye gaze patterns as a potential mediator for music experience and the extent of visual influence on auditory judgments. 108 participants (ages 18-40) completed a questionnaire and melody/rhythm perception tasks to determine music experience and abilities, and then completed speech and musical McGurk tasks. Stimuli were recorded from five sounds produced by a speaker or musician (cellist and trombonist) that ranged incrementally along a continuum from one type to another (e.g. non-vibrato to strong vibrato). In the audiovisual condition, these sounds were paired with videos of the speaker/performer producing one type of sound or another (representing either end of the continuum) such that the audio and video matched or mismatched to varying degrees. Participants indicated, on a 100-point scale, the extent to which the auditory presentation represents one end of the continuum or the other. Auditory judgments for each sound were then compared based on their visual pairings to determine the impact of visual cues on auditory judgments. Additionally, several types of music experience were evaluated as potential predictors of the degree of influence visual stimuli had on auditory judgments. Finally, eye gaze patterns were measured in a different sample of 15 participants to assess relationships between music experience and eye gaze patterns, and eye gaze patterns and extent of visual on auditory judgments. Results indicated a reliable “musical McGurk Effect” in the context of cello vibrato sounds, but weaker overall effects for trombone vibrato sounds and cello pluck and bow sounds. Limited evidence was found to suggest that music experience impacts the extent to which individuals are influenced by visual stimuli when making auditory judgments. The support that was obtained, however, indicated the possibility for diminished visual influence on auditory judgments based on variables associated with music “production” experience. Potential relationships between music experience and eye-gaze patterns were identified. Implications for audiovisual integration in the context of speech and music perception are discussed, and future directions advised.
APA, Harvard, Vancouver, ISO, and other styles
4

Nordstrom, Lauren Donelle. "Brain Mapping of the Latency Epochs in a McGurk Effect Paradigm in Music Performance and Visual Arts Majors." BYU ScholarsArchive, 2015. https://scholarsarchive.byu.edu/etd/4447.

Full text
Abstract:
The McGurk effect is an illusion that occurs when an auditory /ba/ is combined with a visual /ga/. The two stimuli fuse together which leads to the perception of /da/, a sound in between /ba/ and /ga/. The purpose of this study was to determine whether music performance and visual arts majors process mismatched auditory and visual stimuli, like the McGurk effect, differently. Nine syllable pairs were presented to 10 native English speakers (5 music performance majors and 5 visual arts majors between the ages of 18 and 28 years) in a four-forced-choice response paradigm. Data from event-related potentials were recorded for each participant. Results demonstrate that there are differences in the electrophysiological responses to viewing the mismatched syllable pairs. The /ga/ phoneme in the music performance group produced more differences while the /da/ phoneme produced more differences in the visual arts group. The McGurk effect is processed differently in the music performance majors and the visual arts majors; processing begins in the earliest latency epoch in the visual arts group but in the late latency epoch in the music performance group. These results imply that the music performance group has a more complex decoding system than the visual arts group. It also may suggest that the visual arts group is better able to integrate the visual and auditory information to resolve the conflict when mismatched signals are presented.
APA, Harvard, Vancouver, ISO, and other styles
5

Stevanovic, Bettina. "The effect of learning on pitch and speech perception influencing perception of Shepard tones and McGurk syllables using classical and operant conditioning principles /." View thesis, 2007. http://handle.uws.edu.au:8081/1959.7/33694.

Full text
Abstract:
Thesis (Ph.D.)--University of Western Sydney, 2007.
A thesis submitted to the University of Western Sydney, College of Arts, School of Psychology in fulfilment of the requirements for the degree of Doctor of Philosophy. Includes bibliography.
APA, Harvard, Vancouver, ISO, and other styles
6

Attigodu, Chandrashekara Ganesh. "Characterization of audiovisual binding and fusion in the framework of audiovisual speech scene analysis." Thesis, Université Grenoble Alpes (ComUE), 2016. http://www.theses.fr/2016GREAS006/document.

Full text
Abstract:
Cette thèse porte sur l’intégration de deux concepts : l’Analyse de Scènes Auditives (ASA) et la fusion audiovisuelle (AV) en perception de parole. Nous introduisons "l’Analyse de Scènes de Parole Audio Visuelles" (AVSSA) comme une extension du modèle à deux étages caractéristique de l’ASA vers des scènes audiovisuelles et nous proposons qu'un indice de cohérence entre modalités auditive et visuelle est calculé avant la fusion AV, ce qui permet de déterminer si les entrées sensorielles doivent être cognitivement liées : c’est le « modèle à deux étages » de la fusion AV. Des expériences antérieures sur la modulation de l'effet McGurk par des contextes AV cohérents vs. incohérents présentés avant la cible McGurk ont permis de valider le modèle à deux étages. Dans ce travail de thèse, nous étudions le processus AVSSA au sein de l'architecture à deux étages dans différentes dimensions telles que l'introduction de bruit, le mélange de sources AV, la recherche de corrélats neurophysiologiques et l’évaluation sur différentes populations.Une première série d'expériences chez les jeunes adultes a permis la caractérisation du mécanisme de liage AV en introduisant du bruit et les résultats ont montré que les participants étaient en mesure d'évaluer à la fois le niveau de bruit acoustique et la cohérence AV et de contrôler la fusion AV en conséquence. Dans une deuxième série d'expériences comportementales impliquant une compétition entre sources AV, nous avons montré que l’AVSSA permet d'évaluer la cohérence entre caractéristiques visuelles et auditives dans une scène complexe, afin d'associer les composants adéquats d'une source de parole AV donné, et de fournir pour le processus de fusion une évaluation de la cohérence de la source AV extraite. Il apparaît également que la fusion dépend du focus attentionnel sur une source ou l'autre. Puis une expérience EEG a cherché à mettre en évidence un marqueur neurophysiologique du processus de liage-déliage et a montré qu’un contexte AV incohérent peut moduler l'effet de l'entrée visuelle sur la composante N1 / P2. Une dernière série d'expériences a été axée sur l’évaluation du liage AV et de sa dynamique dans une population âgée, et a fourni des résultats similaires à ceux des adultes plus jeunes mais avec une plus grande dynamique de déliage. L'ensemble des résultats a permis de mieux caractériser le processus AVSSA et a été intégré dans la proposition d'une architecture neurocognitive améliorée pour la fusion AV dans la perception de la parole
The present doctoral work is focused on a tentative fusion between two separate concepts: Auditory Scene Analysis (ASA) and Audiovisual (AV) fusion in speech perception. We introduce “Audio Visual Speech Scene Analysis” (AVSSA) as an extension of the two-stage ASA model to- wards AV scenes, and we propose that a coherence index between the auditory and the visual input is computed prior to AV fusion, enabling to determine whether the sensory inputs should be bound together. This is the “two-stage model of AV fusion”. Previous experiments on the modulation of the McGurk effect by AV coherent vs. incoherent contexts presented before the McGurk target have provided experimental evidence supporting the two-stage model. In this doctoral work, we further evaluate the AVSSA process within the two-stage architecture in various dimensions such as introducing noise, considering multiple sources, assessing neurophysiological correlates and testing in different populations.A first set of experiments in younger adults was focused on behavioral characterization of the AV binding process by introducing noise and results showed that the participants were able to evaluate both the level of acoustic noise and AV coherence and to monitor the AV fusion accordingly. In a second set of behavioral experiments involving competing AV sources, we showed that the AVSSA process enables to evaluate the coherence between auditory and visual features within a complex scene, in order to properly associate the adequate components of a given AV speech source, and provide to the fusion process an assessment of the AV coherence of the extracted source. It also appears that the modulation of fusion depends on the attentional focus on one source or the other.Then an EEG experiment aimed to display a neurophysiological marker of the binding and un- binding process and showed that an incoherent AV context could modulate the effect of the visual input on the N1/P2 component. The last set of experiments were focused on measurement of AV binding and its dynamics in the older population, and provided similar results as in younger adults though with a higher amount of unbinding. The whole set of results enabled better characterize the AVSSA process and were embedded in the proposal of an improved neurocognitive architecture for AV fusion in speech perception
APA, Harvard, Vancouver, ISO, and other styles
7

Nahorna, Olha. "Analyse de scènes de parole multisensorielle : mise en évidence et caractérisation d'un processus de liage audiovisuel préalable à la fusion." Thesis, Grenoble, 2013. http://www.theses.fr/2013GRENS039/document.

Full text
Abstract:
Dans la parole audiovisuelle, les flux auditifs et visuels cohérents sont généralement fusionnés en un percept unifié. Il en résulte une meilleure intelligibilité dans le bruit, et cela peut induire une modification visuelle du percept auditif dans le célèbre « effet McGurk » (le montage d'un son « ba » avec une image d'un locuteur prononçant « ga » est souvent perçu comme « da »). La vision classique considère que le traitement est effectué indépendamment dans les systèmes auditif et visuel avant que l'interaction ne se produise à un certain niveau de représentation, ce qui résulte en un percept intégré. Cependant certaines données comportementales et neurophysiologiques suggèrent l'existence d'un processus à deux niveaux. Le premier niveau implique le liage des éléments d'information auditive et visuelle appropriés avant de donner naissance à un percept fusionné au second niveau. Pour démontrer l'existence de ce premier niveau, nous avons élaboré un paradigme original qui vise à tenter de « délier » ces deux flux. Notre paradigme consiste à faire précéder l'effet McGurk (indicateur de la fusion audiovisuelle) par un contexte soit cohérent soit incohérent. Dans le cas du contexte incohérent on observe une diminution significative de perception d'effet McGurk, donc une décroissance de la fusion audiovisuelle. Les différent types d'incohérence (syllabes acoustiques superposées à des phrases visuelles, modifications phonétiques ou temporelles du contenu acoustique de séquences régulières de syllabes audiovisuelles) peuvent réduire significativement l'effet McGurk. Le processus de déliage est rapide, une unique syllabe incohérente suffisant pour obtenir un résultat de déliage maximal. Par contre le processus inverse de « reliage » par un contexte cohérent suivant le déliage est progressif, puisqu'il apparaît qu'au minimum trois syllabes cohérentes sont nécessaires. Nous pouvons également geler le sujet dans son état délié en rajoutant une pause entre un contexte incohérent et l'effet McGurk. Au total 7 expériences ont été effectuées pour démontrer et décrire le processus de liage dans la parole audiovisuelle. Les données sont interprétées dans le cadre du modèle à deux niveaux « liage et fusion »
In audiovisual speech the coherent auditory and visual streams are generally fused into a single percept. This results in enhanced intelligibility in noise, or in visual modification of the auditory percept in the famous “McGurk effect” (the dubbing of the sound “ba” on the image of the speaker uttering “ga” is often perceived as “da”). It is classically considered that processing is done independently in the auditory and visual systems before interaction occurs at a certain representational stage, resulting in an integrated percept. However, some behavioral and neurophysiological data suggest the existence of a two-stage process. A first stage would involve binding together the appropriate pieces of audio and video information, before fusion in a second stage. To demonstrate the existence of this first stage, we have designed an original paradigm aiming at possibly “unbinding” the audio and visual streams. Our paradigm consists in providing before a McGurk stimulus (used as an indicator of audiovisual fusion) an audiovisual context either coherent or incoherent. In the case of an incoherent context we observe a significant decrease of the McGurk effect, implying a reduction of the amount of audiovisual fusion. Various kinds of incoherence (acoustic syllables dubbed on video sentences, phonetic or temporal modifications of the acoustic content of a regular sequence of audiovisual syllables) can significantly reduce the McGurk effect. The unbinding process is fast since one incoherent syllable is enough to produce maximal unbinding. On the other side, the inverse process of “rebinding” by a coherent context following unbinding is progressive, since it appears that at least three coherent syllables are needed to completely recover from unbinding. The subject can also be “freezed” in an “unbound” state by adding a pause between an incoherent context and the McGurk target. In total seven experiments were performed to demonstrate and describe the binding process in audiovisual speech perception. The data are interpreted in the framework of a two-stage “binding and fusion” model
APA, Harvard, Vancouver, ISO, and other styles
8

Klitsch, Julia Ulrike. "Open your eyes and listen carefully auditory and audiovisual speech perception and the McGurk effect in Dutch speakers with and without aphasia /." [S.l. : [Groningen : s.n.] ; University Library Groningen] [Host], 2008. http://irs.ub.rug.nl/ppn/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Bedard-Giraud, Kimberly. "Troubles du traitement de la parole chez le dyslexique adulte." Toulouse 3, 2007. http://www.theses.fr/2007TOU30334.

Full text
Abstract:
Les troubles du traitement de la parole peuvent jouer un rôle causal dans certains cas de dyslexie. Cette recherche porte sur le traitement auditif de consonnes occlusives chez le dyslexique adulte. Dans la première étude [décours temporel des Potentiels Evoqués Auditifs (PEAs)], nous analysons le traitement cortical des indices temporels constituant le "Voice Onset Time" de consonnes voisées et non-voisées. On constate deux profils atypiques: (i) "PEA Pattern I": traitement différentiel sur la base d'indices temporels, mais avec plus de composantes et des délais; (ii) "PEA Pattern II": absence de traitement différentiel sur la base d'indices temporels. Dans la deuxième étude [modélisation des sources et analyse d'asymétries], on constate que ce traitement temporel peut être latéralisé atypiquement chez le dyslexique - même dans des cas compensés avec décours temporels relativement normaux. La troisième étude [Perception Catégorielle & MMN] suggère une correspondance entre nos profils électrophysiologiques et la discrimination de sons occlusives: le "PEA Pattern I" peut être associé à un traitement superflu d'indices acoustiques sans pertinence phonétique, tandis que le "PEA Pattern II" semble lié à un trouble de discrimination sévère. Dans la quatrième étude [l'effet McGurk], nous analysons l'intégration des indices acoustiques et visuels dans la perception de la parole face-à-face. Chez le dyslexique, on constate moins d'intégrations audiovisuelles et plus de dépendance sur des indices strictement acoustiques. Dans l'ensemble, ces résultats sont cohérents avec un dysfonctionnement affectant plusieurs niveaux du traitement de la parole chez le dyslexique
Speech perception deficits may play a causal role in certain cases of developmental dyslexia. This research focuses on the perception of stop consonants in the adult dyslexic. In the first study [temporal course of Auditory Evoked Potentials (AEPs)], the cortical processing of temporal cues (Voice Onset Time) differentiating voiced and voiceless stops is analysed in dyslexics with persistent deficits. Two atypical electrophysiological patterns are observed: (i) AEP Pattern I is characterised by a differential coding of stimuli on the basis of some temporal cues but with more AEP components and a delay in termination time; (ii) AEP Pattern II is characterised by an absence of differential coding based on temporal cues. The second study [source modelling and asymmetry of temporal processing] shows an atypical functional asymmetry of this temporal cue processing in adult dyslexics - even in compensated cases with relatively normal AEP timecourses. The third study [Categorical Perception and MMN] suggests how atypical temporal cue processing may affect stop consonant discrimination: AEP Pattern I may be associated with the coding of superfluous non-phonetically pertinent cues, while AEP Pattern II may be associated with a severe voiced/voiceless discrimination deficit. In the fourth study [McGurk Effect], the integration of acoustic and visual cues in face-to-face speech perception is analysed in adult dyslexics. Compared to controls, dyslexics demonstrated less audiovisual integration, relying preferentially on acoustic cues. Together, these results are consistent with a speech perception deficit that affects multiple levels of processing in the developmental dyslexic
APA, Harvard, Vancouver, ISO, and other styles
10

Deonarine, Justin. "Noise reduction limits the McGurk Effect." Thesis, 2011. http://hdl.handle.net/10012/6046.

Full text
Abstract:
In the McGurk Effect (McGurk & MacDonald, 1976), a visual depiction of a speaker silently mouthing the syllable [ga]/[ka] is presented concurrently with the auditory input [ba]/[pa], resulting in “fused” [da]/[ta] being heard. Deonarine (2010) found that increasing the intensity (volume) of the auditory input changes the perception of the auditory input from [ga] (at quiet volume levels) to [da], and then to [ba] (at loud volume levels). The present experiments show that reducing both ambient noise (additional frequencies in the environment) and stimulus noise (excess frequencies in the sound wave which accompany the intended auditory signal) prevents the illusory percept. This suggests that noise is crucial to audiovisual integration and that the McGurk effect depends on the existence of auditory ambiguity.
APA, Harvard, Vancouver, ISO, and other styles

Books on the topic "Perception/McGyrk effect"

1

Dias, James W., Theresa C. Cook, and Lawrence D. Rosenblum. The McGurk Effect and the Primacy of Multisensory Perception. Oxford University Press, 2017. http://dx.doi.org/10.1093/acprof:oso/9780199794607.003.0115.

Full text
Abstract:
The McGurk effect is an audiovisual speech illusion in which tacit lip-reading of a spoken syllable influences perception of a synchronized, but discrepant, auditory syllable. The effect works on speakers of different native languages and on prelinguistic infants. The effect is also robust to various image manipulations and to observers’ conscious awareness of the stimulus composition. For these reasons, the effect has become one of the preeminent demonstrations of multisensory perception. This chapter discusses research on the McGurk effect, as well as its influence on theories of speech perception and its importance to the reconceptualization of the perceptual brain as designed around multisensory input.
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "Perception/McGyrk effect"

1

Keil, Julian, Niklas Ihssen, and Nathan Weisz. "Prestimulus Oscillatory Brain Activity Influences the Perception of the McGurk Effect." In IFMBE Proceedings, 219–22. Berlin, Heidelberg: Springer Berlin Heidelberg, 2010. http://dx.doi.org/10.1007/978-3-642-12197-5_49.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Burnham, Denis, and Barbara Dodd. "Auditory-Visual Speech Perception as a Direct Process: The McGurk Effect in Infants and Across Languages." In Speechreading by Humans and Machines, 103–14. Berlin, Heidelberg: Springer Berlin Heidelberg, 1996. http://dx.doi.org/10.1007/978-3-662-13015-5_7.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Cavedon-Taylor, Dan. "High-Level Perception and Multimodal Perception." In Purpose and Procedure in Philosophy of Perception, 147–73. Oxford University Press, 2021. http://dx.doi.org/10.1093/oso/9780198853534.003.0008.

Full text
Abstract:
What is the correct procedure for determining the contents of perception? Philosophers tackling this question increasingly rely on empirically oriented procedures. This chapter argues that this strategy constitutes an improvement over the armchair methodology of phenomenal contrast arguments, but that there is a respect in which current empirical procedures remain limited: they are unimodal in nature, wrongly treating the senses as isolatable. The chapter thus has two aims: first, to motivate a reorientation of the admissible contents debate into a multimodal framework. The second is to explore whether experimental studies of multimodal perception support a so-called Liberal account of perception’s admissible contents. The chapter concludes that the McGurk effect and the ventriloquist effect are both explicable without the postulation of high-level content, but that at least one multimodal experimental paradigm may necessitate such content: the rubber hand illusion. One upshot is that Conservatives who claim that the Liberal view intolerably broadens the scope of perceptual illusions, particularly from the perspective of perceptual psychology, should pursue other arguments against that view.
APA, Harvard, Vancouver, ISO, and other styles
4

O'Callaghan, Casey. "Processes." In A Multisensory Philosophy of Perception, 19–52. Oxford University Press, 2019. http://dx.doi.org/10.1093/oso/9780198833703.003.0002.

Full text
Abstract:
Crossmodal perceptual illusions such as ventriloquism, the McGurk effect, the rubber hand, and the sound-induced flash demonstrate that one sense can causally impact perceptual processing and experience that is associated with another sense. This chapter argues that such causal interactions between senses are not merely accidental. Interactions between senses are part of typical perceptual functioning. Unlike synesthesia, they reveal principled perceptual strategies for dealing with noisy, fallible sensory stimulation from multiple sources. Recalibrations resolve conflicts between senses and weight in deference to the more reliable modality. Coordination between senses thus improves the coherence and the reliability of human perceptual capacities. Therefore, some perceptual processes of the sort relevant to empirical psychology are multisensory.
APA, Harvard, Vancouver, ISO, and other styles
5

Nicholls, Michael E. R., Dara A. Searle, and John L. Bradshaw. "Read My Lips: Asymmetries in the Visual Expression and Perception of Speech Revealed through the McGurk Effect*." In Language in Use, 350–58. Routledge, 2020. http://dx.doi.org/10.4324/9781003060994-33.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Perception/McGyrk effect"

1

AlAnsari, Noora Essa, Ali Idrissi, and Michael Grosvald. "The McGurk Effect in Qatari Arabic: Influences of Lexicality and Consonant Position." In Qatar University Annual Research Forum & Exhibition. Qatar University Press, 2020. http://dx.doi.org/10.29117/quarfe.2020.0279.

Full text
Abstract:
The McGurk effect is a psycholinguistic phenomenon where an illusion is made by dubbing an auditory element of one sound on a visual element of another sound, which leads to hearing a third sound. The phenomenon demonstrates how the perception of speech does not depend on audio inputs only. Rather, it shows how seeing the shape of the mouth while producing a certain sound can influence what we hear. Thus, it proves the interaction of both vision and auditory parameters in understanding language. In addition, what is known as “lexicality – the property of a word being real or not” influences speech perception. People, unconsciously, tend to alter nonwords to real words. For example, if one said “shtrength” instead of “strength”, a listener would alter and understand it as “strength”. For the purpose of the research, these two phenomena were combined. In this study, we test how effective is the McGurk effect on the Qatari Arabic dialect, which has not been investigated before. The data used were 24 minimal pairs of real and fake words with the substitution of the phonemes: /b/ and /g/ at three different positions: first, middle, final. Videos were made by dubbing audio recordings of the sound /b/ into video recordings of the sound /g/ in order to test if this creates an illusion of the sound /d/. We ran the experiment on 25 native Qatari female students, they had to sit on a computer with headphones on, watch and hear clearly what the person on the video is saying, and then they had to preform two tasks: first, lexical decision task: decide if the word is real or fake. Second, sound discrimination task: choose what sound did they hear. In general, the participants captured audio (which means they heard /b/) were only 16% of target trials, while visual capture occurred (which means they heard /g/) 45%, and the McGurk fusion (which means they heard /d/) happened on 39%. Interestingly, perceiving McGurk fusion was gradually less common at later consonants positions. A significant effect of lexicality was also found, as fusion was more likely to occur if the results of the fusion was a real word.
APA, Harvard, Vancouver, ISO, and other styles
2

Massaro, Dominic. "The McGurk Effect: Auditory Visual Speech Perception’s Piltdown Man." In The 14th International Conference on Auditory-Visual Speech Processing. ISCA: ISCA, 2017. http://dx.doi.org/10.21437/avsp.2017-25.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Rahmawati, Sabrina, and Michitaka Ohgishi. "Cross cultural studies on audiovisual speech processing: The Mcgurk effects observed in consonant and vowel perception." In 2011 6th International Conference on Telecommunication Systems, Services, and Applications (TSSA). IEEE, 2011. http://dx.doi.org/10.1109/tssa.2011.6095406.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography