User:Fives enough/sandbox

McGurk effect

I would like to expand this article to include more information on its effect in whole sentences and add in some missing citations. I've also been looking at information and studies regarding how this phenomenon effects listeners in various languages besides English, particularly Italian; German; Spanish; Japanese; and Chinese, and would like to add information about this. Along with these two areas I would like to add information on how this effect compares in spoken and sung phonemes. My hope is to also weave some info into the article on how/why the brain transfers the visual phoneme and auditory phoneme into the third phoneme. I will be using the sources listed below.

Annotated Bibliography Section

1. The authors of this article are both researchers and have written numerous articles relating to speech, hearing, and computer or technology use; while each has a different background: studying how people integrate sensory information in speech perception and building lipreading machines that can decode speech. I did not detect any particular bias in their article. The article was mainly discussing how people and machines can integrate auditory and visual information to understand speech. While the article was posted in the American Scientist it does not contain much technical jargon and could be easily understand by a college level student or someone with an interest in the subject.

2. The authors of this article are Italian researchers in the Audiology and Phoniatrics Department at University Hospital of Ferrara. I cannot locate further credentials that are in English, but came across multiple publications and articles written by them individually and collectively. This article discusses the phenomenon as it occurs in various other languages with the main focus of their study on Italian subjects. The article starts out fairly basic but it soon becomes apparent that it is a journal publication and becomes much more technical and over the heads of the average reader. I did not notice any bias to the article.

3. This article is written by various authors with a wide range of experience. I think this group is a reliable source due to their combined vast knowledge of the music and sound field as well it's relation to auditory research. Each member has authored numerous articles and publications. The article appears to give a fair assessment of the study and does not seem biased to me. This article is found in a professional journal and does include quite a bit of technical terminology, but with a little further research can be easily understood by college students or anyone familiar with speech therapy, or has an auditory or music background. This article discusses how phonemes are heard and visualized in sung syllables versus the studies of this same phenomenon in speech patterns.

**Current work on McGurk effect article: Previous article in regular type; new material in bold (tables are new material)**
The McGurk effect is a perceptual phenomenon which demonstrates an interaction between hearing and vision in speech perception. It is a compelling illusion in which humans perceive mismatched audiovisual speech as a completely different syllable. It suggests that speech perception is multimodal, that is, that it involves information from more than one sensory modality. This effect may be experienced when a video of one phoneme's production is dubbed with a sound-recording of a different phoneme being spoken. Often, the perceived phoneme is a third, intermediate phoneme. As an example, the syllable /ba-ba/ is spoken over the lip movements of /ga-ga/, an the perception is of /da-da/. The McGurk effect is sometimes called the McGurk-MacDonald effect.

It was first described in a paper by Harry McGurk and John MacDonald in 1976. This effect was discovered by accident when McGurk and his research assistant, MacDonald, asked a technician to dub a video with a different phoneme other than the one spoken while conducting a study on how infants perceive language at different development stages. When the video was played back, both researchers heard a third phoneme rather than the one spoken or mouthed in the video. McGurk and MacDonald originally believe that this effect resulted from the common phonetic and visual properties of /b/ and /g/. '''The researchers explain that the sound 'bah' is more acoustically in line with 'dah' and that 'gah' is more visually in line with 'dah' rather than 'bah'. These similarities create mixed signals in the brain - in the auditory and visual processing centers, and the brain needing to find a common factor between these two centers processes 'duh.' Research suggests that visual information plays an important role in how well we process auditory information even when sufficient auditory information in presented in a clear and accurate format. Listeners unconsciously watch mouth and facial movements as a form of lip reading to understand a speaker's meaning. When the facial area or facial movements are obscured, whether through the McGurk effect or lack of visual contact with the face miscommunication is more likely.'''

Further studies have shown that the McGurk effect appears with other consonants and vowels and that it can exist throughout whole sentences, but the illusion is greater with certain vowels, consonants, and words. '''Studies have shown that the McGurk effect is stronger when matching vowel combinations are used in the auditory and visual syllable sound stimuli versus using non matching vowel combinations. The same was true for matching and nonmatching vowel combinations in the spoken word stimuli.'''

When subjects watch a video with the visual "My gag kok me koo grive" dubbed with the audio "My bap pop me poo brive" most subjects reported hearing "My dad taught me to drive" on the audiovisual. The effect is very robust; that is, knowledge about it seems to have little effect on one's perception of it. '''Subjects can be told of the effect before watching a dubbed video and they will still hear the third phoneme while watching, yet if they close their eyes they can hear the correct auditory stimuli. Subjects report hearing a third phoneme even when watching a dubbed video of themselves mouthing the sound. This illusion''' is different from certain optical illusions, which break down once one "sees through" them. '''Precise synchronization of mouth movements and dubbed words, clarity of image, or the gaze patterns of the subjects do not play a role in whether or not subjects successfully hear the third phoneme. It has been have shown that a delay of 250 msec or advance of 60 msec between visual and auditory stimuli can appear before results begin to be effected.'''

'''Research has shown that the McGurk effect is prevalent in other languages as well as English. It has been demonstrated that it has a strong effect in Italian, German, Spanish, and Hungarian languages, while it has a weaker effect in Japanese, Chinese, and Thai languages. These latter languages have simpler phonological cues with less consonant contrasts along with fewer visually distinct contrasts. It is possible that languages with more complex phonological characteristics require more attention to visual cues. Using the McGurk effect to decipher lyrics to a song has a greater outcome then auditory stimuli alone, but it's effect is not as strong as with the spoken word. This may be due to the difference in movement of the jaw and lips. In speech for the most part these movements are minimal, where as in singing the mouth is opened wider and lip movements are exaggerated for the production of higher pitches and sound fullness,and articulating sung vowels.'''

Study into the McGurk effect is being used to produce more accurate speech recognition programs by making use of a video camera and lip reading software. It has also been examined in relation to witness testimony; Wareham & Wright's 2005 study showed that inconsistent visual information can change the perception of spoken utterances, suggesting that the McGurk effect may have many influences in everyday perception.

References