Download Chapter 12 – Auditory Localization and Organization

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Premovement neuronal activity wikipedia , lookup

Biological neuron model wikipedia , lookup

Development of the nervous system wikipedia , lookup

Cortical cooling wikipedia , lookup

Bird vocalization wikipedia , lookup

Neural coding wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Stimulus (physiology) wikipedia , lookup

Nervous system network models wikipedia , lookup

Synaptic gating wikipedia , lookup

Sound wikipedia , lookup

Feature detection (nervous system) wikipedia , lookup

Cognitive neuroscience of music wikipedia , lookup

Animal echolocation wikipedia , lookup

Perception of infrasound wikipedia , lookup

Sensory cue wikipedia , lookup

Sound localization wikipedia , lookup

Transcript
Chapter 12 – Auditory Localization and Organization
Localization of sounds
Localizing Visual stimuli in vertical and horizontal dimensions – Too easy!! Visual stimuli at
different places activate different locations on the retina. Each different place on the retina is
connected to a different area of the cortex. So it’s kind of easy to see how the brain can locate
different visual stimuli – by noting where the activity occurs on the retina, LGN, area V1.
Localizing in the 3rd dimension – depth – is more difficult. It requires “neural deduction” based
on monocular cues and comparison of images in the two eyes – binocular disparity.
The issues involving perceiving sound are like those involved in using binocular disparity to
perceive depth or distance . . .
A sound originating from the left activates ALL the receptors on both the left cochlea and right
cochlea.
Similarly, a sound originating from the right activates ALL the receptors on both the left
cochlea and right cochlea.
So, perceiving location requires a comparison of the input to the two ears.
Elevation –
location
above or
below the
head.
Some terms . . .
Azimuth –
location on
a circle
around the
head
0◦
270◦
Or 90◦ left
45◦
◦
90 right
0◦
Azimuth –
Location in a
horizontal
circle around
the listener.
Elevation –
Location in a
vertical circle
around the listener.
0◦ is immediately in
front of us.
180◦
Auditory Locatization and Organization - 1
3/21/5
Cues for localization in the azimuth
Sound localization in the azimuth involves comparison of the sounds in the two ears.
The primary cues are binaural cues: Specific differences in the sounds between the two ears.
1. Interaural Time Differences (ITDs): G9 p 291
Also called time of arrival differences
A sound originating on the left will reach the left ear slightly sooner than it reaches the right ear.
Vice versa for sounds originating on the right.
Maximum difference in time of arrival is associated with a sound on immediately to the left of
the left ear or the right of the right ear. . . .
600 µsec
600 µsec = .000600 sec. = 600
millionths of a second
The maximum difference associated with a sound from one side is .000600 seconds, 600
millionths of a second, or 600 microseconds.
Time of arrive differences are by themselves enough to allow precise localization of sounds.
Careful experiments in which the only cue for location was time of arrival have demonstrated
that persons can accurately locate sounds based on this cue only.
The threshold for ITD is about 100 µsec – sounds which differ in ITD by less than 100 µsec are
perceived to be directly in front of us 50% of the time.
Interaural time differences (ITDs) are most important for localization of low frequency
sounds.
Auditory Locatization and Organization - 2
3/21/5
2. Interaural Level Differences (ILDs) G9 p 291
Differences in Level (intensity) between the ears
Sounds which originate on the left will tend to be more intense at the left ear.
Sounds which originate on the right will tend to be more intense at the right ear.
Our perception corresponds to reality . . .
All other things equal, a sound localized to the left will be more intense in the left ear.
A sound on the right will be more intense in the right ear.
But: How different the intensities will be in the two ears depends on the frequency of the sound.
A sound originating on the left has two ways of getting to the right ear (assuming no reflective
surfaces) . . .
a. bending around the head.
Low frequency sounds
bend easily around the head
b. going through the head.
High frequency sounds must go through
the head, reducing intensity.
Ear in
“shadow”
The result is the following:
Frequency
Maximum Intensity difference between ears
300 Hz
1 db (barely noticeable) Low frequency sound easily bend around the
head.
4000 Hz
5 – 10 db (noticeable)
10,000 Hz
10-20 db (quite noticeable) High frequency sounds are blocked by the
head.
Auditory Locatization and Organization - 3
3/21/5
Graphically . . .
So interaural level differences provide good information on location of high frequency
sounds.
But interaural intensity differences provide little information on location of low frequency
sounds
This is one of the reasons that some persons with high quality sound systems have only one
speaker designed for very low frequency sounds. It’s because we are less able to identify the
location of such sounds, so there is no need for two in a stereo system.
Auditory Locatization and Organization - 4
3/21/5
Perceiving Elevation G9 p 293
The Cone of Confusion
ITDs and ILDs give great information about localization in the azimuth.
Alas, they don’t provide much information
about elevation.
The collection of locations for which ITD
and ILD are equal is called the cone of
confusion.
The locations on the cone differ in elevation.
A and B have exactly the same ITD and ILD.
This means that ITD and ILD are of little use
in determining elevation.
Spectrum cues used to judge elevation and whether sound is in front of us or in back
The shape of the pinna differentially attenuates different frequency components of the sound
stimulus.
This shaping of the sound stimulus creates information for the location of the sound called the
spectral shape cue – a cue for location based on the shape of the spectrum. This cue is
particularly important for perception of the elevation of the sound stimulus.
The shaping of the sound spectrum by the pinna is specific to each individual ear.
This means that using the shape of the sound spectrum to perceive elevation must be learned.
It also implies that if the shape of an adult’s pinna were changed, that adult would have difficulty
locating the elevation of sounds they could previously locate very easily.
Spectra of same sound from two elevations
Note that the sound from
above is generally more
intense n the ear canal.
Note also that the sound
from below lacks
intensity in the 6-8 KHz
frequency range relative
to the sound from above.
Auditory Locatization and Organization - 5
3/21/5
Influence of shape of the pinna in judging elevation
Figure 12.9 G9 p 295 shows the effects of wearing an artificial pinna on ability to localize
sounds.
The crossings of the blue grid are where sounds were presented.
The red dots are where the sounds were reported to be located.
Note the close correspondence for this control condition.
On the first day that observers wore an implant, their reports
were horribly inaccurate in the vertical dimension, although
pretty nearly correct in the horizontal dimension.
After 5 days of wearing the implant, observers’ reported
locations in the vertical dimension became more accurate.
After 19 days wearing the implant, observers’ reports of both
vertical and horizontal location were pretty accurate.
Immediately after removing the implant, observers’ reports were
still quite accurate, suggesting that they had not “lost” an ability,
but simply gained a new ability.
This study suggests that the timbre (the experienced frequency spectrum) of the sound is a cue
for that sound’s elevation
Auditory Locatization and Organization - 6
3/21/5
Using timbre to determine whether a sound is in front of us or in back.
The pinna attenuates the high frequency components of sounds located in back of the head.
This means that we can tell when a familiar sound is behind us because it sounds slightly
muffled compared to the way it sounds when it’s in front of us.
A study demonstrating the effect of the ear flap . . .
Batteau (1967) as reported in L&S, p. 424.
Batteau created artificial pinnae out of plaster and put them on an artificial head. They then put
microphones in the hole in the artificial pinna where the ear drum would be.
Condition 1: Recorded sounds played in different locations around the head with the plaster
pinnas. Recall that there was a microphone in each plaster ear canal.
Condition 2: Recorded the same sounds played in different locations around a head in which
microphones had been embedded without plaster pinnae. That is, the artificial head had no
ears, just ear canals with microphones at the end of the canals.
Played back the recordings of those sounds to human observers.
Results:
Condition 1: Participants were able to identify locations of the sounds in front of and behind
the artificial head.
Condition 2: Participants were not able to localize the sounds correctly.
Conclusion: The plaster, artificial pinnae muffled the sounds played behind the head in a way
that was similar to how sounds are muffled by the real pinna.
Effect of differences between the two pinnae on location in azimuth
We’d be much better at locating sounds in the azimuth if our ears had different shapes.
Perceiving distance (Not in G9, from K&S p 296)
Familiarity – comparison of perceived sound with memories of similar sounds at different
distances
Spectral analysis – High frequency sounds decrease in intensity over distance more than low
frequency sounds. So we hear the bass guitar across the lake, but not the singer. We hear the
rumble of thunder but not the high frequency components of a crack of lightning.
Ratio of direct to reflected sound – Nearer sounds have a different ratio of direct to reflected
than far sounds.
Auditory Locatization and Organization - 7
3/21/5
Possible neural circuits for ITD and ILD – G9 p 295
Study of barn owls has provided evidence for existence of neurons that respond only to
sounds presented in a particular location.
The superior olivary complex contains the medial superior olive, a collection of neurons that
have been identified as responding to ITD.
Neurons have been found which respond only when one ear receives sound slightly before the
other – to time of arrival differences – that is, ITDs.
From the web site of David Heeger, NYU.
Auditory Locatization and Organization - 8
3/21/5
A circuit for neurons responding to Interaural Time Differences. – G9 p 297
Issue: Design a neural circuit so that a neuron responds most actively when a sound arrives at
the right ear before it arrives at the left ear.
Sound
on the
right
Stimulus
Back of
head
Ears
L
Short axon from
left cochlear
nucleus
R
MSO
MSO
Neuron
Long axon from
right cochlear
nucleus
A simple way to create time of arrival signaling is to have a longer path from neurons
originating in one ear to the postsynaptic neuron than the path from the other ear. This is
analagous to the motion detectors we studied a couple of chapters ago.
Sounds in front of us
If a sound strikes both L and R at the same time, L’s action potential will arrive at the end of its
axon, followed by R’s. That’s because the action potential has further to travel in R than it does
in L. Thus, the total release of neurotransmitter substance at MSO will be long-term, but low
intensity, resulting in low probability of MSO neuron responding.
Sounds to the right of us
Sound strikes the right ear first, but the path from the ear to the MSO from the right ear is longer.
So R’s action potential will reach the end of its axon at about the same time as L’s, causing a
massive release of neurotransmitter, enough to cause the MSO neuron to fire.
Sounds to the left of us
But if a sound strikes L first, then strikes R shortly afterward, L’s action potential will reach the
end of its axon first and much much later (in neural time) R’s action potential will reach the end
of its axon, resulting in insufficient neurotransmitter to cause the MSO neuron to respond.
Note that in such circuits, the long path is from the ear closest to the sound. (In case you need
to construct such a neural circuit.)
Play ..\..\..\MDBT\P312\Ch14AuditoryPhenomena\ITD Neuron.pptx (It’s in Class Videos)
Auditory Locatization and Organization - 9
3/21/5
The Jeffress Neural Coincidence Model – combining many of the above circuits.
Figure 12.12, G9 p 297
Each neuron – 1, 2, 3, etc except 5
– has a different length path to
each ear.
Neuron 5 has an equal length path
from each ear, so it will respond
best to sounds from the front.
.
Neuron 3 has a long path from the
right ear and a short path from the
left ear, so 3 will respond best to
sound that reach the right ear
first.
Imagine 1000s of neurons like 1,2,3,4,5,6,7,8,9 above,
each with a different length path to each ear. Each of
those neurons will respond best to a sound at a particular
location on the azimuth.
Each curve is the firing rate of a neuron to different
ITDs.
Auditory Locatization and Organization - 10
3/21/5
Mammalian Neurons
It appears that tuning curves of neurons in mammals are not as sharp as those of owls.
The gerbil curve is about twice as wide as the owl curve in Figure 12.14 p 298.
It’s been suggested that in mammals, there are fewer more broadly tuned neurons, ones that fire
Forgettabout what’s below this box.
over a wide range of locations to the right and others that fire over a wide range of locations to
the left.
The blue curve is of a neuron that responds above its
base rate to sounds from the right and below its base rate
to sounds from the left.
The left curve has the opposite response characteristics.
Note – each of these curves is like the response curve of
an opponent processes neuron in the LGN.
In this model, the location of a sound is signaled by the
pattern of responses of these two types of neurons . . .
Sound 1 is on the left- resulting in a large response of the
L neuron, but a small response of the R neuron.
Sound 3 is on the right – resulting in a small response of
the L neuron but a large response of the R neuron.
Auditory Locatization and Organization - 11
3/21/5
A Neural Circuit responding to Interaural Level Differences
(This one is almost too easy.)
Below is a simple-minded circuit to process intensity differences.
L
R
ILD
Sounds in front of us
Neurotransmitter from L and R counteract each other, resulting in base rate activity of LSO. So
a base rate of activity would signal that a sound is in front of us.
Sounds to the right
Since the sound is more intense at R, R releases more excitatory neurotransmitter substance the
L, making ILD more active. So through its heightened activity, ILD signals “sounds to the
right”.
Sounds to the left
Since the sound is more intense at L, L releases more inhibitory neurotransmitter substance than
R, reducing LSO activity. So through its lowered rate of activity ILD signals “sounds to the
left”.
Obviously, the neuron ILD by itself could not completely identify the location of a sound, but its
activity, in conjunction with that of other neurons responding to other aspects of the auditory
scene would help signal sound location in the azimuth.
.
Auditory Locatization and Organization - 12
3/21/5
The auditory pathway to the auditory cortex –
Note that each primary
auditory projection area (A1)
is buried in a sulcus.
An important point to get from this is that there is a LOT of processing of the auditory signal in
each of these structures.
A natural question is: What exactly are these structures doing?
As Yantis, p 325 says, “The exact functions served by the different subcortical structures and the
different response patterns of their neurons are the subject of considerable ongoing research
and debate, but it’s clear that these structures and the neural signal processing that takes place
within them along the auditory pathways play a critical role in encoding, with exquisitely high
fidelity, the rapidly changing stimuli that typically make up the auditory environment.”
Auditory Locatization and Organization - 13
3/21/5
Tonotopic maps are found in almost every cortical area.
An image of the initial processing of sound that emphasizes the tonotopic maps in the various
areas . . .
High Freq
Low Freq
Reference is journal.frontiersin.org
Red represents low frequency sounds
Blue represents high frequency sounds
AN: Auditory Nerve
CN: Cochlear Nucleus
LSO: Lateral Superior Olive
MSO: Medial Superior Olive
MNTB: Medial Nucleus of the Trapezoidal Body – not mentioned in Y1
The point is that every area through which the auditory signal passes is organized tonotopically –
by frequency.
Auditory Locatization and Organization - 14
3/21/5
The auditory Cortex – Y1 p 327
The auditory cortex is located on the upper surface of the temporal lobe tucked into the lateral
sulcus – the sulcus separating the temporal lobe from the frontal and parietal lobes.
A primary discovery associated with the auditory cortex is the tonotopic map – the neurons are
organized by the frequency of sound to which they’re most sensitive.
Note that the maps reverse as you move through the auditory cortex – high to low; low to high;
high to low
The region is separated into three subregions – the core, the belt, and the parabelt.
Neurons in the core respond best to pure tones.
Neurons in the belt and parabelt respond to more complex sound stimuli.
This is analogous to the relationships of visual neuron responses in V1 vs. V4, MT, and other
cortical areas. The complexity of the stimuli to which neurons respond becomes greater, the
farther you go from the initial projection area.
Auditory Locatization and Organization - 15
3/21/5
What and where pathways
Auditory cortex neurons that carry information about sound localization send their axons to the
parietal lobe.
Auditory cortex neurons that carry information that might help in identifying what is in the
external environment send their axons to the frontal lobe.
The auditory pathways meet the visual “what” and “where” pathways.
The “where” pathways meet in the parietal lobe (in the monkey).
The “what” pathways meet in the prefrontal cortex (again, in the monkey).
This, of course, means that there are neurons that receive both visual and auditory information –
in the blue areas in the above figure.
Images of brain activity of a human performing
either identification of sounds or
location of sounds.
From Goldstein’s 8th Ed.
Auditory Locatization and Organization - 16
3/21/5
Architectural Acoustics G9 p 302
Reverberation – Delay and strength of echos
Definition: The time it takes for the reflections of the sound to become 60 decibels less than the
original sound.
Play VL 12.14
Best is about 2 seconds for concert halls and 1.5 seconds for opera houses.
Best is 0.4-0.6 seconds for a classroom, about 1.0-1.5 seconds for an auditorium.
Intimacy time: The time between arrival of the sound directly from the source and the arrival of
the first reflection. Like reverberation, but focusing on the time of the first reflection, not the
total amount of reflected sound.
Best is about 20 milliseconds. (.020 sec).
Bass ratio: The ratio of low frequencies to middle frequencies of reflected sounds.
Best is a high ratio.
Spaciousness factor: The fraction of all the sound received by a listener that is reflected, as
opposed to direct.
Best is high.
Signal to Noise Ratio: The decibel difference between the amplitude of the signal (voice in a
classroom) vs. the amplitude of sounds without the signal – the sound of the heating system,
surrounding cars, people in the hallway.
Need about 10-15 S/N ratio – but higher is better.
Auditory Locatization and Organization - 17
3/21/5
Auditory Scene Analysis – Y1 p 341
page
skipped
in Summer 2016
Identifying the objectsThis
in the
auditory
scene.
The different pieces of the auditory experience.
Example – Musical scenes . . .
American Woman by the Guess Who – (Play from 1:11 on . . .)
..\..\..\MDBT\P312\Ch14AuditoryPhenomena\American Woman Whole Version by the Guess
Who.mp4
Taken from YouTube.
Turtle Blues – 2:56 – breaking glass
..\..\..\MDBT\P312\Ch14AuditoryPhenomena\Turtle Blues from Cheap Thrills by Janis
Joplin.mp4
The problem . . .
what we hear has multiple sources, each broadcasting its own waveform,
but
the sound waveform arrives as a single waveform, a mixture of all of the original waveforms.
How do we break up the composite waveform into its pieces, identifying the different sources so
that we hear different things simultaneously?
Auditory Locatization and Organization - 18
3/21/5
Factors affecting how we identify auditory “objects”.
These are analagous to the Gestalt Laws of Grouping
This page skipped in Summer 2016
1. Location – aka spatial segregation.
Sounds from the same location will be more likely to be perceived as coming from the same
source.
2. Onset Time – aka temporal segregation.
Sounds which begin at the same time are grouped together.
3. Pitch and timbre similarities – spectral segregation.
Play VL 12.3
Play VL 12.6
Harmonic coherence – Most sounds are fundamental + harmonics
So having a sound which is not one of the harmonics makes it seem
separate.
Sigview – Load 200,400,500,800 vs 200,400,600,800 workspaces. Play each.
4. Auditory Continuity.
VL 12.9 / Web 12.10
5. Experience.
The two melodies – Three Blind Mice and Mary had a Little Lamb – are more easily identified
by those familiar with the two melodies.
Web 12.11 / VL 12.11
Auditory Locatization and Organization - 19
3/21/5
Hearing and Vision.
Sound influences what we see
This page skipped in Summer 2016
Play VL 12.15 / Web 12.12
First, play with sound turned down.
Then, play with sound turned on.
I’m not thrilled with the other two Web VLs.
Areas of the brain activated by both vision and sound. G9 p 312, Figure 12.34
Neurons in the monkey parietal lobe were recorded during auditory and visual stimulation.
The red areas are areas in which stimulation – auditory in (a), visual in (b) – resulted in high
activity above the base rate.
The rightmost panel (c) shows that the two areas overlap considerably.
Auditory Locatization and Organization - 20
3/21/5
Echolocation by persons who are blind.
Thaler et al (2011) study
The researchers recorded from the ear canals of experienced echolocators.
So they recorded both the clicking sounds made by the echolocators and the echos of the clicks
as those echos reached their ears.
They then played the recordings of the clicks and their echos back to the echolocators and to
normally sighted volunteers and recorded cortical activity associated with what was heard.
They found . . .
1. Activity in the auditory cortexes of both.
2. Activity in the visual cortex of the echolocators, but not the sighted volunteer.
Visual cortex activity in echolocator, as he/she listened
to the clicks and their echos.
Absence of visual cortex activity in normally sighted
volunteer.
Auditory Locatization and Organization - 21
3/21/5