Listening without hearing Nadia M. Biassou MD, PhD Senior Research - - PowerPoint PPT Presentation

listening without hearing
SMART_READER_LITE
LIVE PREVIEW

Listening without hearing Nadia M. Biassou MD, PhD Senior Research - - PowerPoint PPT Presentation

Listening without hearing Nadia M. Biassou MD, PhD Senior Research Physician Dept of Radiology Division of Neuroradiology NIH Clinical Center and Senior Fellow Linguistics Data Consortium Graduate School of Arts and Sciences University of


slide-1
SLIDE 1

Listening without hearing

Nadia M. Biassou MD, PhD

Senior Research Physician Dept of Radiology Division of Neuroradiology NIH Clinical Center and Senior Fellow Linguistics Data Consortium Graduate School of Arts and Sciences University of Pennsylvania

slide-2
SLIDE 2

Financial Disclosure

l None

slide-3
SLIDE 3

Learning Objectives

l The basics of the physics of speech l What is currently known about conscious

neurobiologic speech perception?

l Can unconscious speech perception by

reliably measured?

l What can its study tell us about the general

nature of speech perception and about the human brain that processes it?

slide-4
SLIDE 4

Introduction

l What is speech and

why is it special?

slide-5
SLIDE 5

Speech is the entryway to human linguistic communication

slide-6
SLIDE 6

1 Pa

  • 2. Ta
  • 3. Ka
slide-7
SLIDE 7
slide-8
SLIDE 8

Formant frequency

l F0 is called the

fundamental frequency and represents the frequency of vocal cord oscillation

slide-9
SLIDE 9

Formant frequencies

l Oscillation of vocal

cords and its harmonics

l F0

1

l F1

3

l F2

5

l F3

7

slide-10
SLIDE 10
slide-11
SLIDE 11
slide-12
SLIDE 12

The speech waveform

l The production of any

sound during word production is simultaneously influenced by the sounds that precede and follow it.

l Liberman et al., 1957

slide-13
SLIDE 13

Coarticulation of sounds

l “ebb” vs. “egg”

slide-14
SLIDE 14
slide-15
SLIDE 15
slide-16
SLIDE 16

The speech spectrograms: formant frequency transitions

l The formant

frequencies transitions reflect coarticulation

slide-17
SLIDE 17

Does the brain listen to every acoustic variation during speech perception?

slide-18
SLIDE 18

Bottom Up processes

Bottom-up processing refers to processing sensory information as it is coming in

slide-19
SLIDE 19

TASK

l PART 1: Actively decided whether real and

nonreal words are real words of English, half of the real and nonreal words are acoustically manipulated

slide-20
SLIDE 20

STIMULI

l EXPERIMENT 1 l 40 REAL WORDS l HALF ARE

ACOUSTICALLY MANIPUALTED

l HALF ARE NON-

MANIPULATED

l 40 NONREAL WORDS l HALF ARE

ACOUSTICALLY MANIPULATED

l HALF ARE NON-

MANIPULATED

slide-21
SLIDE 21

RESULTS

10 20 30 40 50 60 70 80 realwords nonwords

l The brain takes

74msecs longer to process the acoustically manipulated realwords, even though subjects could not consciously distinguish the word types

slide-22
SLIDE 22

Sensory changes affect higher

  • rder language processing

BLUMSTEIN and colleagues

  • LEXICAL DECISION TASKS IN WHICH LEXICAL

ITEMS WERE MANIPULATED ACOUSTIC GAP DETECTION (I.E. VOT) below the conscious level

  • Sensory alteration can affect activation semantic priming and

lexical access.

  • FARAH et al argue that words may also be stored with visual

associated information.

slide-23
SLIDE 23

What are the neural networks that subserve subconscious processing of speech during conditions of increased effort?

l

LEFT INFERIOR FRONTAL CORTEX, ANTERIOR CINGULATE AND THALAMUS

l

POSTERIOR SUPERIOR TEMPORAL LOBES BILATERALLY

l

OCCIPITAL LOBES

l

LEFT CEREBELLUM

slide-24
SLIDE 24

PART 2: Passively listen to real and nonreal words of English, half of which had been acoustically manipulated. STIMULI- PARTS 1 & 2 ARE MATCHED IN WORD FREQUENCY, WORD LENGTH, NUMBER OF SYLLABLES, AND IMAGEABILITY

l EXPERIMENT 2 l 40 REAL WORDS l HALF ARE

ACOUSTICALLY MANIPULATED

l HALF ARE NOT

MANIPULATED

l EXPERIMENT 2 l 40 NONREAL WORDS l HALF ARE

ACOUSTICALLY MANIPULATED

l HALF ARE NOT

MANIPULATED

slide-25
SLIDE 25

Are the same networks activated in conditions of less effort?

l Activation in (b) posterior

superior temporal lobes and anterior cingulate are sufficiently robust even for the passive presentation of subconsciously manipulated realwords. But right frontal and right parietal lobe networks are activated

slide-26
SLIDE 26

BUT IS SPEECH PERCEPTION ALL BOTTOM UP?

slide-27
SLIDE 27

Top Down Processes

Visual Cues and Speech Perception McGuck Effect Baysan, U. (July 2017) "McGurk Effect" in F. Macpherson (ed.), The Illusions Index. Retrieved from https://www.illusionsindex.org/i/ mcgurk-effect.

slide-28
SLIDE 28

McGuck Effect

BBC – Horizon: Is Seeing Believing Nov 2010

slide-29
SLIDE 29

TOP DOWN PROCESSES

CONTEXT AND SPEECH PERCEPTION PHONEME RESTORATION (Warren & Warren 1970)

slide-30
SLIDE 30

Phoneme Restoration Effect

Suboptimal environment input is

  • verridden by context of speech

to hear stimuli that is in fact absent. “The State Governors met with their respective legislatures convening in the capitol city.” “The State Governors met with their respective le…latures convening in the capitol city.”

slide-31
SLIDE 31

Ed Chang and colleagues

Leonard et al 2016, Nature Communications, 7:13619

slide-32
SLIDE 32

Top Down Processes

Yanny vs Laurel Left: YANNY Right: LAUREL Middle spectrogram is a simulated ambiguous spectrogram – BUT listeners hear Yanny or Laurel

slide-33
SLIDE 33

PERCEPTION is the point of contact between multisensory information:

BOTTOM UP (Objective)

l Processing of

sensory input

l Can affect higher

  • rder cognitive and

linguistic processes such as vision and semantics.

TOP DOWN (Subjective)

l Visual Input l Context l Linguistic

phonotactics (the language that you speak) can all affect the interpretation of sensory cues.

slide-34
SLIDE 34

This point of contact is dynamic in time and in space

l Different neural networks can process the

same types of speech cues depending on the conditions under which the cues are being processed.

l Neural networks involved in processing

subconscious fine grain speech cues can involve the right hemisphere under passive listening (or lighter attentional load)

slide-35
SLIDE 35

Attentional networks are always being recruited to varying degrees?

l Even for the passive listening of speech

cues.

slide-36
SLIDE 36

We hear want we want to hear This doesn’t

  • nly apply to

cats or dogs!

slide-37
SLIDE 37

Resting State fMRI

l Passive neural networks

may not be fully representative of the neural networks that subserve linguistic/cognitive processes because the network dynamics change depending on the attentional load to achieve the task at hand. It is NOT solely driven by the stimulus.

slide-38
SLIDE 38

CLINICAL IMPLICATIONS

l Language mapping for

neurosurgery should reflect natural state of language processing as closely as possible including masked stimuli

slide-39
SLIDE 39

New research

l Normal aging l Alzheimer’s and other neurodegenerative

disorders

l Autism l Is there a genetic basis for the balance

between objective and subjective speech perception?

slide-40
SLIDE 40

Future research

l Can we develop new wearable technologies

that can diagnosis changes in the processing

  • f sensory input in the preclinical stage of

disease?

slide-41
SLIDE 41

l “Our imaginations are limited by the

knowledge that we currently possess”

  • Helen Neville (IRCS Talk, University of

Pennsylvania, 1995)

slide-42
SLIDE 42

l THANK YOU!