What do hidden representations learn? Other animals dont like onions - - PowerPoint PPT Presentation

what do hidden representations learn other animals don t
SMART_READER_LITE
LIVE PREVIEW

What do hidden representations learn? Other animals dont like onions - - PowerPoint PPT Presentation

What do hidden representations learn? Other animals dont like onions (but primates do) Plaut and Shallice (1993) Mapped orthography to semantics (unrelated similarities) Compared similarities among hidden representations to those among


slide-1
SLIDE 1

What do hidden representations learn?

Plaut and Shallice (1993) Mapped orthography to semantics (unrelated similarities) Compared similarities among hidden representations to those among orthographic and semantic representations (over settling) Hidden representations “split the difference” between input and

  • utput similarity

16 / 25

Other animals don’t like onions (but primates do)

No correlation for “specific” weights because inputs vary No correlation for “more general” weights because output varies Only intermediate “general” weights build up due to correlations

25 / 29

slide-2
SLIDE 2

Semantic hierarchy Progressive differentiation in development Progressive deteroriation in semantic dementia Rumelhart and Todd (1993)

slide-3
SLIDE 3

Progressive differentiation of concepts Internal representations “Basic” level Perceptual to conceptual shift

slide-4
SLIDE 4

Semantic memory

Unitary, amodal semantic system

  • General conceptual knowledge

abstracted from a large number of individual episodes or experiences

(Tulving, 1972).

  • Mediates among multiple input and
  • utput modalities
  • Can be selectively impaired by

brain damage, usually to the anterior temporal lobes (Warrington,

1975).

Vision Touch Hearing Action Writing Speaking Semantics Challenges

  • Modality-specific effects in semantic priming
  • Category- and modality-specific semantic deficits
  • Modality-specific naming disorders (e.g., optic aphasia)

Optic aphasia

Selective impairment in visual object naming

  • Brain damage to left medial occipital lobe (visual cortex and underlying white matter)
  • Not visual agnosia—relatively preserved visual gesturing and other tests of visual

comprehension

  • Not general anomia—relatively preserved naming from other modalities (e.g., touch,

spoken definitions)

  • Relatively preserved naming of actions associated with visually presented objects

(Manning & Campbell, 1996)

% Correct Performance Visual Visual Tactile Action Study Naming Gesturing Naming Naming Lhermitte & Beauvois (1973) 73 100 91 Teixeira Ferreira et al. (1997) 53 95 81 75 Manning & Campbell (1996) 27 75 90 67 Coslett & Saffran (1989) 50 92

Analogous selective naming deficits have been observed for tactile input (Beauvois et al.,

1978) and for auditory input (Denes & Semenza, 1975)

Challenge to unitary semantics account

Post-semantic lesion

  • No basis for sensitivity to input modality

Action (Gesturing) Phonology (Naming) Visual Input Verbal Input Tactile Input Semantics

Semantic lesion

  • Would impair visual gesturing and non-visual

naming

Action (Gesturing) Phonology (Naming) Visual Input Verbal Input Tactile Input Semantics

Pre-semantic lesion

  • Would impair visual gesturing (and other

measures of comprehension) – But might be preserved relative to naming due to

priviledged access (Caramazza et al., 1990)

Action (Gesturing) Phonology (Naming) Visual Input Verbal Input Tactile Input Semantics

Alternative view of semantic organization

Multiple modality-specific semantic systems (Beauvois, 1982; Lhermitte &

Beauvois, 1973; Shallice, 1987; Warrington, 1975)

Action (Gesturing) Phonology (Naming) Visual Semantics Tactile Semantics Verbal Semantics Visual Input Verbal Input Tactile Input

Optic aphasia: Disconnection from visual to verbal semantics Problems

  • Unparsimonious and post-hoc
  • Poor accounts of acquisition, cross-modal generalization and priming
  • No account of relative sparing of visual action naming
slide-5
SLIDE 5

Current approach

  • The semantic system operates according to connectionist/parallel distributed

processing (PDP) principles: – Processing: Responses are generated by the interactions of large numbers of simple,

neuron-like processing units.

– Representation: Within each modality, similar objects are represented by overlapping

distributed patterns of activity.

– Learning: Knowledge is encoded as weights on connections between units, adjusted

gradually based on task performance.

  • Semantic representations develop a graded degree of modality-specific

specialization in learning to mediate between multiple input and output modalities.

  • Graded specialization derives from two factors:
  • 1. Task systematicity: Whether similar inputs map to similar outputs

⇒ Naming is an unsystematic task

  • 2. Topographic bias: Learning favors “short” connections (Jacobs & Jordan, 1992)

⇒ Mappings rely most on regions of semantics “near” relevant modalities Vision Action (Gesturing) Phonology (Naming) Touch Task

  • Continuous recurrent

attractor network

  • Two input groups

(Vision, Touch) equidistant from two

  • utput groups (Action,

Phonology)

  • Task units: object vs.

action tasks

Tasks

  • Naming objects from vision or touch
  • Naming and gesturing action associated with objects from vision or touch

Stimulus (Vision or Touch) Task

Phonology Action

  • bject

“bed” – action

“sleep”

Training procedure

  • Object (N = 100), modality of presentation (Vision vs. Touch), and task (object vs.

action) chosen randomly during training

  • Activations clamped on appropriate input modality; network settled for 5.0 units of

time (τ = 0.2); error injected only over last time unit

  • Error derivatives for each weight calculated by back-propagation-through-time

adapted for continuous-time networks (Pearlmutter, 1989)

  • Weight changes scaled by Gaussian function (SD = 10) of connection length

(cf. Jacobs & Jordan, 1992)

  • 110,000 total object presentations (≈275 per condition);

all output activations on correct side of 0.5 for all objects and tasks

slide-6
SLIDE 6

Vision Action (Gesturing) Phonology (Naming)

size of white square = unit activity level

Touch Task

1 3 5 7 9 11 13 15

X Position

0.0 0.2 0.4 0.6 0.8 1.0 Weight-Change Factor

Vision Inputs Touch Inputs

Acquisition

10 20 30 40 50 60 70 80 90 100

Object Presentations (x 1000)

10 20 30 40 50 60 70 80 90 100

Percent Correct

Visual Gesturing Visual Action Naming Visual Object Naming

Semantic similarity

Mean correlations among pairs of semantic representations generated by each object in each input modality (note: relatedness is relative to visual categories).

Identical Same Category Unrelated

Item Pairs

0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Mean Semantic Correlation Same Modality Different Modalities

.63 (.08) .36 (.19) .22 (.16) .03 (.12).02 (.11)

  • An object is most similar to itself regardless of modality of presentation

(but cross-modal representations are not identical)

  • Accounts for reduction in semantic priming with cross-modal presentation

Magnitude of incoming weights to semantics

Vision Input Vision Touch Input Touch

1 3 5 7 9 11 13 15

X Position

0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Incoming Weight Magnitude Visual Input Tactile Input

slide-7
SLIDE 7

Topographic lesions

Lesions applied to Vision ⇒ Semantics connections

  • Probability of removing each connection is a Gaussian function of the distance of unit from

lesion location; SD of Gaussian controls severity (1.5 below).

Vision Action (Gesturing) Phonology (Naming) Touch Task

Effects of lesion location

Performance on visual naming and visual gesturing after lesions to Vision ⇒ Semantics connections centered at each location (10 repetitions each; SD = 1.5). Visual Naming Action Vision Phonology Visual Gesturing Action Touch Phonology

Effects of lesion location

% correct visual gesturing − % correct visual naming (white: gesturing > naming; black: naming > gesturing) Action Vision Touch Phonology

Effects of lesion severity

0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0

Vision−to−Semantics Lesion Severity (SD)

10 20 30 40 50 60 70 80 90 100

Percent Correct

Tactile Naming Visual Gesturing Visual Action Naming Visual Object Naming

1 2 3 4

  • Highly selective impairment
  • f visual object naming

relative to visual gesturing and tactile naming

  • Relative preservation of

visual action naming

% Correct Performance Visual Visual Tactile Action Study Naming Gesturing Naming Naming

  • 1. Lhermitte & Beauvois (1973)

73 100 91

  • 2. Teixeira Ferreira et al. (1997)

53 95 81 75

  • 3. Manning & Campbell (1996)

27 75 90 67

  • 4. Coslett & Saffran (1989)

50 92

slide-8
SLIDE 8

Control simulation: No topographic bias

  • Exact replication of simulation except without topographic bias

0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0

Vision−to−Semantics Lesion Severity (SD)

10 20 30 40 50 60 70 80 90 100

Percent Correct

Tactile Naming Visual Gesturing Visual Action Naming Visual Object Naming

1 2 3 4

  • Relative degrees of impairment due to differences in task systematicity
  • Effect is not as large as in patients

Generating object names vs. action names

Performance after lesions to semantics (SD = 2.0). Visual Object Naming Action Vision Visual Action Naming Action Touch

  • Generating names of actions associated with objects involves interactions with

Action representations.

Modality-specific impairment of grammatical categories

Selective impairment on nouns vs. verbs restricted to a particular input or output modality (Caramazza & Hillis, 1991; Hillis & Caramazza, 1995; Rapp & Caramazza, 1997)

  • Patient HW (Caramazza & Hillis, 1991)

– nouns > verbs in spoken output but not in written output

  • Patient EBA (Hillis & Caramazza, 1995)

– two lesions: left frontal and left temporal

Modality-specific impairment of grammatical categories

Account 1: Modality-specific lexical representations divided by grammatical category

(Caramazza & Hillis, 1991; Hillis & Caramazza, 1995; Rapp & Caramazza, 1997)

  • rthographic

lexical form nouns verbs A P P L E G I R L E A T G I V E phonological lexical form / it / / gIv / nouns verbs / girl / / apl / [eat]V [give]V [girl]N [apple]N semantic/syntactic lexical specification nouns + verbs − nouns − verbs +

slide-9
SLIDE 9

Modality-specific impairment of grammatical categories

Account 2: Modality-specific access pathways divided by grammatical category (Rapp

& Caramazza, 1997)

  • rthographic

lexical form phonological lexical form / it / / gIv / nouns verbs nouns verbs [eat]V [give]V [girl]N [apple]N semantic/syntactic lexical specification A P P L E G I R L nouns + verbs − nouns − verbs +

Modality-specific impairment of grammatical categories

Object naming − Action naming (white: nouns > verbs; black: verbs > nouns) Incoming Lesions Vision ⇒ Semantics (SD = 2.0) Vision Outgoing Lesions Semantics ⇒ Phonology (SD = 1.5) Touch

  • Nouns depend more on semantics near Vision/Phonology; Verbs depend more on

semantics near Action/Touch

  • Lesions to incoming/outgoing connections produce modality-specific grammatical category

impairments

Modality-specific impairment of grammatical categories

[4,14] [10,5] [5,9] [14,6]

20 40 60 80 100

Percent Correct Incoming Lesions Outgoing Lesions

Nouns (visual object naming) Verbs (visual action naming)

Summary

  • Differences in task systematicity and a topographic bias on learning produce a

graded degree of modality-specific specialization within semantics.

⇒ The network does not develop separate modality-specific semantic

systems—both visual and tactile input engage all of semantics.

  • Damage to connections from vision to regions of semantics near phonology

produces a highly selective visual naming deficit as observed in optic aphasia.

  • The relative sparing of naming visual actions in optic aphasia results from the

preserved support of action representations.

  • Modality-specific impairments to nouns vs. verbs can arise from lesions to/from

regions of semantics partially specialized for these grammatical classes.