Computational Thinkers a theoretical conception as deep as it is - - PowerPoint PPT Presentation

computational thinkers
SMART_READER_LITE
LIVE PREVIEW

Computational Thinkers a theoretical conception as deep as it is - - PowerPoint PPT Presentation

Computational Thinkers a theoretical conception as deep as it is daring: namely, we are, at root, computers ourselves Haugeland, 1981 Mind as a computer As described by e.g.Craik (1943) Thinking involves manipulation of internal


slide-1
SLIDE 1

Computational Thinkers

“a theoretical conception as deep as it is daring: namely, we are, at root, computers ourselves” Haugeland, 1981

slide-2
SLIDE 2

Mind as a computer

  • As described by e.g.Craik (1943)

– Thinking involves manipulation of internal models of external situations – Explains ability to act towards things, beyond the current stimulus and history of reinforcement (challenging behaviourism) – Computer is more than metaphor: it has the exactly the right kind of capabilities for flexible model representation and manipulation

slide-3
SLIDE 3

However…

  • Should we consider all behaviour as falling under

this description, i.e. all nervous systems are computers? – The internal model has to be produced/updated and read out from: at minimum need computer plus transduction processes. – And is it right to assume all behaviour is described by:

sense - construct model - manipulate model - act?

slide-4
SLIDE 4

Bottom up view

  • Nervous systems perform a transfer

function from stimuli to actions

Nervous system f(s)=a Environment f(a)=s

slide-5
SLIDE 5

Mechanical example

  • Lotka (1925) described a simple toy insect

that detected and avoided the edges of table tops:

slide-6
SLIDE 6

Electronic example

Can get surprising capability from a couple of vacuum tubes and relays… Grey Walter’s ‘tortoise’1950

slide-7
SLIDE 7

Starts with: drive motor in series with lamp and turning motor full on; get cycloid movement that scans for light. Light input: passes through two amplifiers, switching relay 2, short circuit; so stops turning and drives double speed.

slide-8
SLIDE 8

Steers at increasingly shallow angle towards light source

slide-9
SLIDE 9

Strong light: switches relay 1, turning motor in series with lamp; turns smoothly away from light.

slide-10
SLIDE 10

Inspects different light sources Approaches then circles light

slide-11
SLIDE 11

If battery low: won’t reach threshold to turn away from light, so enters hutch to recharge. Replica tortoise (original hutch) Holland, 1995

slide-12
SLIDE 12

During scanning for light, own lamp is on. When moving to light, own lamp is off.

slide-13
SLIDE 13

Complex interactions of two robots ‘Recognises’ self in mirror and ‘dances’

slide-14
SLIDE 14

Shell collision: closes touch contact, output of amplifier 2 becomes input to amplifier 1; produces oscillator. Rapidly alternates driving and turning speeds, overriding effects of light input, till clear of obstacle.

slide-15
SLIDE 15

Can get round

  • bstacles to find

light. Also tends to push small obstacles out

  • f the way, gradually

clearing the area.

slide-16
SLIDE 16

Biological example

  • Female crickets recognise male calling

song and move towards it

slide-17
SLIDE 17

Reactive response to sound tested in treadmill experiments

slide-18
SLIDE 18

Pressure difference receiver

slide-19
SLIDE 19

Suggested neural circuit

slide-20
SLIDE 20
slide-21
SLIDE 21

When tested on the robot, can choose between sounds,

4.7Hz 4.7Hz 4.7Hz 6.7Hz 4.7Hz 6.7Hz

  • preferring correct carrier frequency

,

slide-22
SLIDE 22

When tested on the robot, can choose between sounds,

  • preferring correct temporal pattern

,

slide-23
SLIDE 23

Should this be called computation?

  • Can choose to view any of these

examples as ‘encoding’ and ‘processing’

  • f information (about table edge, light

direction, sound location…)

  • But if this is ‘computation’, then so is every

kind of causal process, or transformation.

  • So we haven’t said anything “deep and

daring” about minds and brains by identifying them as computers.

slide-24
SLIDE 24

A reasonable objection

  • The simple behaviours I have described are not

the kind of behaviours Craik was talking about.

  • Perhaps insects are not real ‘thinkers’. That

simple nervous systems are not computing (in any interesting sense) does not necessarily mean that no part of our nervous system is computing (in some interesting sense).

  • But then we need to identify the tasks and

nervous system structures that do require a computational interpretation…