Observations and Inspirations
Mutual Inspirations between Cognitive and Statistical Sciences
Shakir Mohamed
Research Scientist, DeepMind @shakir_za Sheffield Machine Learning Research Retreat 2017 shakir@deepmind.com
Observations and Inspirations Mutual Inspirations between Cognitive - - PowerPoint PPT Presentation
Observations and Inspirations Mutual Inspirations between Cognitive and Statistical Sciences Shakir Mohamed Research Scientist, DeepMind She ffi eld Machine Learning shakir@deepmind.com Research Retreat 2017 @shakir_za Abstract Observations
Research Scientist, DeepMind @shakir_za Sheffield Machine Learning Research Retreat 2017 shakir@deepmind.com
Observations & Inspirations: The Mutual Inspirations between Cognitive and Statistical Sciences Where do we obtain our inspiration in cognitive science? And in Machine Learning? These questions look at the parallels between these two fields. Fortunately, seeking out the parallels between minds and machines is one of our long-established scientific traditions, and this talk will explore the exchange of ideas between the two fields. The parallels between the cognitive and statistical sciences appear in all aspects of our practice, from how we conceptualise our problems, to the ways in which we test them, and the language we use in communication. One of these mutually useful tools are the conceptual frameworks used in the two fields. In cognitive science the most established frameworks are the classical cognitive architecture and Marr's levels of analysis, and similarly in machine learning, that of Box's loop and the model-inference-algorithm paradigm; these will be our starting point. The parallels between our fields appear in other more obvious forms, from cognitive revolutions and dogmas of information processing, to neural networks and embodied robotics. Recurring principles appear: prediction, sparsity, uncertainty, modularity, abduction, complementarity; and we'll explore several examples of these principles. From my own experience, we'll explore the probabilistic tools that connect to one-shot generalisation, grounded cognition, intrinsic motivation, and memory. Ultimately, these connections allow us to go from observation to inspiration: to make observations of cognitive and statistical phenomena, and, inspired by them, to strive towards a deeper understanding
2
3
machine learning, AI.
processing, statistical physics
4
5
6
Classical Cognitive Architecture
7
Marr’s Levels
8
Sun’s Phenomenological Levels
9
Problem Machine Learning Core
Data Implement and Test
Inference
Application/ Production
Model
Problem Machine Learning Core
Data Implement and Test
Inference
Application/ Production
Model
10
Principles
11
A given model and learning principle can be implemented in many ways.
zi xi xj zj xk
Restricted Boltzmann Machine + maximum likelihood
z x
f(z)
Latent variable model + variational inference
Convolutional neural network + penalised maximum likelihood
12
13
dopamine and learning.
14
15
robustness.
16
hippocampus for episodic memory and abstract representations.
shifts to dopaminergic neurons in striatum.
17
Primary Secondary
learning, risk, value-at-risk and sensitivity.
understanding.
decisions attitudes.
18
19
20
21
Original Oxygen/Swimmers Score Score/Lives Moving Up Moving Left
22
23
Model p(x |z) log p(x|z)
Prior p(z) log p(z)
Inference q(z |x) H[q(z)] z
Data x
Penalty Reconstruction
qφ(z)
KL[q(z|y)kp(z|y)]
Approximation class True posterior
Variational inference is scalable and robust as a default approach for inference in deep probabilistic models.
24
differentiable, like a graphics engine.
colour channels, volumes, time.
convolutions.
Model p(x |z) log p(x|z)
Prior p(z1) Prior p(z2) Prior p(zT) State h(z) State h(z) State h(z)
Inference q(z1⎜x) State s(x) State s(x) Inference q(z2⎜x) State s(x) Inference q(zT⎜x)
Prior
p(zwhere
1
) p(zwhat
1
)
p(zcont
1
)
25
26
27
Action-conditional and latent-only transitions. Grounded representations in actions and observations, using simulation to support grounding.
28
Data
xt-1 State st State st-1 Action at-1 mt
Data
xt State st+1 mt+1
Data
xt+1 Action at Action at+!
mt-1
29
30
Computational perception-action loop Biological perception-action loop
31
Escaping a Predator
1 1 2 3 4 5 6 6
True MI
32
33
34
Recall One-shot generalisation
35
36
Macro-actions and Planning Visual Concept Learning World Simulation Data-efficient Learning Exploration Complementary Learning Relational learning Hypothesis formation Causal Reasoning
Research Scientist, DeepMind @shakir_za Sheffield Machine Learning Research Retreat 2017 shakir@deepmind.com