Introduction Jean Petitot CAMS, EHESS, Paris J. Petitot - - PowerPoint PPT Presentation

introduction
SMART_READER_LITE
LIVE PREVIEW

Introduction Jean Petitot CAMS, EHESS, Paris J. Petitot - - PowerPoint PPT Presentation

Geometrical Models in Vision IHP, October 22nd-24th, 2014 Introduction Jean Petitot CAMS, EHESS, Paris J. Petitot Introduction Acknowledgements Thanks to the organizers of the Trimester for having proposed such a workshop, to the IHP


slide-1
SLIDE 1

Geometrical Models in Vision IHP, October 22nd-24th, 2014

Introduction

Jean Petitot CAMS, EHESS, Paris

  • J. Petitot

Introduction

slide-2
SLIDE 2

Acknowledgements

Thanks to the organizers of the Trimester for having proposed such a workshop, to the IHP team, to Francesca Carlotta Chittaro (Universit´ e du Sud Toulon-Var) for the organization.

  • J. Petitot

Introduction

slide-3
SLIDE 3

Sponsors

CNRS INRIA Marie Curie Initial Training Net MAnET (Metric Analysis for Emergent Technologies), Department of Mathematics of the University of Bologna. Seventh European Framework Programme Team GECO (GEometric COntrol design) of the CMAP (Ecole Polytechnique) and the ERC GeCo Methods (U. Boscain) CAMS (Centre d’Analyse et de Math´ ematique Sociales), EHESS. Fondation Jacques Hadamard

  • J. Petitot

Introduction

slide-4
SLIDE 4

Neurogeometry

In this short introduction I only will try to explain what can be the link between natural vision of mammals and sub-Riemannian

  • geometry. I will introduce some very basic and elementary

experimental facts and theoretical concepts. QUESTION: How the visual brain can be a neural geometric engine?

  • J. Petitot

Introduction

slide-5
SLIDE 5

Neurogeometry

Neurons are very local detectors and even “point processors” (Jan Koenderink). They can code a numerical value with their firing

  • rate. How such neurons can integrate differential features into

global geometric structures? The classical intuition of integration cannot be used. We need a much deeper concept of integrability.

  • J. Petitot

Introduction

slide-6
SLIDE 6

Neurogeometry

ANSWER Low dimensional jet spaces are neurally implemented and jet spaces are naturally endowed with integrability conditions and SR structures. Jets are “prolongations” in the sense of Cartan. The visual brain is a “Lie-Cartan” geometric engine.

  • J. Petitot

Introduction

slide-7
SLIDE 7

The visual brain

Here is an image of the human brain. It shows the neural pathways from the retina to the lateral geniculate nucleus (thalamic relay) and then to the occipital primary visual cortex (area V 1).

  • J. Petitot

Introduction

slide-8
SLIDE 8

Receptive fields and receptive profiles

Through this very complex connectivity, the neurons of the cortical layers are retinotopically (topographically) connected to small domains of the retina called their receptive fields (minimal discharge field). In a very rough linear approximation, neurons act as filters on the

  • ptical signal transduced by the photoreceptors of the retina. Their

transfer function is called their receptive profile by neurophysiologists.

  • J. Petitot

Introduction

slide-9
SLIDE 9

Receptive fields and receptive profiles

Typically, in the layer of V 1 which receives axons from the lateral geniculate nucleus (layer IV in higher mammals as cats and monkeys), there are neurons called “simple” whose receptive profile is highly anisotropic. They are well modeled by Gabor patches or derivatives of Gaussian.

  • J. Petitot

Introduction

slide-10
SLIDE 10

Example of receptive field

Here is an example. Left: Recording of level sets (Gregory DeAngelis, Berkeley). Right: model (third derivative ϕ(x, y) = ∂3G

∂x3 ).

  • J. Petitot

Introduction

slide-11
SLIDE 11

Wavelet analysis

The filtering of the signal is like a wavelet analysis using oriented wavelets.

  • J. Petitot

Introduction

slide-12
SLIDE 12

Contact elements

Simple neurons are parametrized by triples (a, p) where a = (x, y) is a position in the retina or the visual field (that you can identify to R2) and p is an orientation at a. These triples are contact elements that is numerical values of 1-jets of smooth plane curves. Experiments show that the implemented (a, p) are regularly distributed in V = R2 × P. So V can be considered as an idealized continuous approximation

  • f the real V 1.
  • J. Petitot

Introduction

slide-13
SLIDE 13

Functional architectures

The key point is that neurons are connected and it is this connectivity which can be described by geometric structures on V. In areas as V 1 the connectivity is extremely specific and is called a functional architecture. In neural nets : functional architecture ⇐ ⇒ geometric structure.

  • J. Petitot

Introduction

slide-14
SLIDE 14

Connections

There are two main classes of connections concerning a cortical layer such as the layer IV of V 1: “vertical” connections relating the layer to the LGN and to

  • ther cortical layers and,

“horizontal” cortico-cortical connections inside the layer itself.

  • J. Petitot

Introduction

slide-15
SLIDE 15

“Vertical” connections and hypercolumns

For the vertical connections, the 1981 Nobel Prizes David Hubel and Torsten Wiesel have shown in the 60s that the neurons detecting all the orientations p at the same position a constitute an anatomically well defined small structure called an “orientation hypercolumn”. This means that the fiber bundle π : V = R2 × P → R2 is neurally implemented.

  • J. Petitot

Introduction

slide-16
SLIDE 16

Fiber bundles and “engrafted” variables

The intuition (not the mathematical concept) of a fiber bundle is explicit in Hubel: “What the cortex does is map not just two but many variables on its two-dimensional surface. It does so by selecting as the basic parameters the two variables that specify the visual field coordinates (...), and on this map it engrafts other variables, such as orientation...”

  • J. Petitot

Introduction

slide-17
SLIDE 17

Pinwheels

But what is most remarkable is that the 3D structure V is implemented in a 2D layer. This dimensional collapse is realized via what is called a pinwheel structure. Here is the pinwheel structure of the area V 1 of a tree-shrew (tupaya) obtained by “in vivo optical imaging” (William Bosking).

  • J. Petitot

Introduction

slide-18
SLIDE 18

Pinwheels figure

  • J. Petitot

Introduction

slide-19
SLIDE 19

Pinwheels figure

The plane is V 1, A colored point represents the mean of a small group of real neurons (mesoscale). Colors code for the preferred orientation at each point. The field of isochromatic lines (i.e. iso-orientation lines) is

  • rganized by a lattice of singular points called pinwheels

where all orientations meet. Pinwheels have a chirality. Adjacent pinwheels have opposed chirality.

  • J. Petitot

Introduction

slide-20
SLIDE 20

Dislocations of phase fields

There are beautiful models of pinwheels and of their density. They are analogous to dislocations of phase fields in optics.

  • J. Petitot

Introduction

slide-21
SLIDE 21

Pinwheels and the fiber bundle V

To get a 2D pinwheel structure from the 3D fiber bundle V, you discretize the positions a, compactify ` a la Kaluza-Klein the fiber Pa, and project it around a in the base plane. In the other direction you blow-up the center of the pinwheel.

  • J. Petitot

Introduction

slide-22
SLIDE 22

“Horizontal” cortico-cotical connections

One of the deepest experimental results concerns the long-range excitatory “horizontal” cortico-cortical connections. Bosking’s image show the diffusion of a marker (biocytin) along them (black marks). The injection site is upper-left in a green domain.

  • J. Petitot

Introduction

slide-23
SLIDE 23

Horizontal connections

  • J. Petitot

Introduction

slide-24
SLIDE 24

“Horizontal” cortico-cotical connections

There are two main results:

1 the marked axons and synaptic buttons cluster in domains of

the same color (same orientation), which means that horizontal connections implement neurally a parallel transport.

2 the global clustering along the upper-left bottom-right

diagonal means that horizontal connections connect neurons with almost parallel and almost aligned orientations.

  • J. Petitot

Introduction

slide-25
SLIDE 25

Co-axial alignement

“The system of long-range horizontal connections can be summarized as preferentially linking neurons with co-oriented, co-axially aligned receptive fields.”(W. Bosking)

  • J. Petitot

Introduction

slide-26
SLIDE 26

The association field

This result is corroborated by experiments in psychophysics about what is called the association field (David Field, Anthony Hayes and Robert Hess). The experiments concern the pop-out (the perceptive saliency) of almost aligned Gabor patches. Here is an example. In a background of random patches, you insert a set of almost aligned patches and a global curve emerge.

  • J. Petitot

Introduction

slide-27
SLIDE 27

The association field

  • J. Petitot

Introduction

slide-28
SLIDE 28

Curve integration

So a set of contact elements ci = (ai, pi) is perceived as a global curve (what is called a binding) if the orientations pi are tangent to a regular curve γ interpolating as straightly as possible between the positions ai. These “joint constraints of position and orientation” correspond to the horizontal connections. horizontal connections, parallel transport, coaxiality, binding, propagation of coherent activity, synchronization, global pop out, saliency ⇐ ⇒ integration of contact elements

  • J. Petitot

Introduction

slide-29
SLIDE 29

The contact structure of 1-jets

For the 3D model V parametrizing the simple neurons, these experimental results show that a skew curve Γ = v(s) = (a(s), p(s)) = (x (s) , y(s), p(s)) in V is perceived as a globally coherent curve (via binding and pop-out) in the base plane R2 if and only if it is the Legendrian lift of its projection γ, that is iff it is an integral curve of the contact structure K = ker(ω) of V (ω = dy − pdx). For curves, to work in V 1 is to work with Legendrian curves. The functional architecture is represented by a 1-form ω = dy − pdx.

  • J. Petitot

Introduction

slide-30
SLIDE 30

Contact structure and point processors

The contact structure K of the space of 1-jets V is neurally implemented. This explains how “point processors” as neurons can do differential geometry.

  • J. Petitot

Introduction

slide-31
SLIDE 31

The contact structure of 1-jets

K is left-invariant for a group law making V isomorphic to the polarized Heisenberg group. The Euclidean group SE(2) = R2 > ⊳ SO(2) of direct isometries of the plane acts naturally on V and it is therefore better to work in this principal bundle. This model is only a first model. If you want to take into account curvature detectors, you will have to use 2-jets and therefore the Engel structure. Etc.

  • J. Petitot

Introduction

slide-32
SLIDE 32

Completion and Gestalttheory

That the functional architecture implements a contact structure explains some strange perceptive phenomena of very long range completion of images. Consider for example this Kanizsa square.

  • J. Petitot

Introduction

slide-33
SLIDE 33

Kanizsa square

  • J. Petitot

Introduction

slide-34
SLIDE 34

Kanizsa square

The red pacmen induce very long-range curved illusory contours (what is called modal completion). Moreover, these contours act as boundaries for a diffusion of color inside the square (what is called the “neon” or “watercolor effect”). Consider also the Koffka cross (Ehrenstein illusion):

  • J. Petitot

Introduction

slide-35
SLIDE 35

Ehrenstein illusion

  • J. Petitot

Introduction

slide-36
SLIDE 36

Ehrenstein illusion

The end points activate locally (area V2 is necessary) the

  • rthogonal orientations.

These very sparse local activation induce a very long-range global modal subjective contour. Moreover, subjects perceive alternately circles and squares, which means that there exists a competition between two completion strategies: circle: illusory contours with a maximal diffusion of curvature, square: piecewise linear illusory contours (curvature = 0) with corners (singularities of curvature).

  • J. Petitot

Introduction

slide-37
SLIDE 37

Variational models

To explain these spectacular completion phenomena, variational models have been introduced since the late 70s. They were models minimizing an energy along curves γ in the base plane R2. The best known is the elastica model proposed in 1992 by David

  • Mumford. The energy to minimize is:

E =

  • γ(ακ2 + β)ds
  • J. Petitot

Introduction

slide-38
SLIDE 38

Geodesic models

But in what concerns neural models (and not only 2D image processing) it is natural to work in V 1, that is with the contact structure and the Legendrian curves of V. Hence the natural idea of introducing sub-Riemannian geodesic models for curve completion and illusory modal contours. For the completion of corrupted images (inpainting) it is therefore natural to use diffusion along the horizontal connections, that is sub-Riemannian Laplacian, heat kernel, etc.

  • J. Petitot

Introduction

slide-39
SLIDE 39

Geometry and Vision

This Introduction focused on a very special problem. Of course the problem of geometry of vision is huge and highly diversified and goes far beyond the link between neural functional architecture and sub-Riemannian geometry. The workshop will show how huge and diversified it is. THANKS

  • J. Petitot

Introduction