Topics in Quantum Machine Learning Vedran Dunjko - - PowerPoint PPT Presentation

topics in quantum machine learning
SMART_READER_LITE
LIVE PREVIEW

Topics in Quantum Machine Learning Vedran Dunjko - - PowerPoint PPT Presentation

Topics in Quantum Machine Learning Vedran Dunjko v.dunjko@liacs.leidenuniv.nl 1 Q uantum M achine L earning (QML) Q uantum I nformation M achine L earning/ AI P rocessing ( QIP ) (ML/AI) ML QIP (quantum-applied ML) [74] QIP ML


slide-1
SLIDE 1

Topics in Quantum Machine Learning

Vedran Dunjko

v.dunjko@liacs.leidenuniv.nl

1

slide-2
SLIDE 2

ML→QIP (quantum-applied ML) [’74] QIP→ML (quantum-enhanced ML) [‘94] QIP↭ML (quantum-generalized learning) [‘00]

ML-insipred QM/QIP

Physics inspired ML/AI

Quantum Information Processing (QIP) Machine Learning/AI (ML/AI)

Quantum Machine Learning (QML)

2

slide-3
SLIDE 3

3

Machine learning is not one thing. AI is not even a few things.

AI

supervised learning unsupervised learning

  • nline learning

generative models reinforcement learning deep learning statistical learning non-parametric learning parametric learning local search Symbolic AI computational learning theory control theory non-convex

  • ptimization

sequential decision theory

ML

big data analysis

slide-4
SLIDE 4

4

Quantum-enhanced ML is even more things

AI

supervised learning unsupervised learning

  • nline learning

generative models reinforcement learning deep learning statistical learning non-parametric learning parametric learning local search Symbolic AI computational learning theory control theory non-convex

  • ptimization

sequential decision theory

ML

big data analysis

Quantum linear algebra Shallow quantum circuits Quantum oracle identification Quantum walks & search Adiabatic QC/ Quantum optimization Quantum COLT

slide-5
SLIDE 5

5

AI

supervised learning unsupervised learning

  • nline learning

generative models reinforcement learning deep learning statistical learning non-parametric learning parametric learning local search Symbolic AI computational learning theory control theory non-convex

  • ptimization

sequential decision theory

ML

big data analysis

Quantum linear algebra

Shallow quantum circuits

Adiabatic QC/ Quantum optimization

Quantum-enhanced ML is even more things

slide-6
SLIDE 6

control and

  • ptimization of

qubits high-energy

QIP

  • Q. Phys

phase diagrams

  • rder

parameters Metrology NISQ optimization, QAOA & VQE Adaptive error correction Experiment synthesis Circuit synthesis Quantum network

  • ptimization

QKD parameter control Efficient decoders Ground state Ansatz Hybrid computation (AI)

6

And then there’s Quantum-applied ML!

slide-7
SLIDE 7

control and

  • ptimization of

qubits high-energy

QIP

  • Q. Phys

phase diagrams

  • rder

parameters Metrology NISQ optimization, QAOA & VQE Adaptive error correction Experiment synthesis Circuit synthesis Quantum network

  • ptimization

QKD parameter control Efficient decoders Ground state Ansatz Hybrid computation (AI)

R e i n f

  • r

c e m e n t l e a r n i n g S u p e r v i s e d l e a r n i n g R e i n f

  • r

c e m e n t l e a r n i n g S u p e r v i s e d l e a r n i n g S u p e r v i s e d l e a r n i n g S u p e r v i s e d l e a r n i n g R e i n f

  • r

c e m e n t l e a r n i n g Unsupervised learning N e u r a l n e t w

  • r

k s

7

slide-8
SLIDE 8

8

What is machine learning

slide-9
SLIDE 9

Learning P(labels|data) given samples from P(data,labels) (also regression)

  • generative models
  • clustering (discriminative)
  • feature extraction

Machine Learning: the WHAT

  • r

Learning structure in P(data) give samples from P(data)

9

?

slide-10
SLIDE 10

10

Beyond data: reinforcement learning

T(s|s0, a) Machine Learning: the WHAT

slide-11
SLIDE 11

Also: MIT technology review breakthrough technology of 2017 [AlphaGo anyone?]

11

slide-12
SLIDE 12

12

slide-13
SLIDE 13

13

Using RL in Real Life

Navigating a city…

https://sites.google.com/view/streetlearn

  • P. Mirowski et. al, Learning to Navigate in Cities Without a Map, arXiv:1804.00168
slide-14
SLIDE 14

14

Machine Learning: the HOW

  • utput hypothesis h on Data x Labels

approximating P(labels|data)

model 
 parameters θ

estimate error

  • n sample

(dataset) Optimizer

In practice

slide-15
SLIDE 15

Support vector machines

separating hyperplane..

15

slide-16
SLIDE 16

Support vector machines

separating hyperplane.. …in higher-dimensional feature space Still (algebraic) optimization over hyperplane and feature function parameters….

16

slide-17
SLIDE 17

17

Machine Learning: the HOW

Learning structure in P(data) give samples from P(data)

slide-18
SLIDE 18

18

  • utput:

hypothesis h on Data x Labels approximating P(labels|data)

  • utput:

hypothesis h on Data “approximating” P(data)

Reinforcement learning

  • utput:

policy π on Actions x States

Machine Learning: the HOW

slide-19
SLIDE 19

Reinforcement learning

(learning behavior, policy, or optimal control)

Supervised learning

(learning how to label datapoints, learning how to approximate a function, how to classify)

Unsupervised learning

(learning a distribution,

  • generate. properties from samples,

feature extraction & dim. reduction)

slide-20
SLIDE 20

That is all ML we need for now What about quantum computers?

20

slide-21
SLIDE 21
  • manipulate registers of

2-level systems (qubits)

  • full description:

n qubits → 2n dimensional vector

  • likely can efficiently compute more things

than classical computers (factoring) e.g. factor numbers, or generate complex distributions

  • even if QC is “shallow”

Banana for scale

cca 50 qubit all-purpose noisy

…and physics …and computer science

…and reality

Quantum computers…

  • manipulation: acting locally (gates)

special-purpose quantum annealers

21

slide-22
SLIDE 22

Quantum computers…

…and physics …and computer science

…and reality

  • can compute things likely beyond BPP (factoring)
  • can produce distributions which are hard-to-simulate

for classical computers (unless PH collapses)

  • even if QC is “shallow”

Banana for scale

special-purpose quantum annealers cca 50 qubit all-purpose noisy

  • manipulate registers of

2-level systems (qubits)

  • full description:

n qubits → 2n dimensional vector

22

slide-23
SLIDE 23

a) The optimization bottleneck b) Big data & comp. complexity c) Machine learning Models

8

Quantum-enhanced supervised learning: the quantum pipeline

23

slide-24
SLIDE 24

a) The optimization bottleneck

— quantum annealers

b) Big data & comp. complexity — universal QC and Q. databases c) Machine learning Models

— restricted (shallow) architectures

24

Quantum-enhanced supervised learning: the quantum pipeline

slide-25
SLIDE 25

a) The optimization bottleneck

— quantum annealers

b) Big data & comp. complexity — universal QC and Q. databases c) Machine learning Models

— restricted (shallow) architectures

25

Quantum-enhanced supervised learning: the quantum pipeline

slide-26
SLIDE 26

The optimization bottleneck

  • Finding ground states of Hamiltonians via adiabatic evolution

  • Very generic optimization problem:

H(s) = sHinitial + (1 − s)Htarget; s(time) argmin|ψihψ|H|ψi

slide-27
SLIDE 27

27

QeML is even more things

AI

supervised learning unsupervised learning

  • nline learning

generative models reinforcement learning deep learning statistical learning non-parametric learning parametric learning local search Symbolic AI computational learning theory control theory non-convex

  • ptimization

sequential decision theory

ML

big data analysis

Quantum linear algebra

Shallow quantum circuits

Quantum walks & search

Adiabatic QC/ Quantum optimization

slide-28
SLIDE 28

a) The optimization bottleneck

— quantum annealers

b) Big data & comp. complexity — universal QC and Q. databases c) Machine learning Models

— restricted (shallow) architectures

28

Quantum-enhanced supervised learning: the quantum pipeline

slide-29
SLIDE 29

Exponential data?

+

Much of data analysis is linear-algebra:

regression = Moore-Penrose PCA = SVD…

Precursors of Quantum Big Data

29

slide-30
SLIDE 30

30

Enter quantum linear algebra

|ψi / PN

i=1 xi|ii

RN 3 x = (xi)i ↓

f(A)|ψi = α0|ψi + α1A|ψi + α0A2|ψi · · · ⇡

U|0i|ψi =  A B C D  ψ

  • =

 Aψ Cψ

  • = |0iA|ψi + |0iC|ψi

· · · ⇡ A−1|ψi amplitude encoding block encoding functions of operators

  • Phys. Rev. Lett. 15,. 103, 250502 (2009)

arXiv:1806.01838

inner products

P(0)ψ = |h0|ψi|2

exp(n) amplitudes in n qubits

interpret QM as linear algebra verbatim

state vector ↔ (data) vector density matrices Hamiltonians unitaries ↔ linear maps projective measurements (swap tests) ↔ inner products prepare states expressible as linear-algebraic manipulations of data-vectors in polylog(N) (when other quantities are well behaved)

slide-31
SLIDE 31

Prediction: 44 zettabytes by 2020. If all data is floats, this is 5.5x1021 float values

If this worked literally…this would make us INFORMATION GODS.

slide-32
SLIDE 32

Prediction: 44 zettabytes by 2020. If all data is floats, this is 5.5x1021 float values

… can be stored in state of 73 qubits (ions, photons….)

If this worked literally…this would make us INFORMATION GODS.

slide-33
SLIDE 33

Clearly there is a catch. Many of them.

slide-34
SLIDE 34

Timeline

2 3 2 8 2 9 2 1 2 2 1 4 2 1 3 2 1 6 2 1 8

Pattern recognition

  • n a QC

QRAM HHL Regression, PCA, SVM

Optimal QLS

Quantum Recommender Systems

QLA, smoothed analysis, De-quantization of low-rank systems

2 1 9 ?

{

Quantum database Linear system solving

Machine learning applications & Improvements

First efficient end-to-end scenario We made it so efficient… that sometimes we don’t need QCs!!

Data-robustness implies

  • q. efficiency
slide-35
SLIDE 35

Summary of quantum (inspired) “big data”

15

interpret QM as linear algebra verbatim manipulate exponentially-sized data-vectors in system (qubit) number HOWEVER need full blown ideal QC need pre-filled database (QRAM) need appropriate condition numbers need robustness to linear error need right preprocessing applied can sometimes be done classically

slide-36
SLIDE 36

Summary of quantum (inspired) “big data”

15

interpret QM as linear algebra verbatim manipulate exponentially-sized data-vectors in system (qubit) number HOWEVER need full blown ideal QC need pre-filled database (QRAM) need appropriate condition numbers need robustness to linear error need right preprocessing applied can sometimes be done classically…

S T I L L A G R E A T I D E A ! !

slide-37
SLIDE 37

37

QeML is even more things

AI

supervised learning unsupervised learning

  • nline learning

generative models reinforcement learning deep learning statistical learning non-parametric learning parametric learning local search Symbolic AI computational learning theory control theory non-convex

  • ptimization

sequential decision theory

ML

big data analysis

Quantum linear algebra

Shallow quantum circuits

Quantum walks & search

Adiabatic QC/ Quantum optimization

slide-38
SLIDE 38

a) The optimization bottleneck

— quantum annealers

b) Big data & comp. complexity — universal QC and Q. databases c) Machine learning Models

— restricted (shallow) architectures

38

Quantum-enhanced supervised learning: the quantum pipeline

slide-39
SLIDE 39

(Quantum) Machine learning Models

Improving ML == speeding up algorithms… or is it?

model 
 parameters θ

estimate error

  • n sample

(dataset) Optimizer

“Machine learning”

39

slide-40
SLIDE 40

Machine learning Models

A lot of machine learning:

  • Take my (training) dataset {(point, label)}
  • Take a model (tensorflow tutorials will suggest), 


e.g. this-that-structure neural network N

  • Train the model (tweak parameters of N, 


until it predicts the training set well) The math behind “cost function”

parametrized family {fθ}

What is this picture missing?

40

slide-41
SLIDE 41

41

Optimization is a part of the method, not the objective

Image: 10.1016/j.compstruct.2018.03.007

best fit v.s. “generalization performance” or classifying well beyond the training set

Data: Models:

Not all models (+training algo) are born equal (for real datasets)…

Challenge:

squeek

  • r

meow?

slide-42
SLIDE 42

Machine learning Models

model 
 parameters θ

estimate error

  • n sample

(dataset) Optimizer

“Machine learning” family of functions. if it’s “good”, we can generalize well

42

slide-43
SLIDE 43

model 
 parameters θ

estimate error

  • n sample

(dataset) Optimizer

How about “shallow quantum circuits”?

  • instead neural network, train a QC!
  • related to ideas from q. condensed-matter physics (VQE)

= = = = =

Quantum Machine learning Models “quantum kernel methods”

  • Phys. Rev. Lett. 122, 040504 2019

Nature 567, 209–212 (2019) (c.f. Elizabeth Behrman in ‘90s)

43

slide-44
SLIDE 44

The quantum feature space

  • relationship between NNs, SVMs and shallow circuits for supervised learning

(embedding - rotation - measurement = feature function - hyperplane - class)

Simple classical kernels A weird quantum kernel

44

slide-45
SLIDE 45

Quantum Machine learning Models “quantum kernel methods”

The good

  • near term architectures
  • seems to be robust 


(noise not inherently critical!)

  • possibly very expressive


The neutral

  • many parameters
  • model advantages less clear

(contrast to variational methods!)
 The bad

  • barren plateaus (also in DNN)

(x1 ∨ x4 ∨ x10) | {z }

(x1 ∨ x4 ∨ x10) | {z } = = = = =

|φ(θin, θclass)i θclass θin(x) {(x, label)i}

estimate error

  • n sample

(dataset) Optimizer

(fiducial)

  • Phys. Rev. Lett. 122, 040504 2019

Nature 567, 209–212 (2019)

CAVEAT: IS IT CLASSICALLY COMPUTATIONALLY HARD?!

45

slide-46
SLIDE 46

A hope… killer app for noisy QCs? ML can be run on small QCs BUT MORE THAN THAT ML good for dealing with noise (in *data*)… Can QML deal with its own noise (in *process*)?

18

46

slide-47
SLIDE 47

47

QeML is even more things

AI

supervised learning unsupervised learning

  • nline learning

generative models reinforcement learning deep learning statistical learning non-parametric learning parametric learning local search Symbolic AI computational learning theory control theory non-convex

  • ptimization

sequential decision theory

ML

big data analysis

Quantum linear algebra

Shallow quantum circuits

Quantum walks & search

Adiabatic QC/ Quantum optimization

slide-48
SLIDE 48

48

Application, match, … conspiracy?

  • Nice analogy Hilbert spaces - big data spaces
  • Hard optimization (needed) - hard optimization (delivered)
  • New learning models (needed) - shallow QC (delivered)
slide-49
SLIDE 49

49

Application, match, … conspiracy?

  • Nice analogy Hilbert spaces - big data spaces


Problem: preparations can offset speed-up; 
 ML: not here! processing must be robust -> low cost

  • Hard optimization (needed) - hard optimization (delivered)


Problem: optimization just heuristic, quality unknown
 ML: well all we do is domain-specific! If it works, it works!

  • New learning models (needed) - shallow QC (delivered)

Problem: noisy models, bad estimates (in VQE) ML: not estimating! Train model, could be even better than exact
 (elements of regularization)

slide-50
SLIDE 50

50

Application, match, … conspiracy?

slide-51
SLIDE 51

51

Application, match, … conspiracy?

Quantum-enhanced reinforcement learning Towards quantum AI Quantum-enhanced unsupervised learning

slide-52
SLIDE 52

52

Application, match, … conspiracy?

still

slide-53
SLIDE 53

53

Application, match, … conspiracy?

slide-54
SLIDE 54

54

Machine learning in the physics domain

slide-55
SLIDE 55

control and

  • ptimization of

qubits

QIP

  • Q. Phys

phase diagrams

  • rder

parameters Metrology NISQ optimization, QAOA & VQE Adaptive error correction Experiment synthesis Circuit synthesis Quantum network

  • ptimization

QKD parameter control Efficient decoders Ground state Ansatz Hybrid computation (AI)

55

Phys

Cosmology

Experimental high-energy Theoretical high-energy

slide-56
SLIDE 56

control and

  • ptimization of

qubits

QIP

  • Q. Phys

phase diagrams

  • rder

parameters Metrology NISQ optimization, QAOA & VQE Adaptive error correction Experiment synthesis Circuit synthesis Quantum network

  • ptimization

QKD parameter control Efficient decoders Ground state Ansatz Hybrid computation (AI)

56

Phys

Cosmology

Experimental high-energy Theoretical high-energy

control and

  • ptimization of

qubits high-energy phase diagrams

  • rder

parameters Metrology NISQ optimization, QAOA & VQE Adaptive error correction Experiment synthesis Circuit synthesis Quantum network

  • ptimization

QKD parameter control Efficient decoders Ground state Ansatz Hybrid computation (AI)

R e i n f

  • r

c e m e n t l e a r n i n g S u p e r v i s e d l e a r n i n g R e i n f

  • r

c e m e n t l e a r n i n g S u p e r v i s e d l e a r n i n g S u p e r v i s e d l e a r n i n g S u p e r v i s e d l e a r n i n g R e i n f

  • r

c e m e n t l e a r n i n g Unsupervised learning & reinforcement N e u r a l n e t w

  • r

k s S u p & u n s u p e r v i s e d

slide-57
SLIDE 57

57

Reinforcement learning

(learning behavior, policy, or optimal control)

Supervised learning

(learning how to label datapoints, learning how to approximate a function, how to classify)

Unsupervised learning

(learning a distribution,

  • generate. properties from samples,

feature extraction & dim. reduction)

slide-58
SLIDE 58

58

Big picture

hard computations new theory & experiments AI/ML assisted computation machine-assisted research 200-petabyte (2017!)

Figure from: https://hackernoon.com/how-big-
 data-is-empowering-ai-and-machine-learning-4e93a1004c8f

slide-59
SLIDE 59

59

Particle physics (and cosmology) Many-body quantum matter Chemistry and materials Facilitating quantum computers

“Machine learning and the physical sciences” Carleo et al., https://arxiv.org/pdf/1903.10563.pdf

slide-60
SLIDE 60

60

Particle physics and cosmology

  • “big data” aspects: event selection, jet tagging, triggering;

(photometric red shift, gravitational lens finding)

  • simulation and inverse problems
  • applications in theory

“Machine learning and the physical sciences” Carleo et al., https://arxiv.org/pdf/1903.10563.pdf

slide-61
SLIDE 61

61

Example: Estimating Cosmological Parameters from the Dark Matter Distribution

(cosm. parameters) − → distr. of matter

ΛCDM What are the cosmological parameters from observed universe?

arXiv:1711.02033v1

“Inverse simulation?”

slide-62
SLIDE 62

62

arXiv:1711.02033v1

Machine learning solution:

Train NN to output correct parameters 
 given the universe; Training set: (universe, parameters) Learning goal: (parameters | universe)

Example: Estimating Cosmological Parameters from the Dark Matter Distribution

slide-63
SLIDE 63

63

Many-body quantum matter

“Machine learning and the physical sciences” Carleo et al., https://arxiv.org/pdf/1903.10563.pdf

  • neural quantum states (approximate the wavefunction)
  • expressivity, learning from data, variational approaches
  • assisted many-body simulations
  • learned hard sampling
  • classification of many-body phases of matter
slide-64
SLIDE 64

64

Machine learning in quantum information processing

slide-65
SLIDE 65

65

Enabling quantum information processing devices

slide-66
SLIDE 66

66

Application, match, … conspiracy?

slide-67
SLIDE 67

Editor-in-Chief Giovanni Acampora, University of Naples Federico II, Italy Field Editors 1) Quantum Machine Learning Seth Lloyd (MIT), USA 2) Quantum Computing for Artificial Intelligence Hans Jürgen Briegel, (Innsbruck, Austria) 3) Artificial Intelligence for Quantum Information Processing Chin-Teng Lin (Sydney, Australia) 4) Quantum- and Bio-inspired Computational Intelligence Francisco Herrera (Granada, Spain) 5) Quantum Optimization Davide Venturelli (USRA, USA) CALL FOR PAPERS

67