Learning Noise in Quantum Information Processors Travis L Scholten - - PowerPoint PPT Presentation

learning noise in quantum information processors
SMART_READER_LITE
LIVE PREVIEW

Learning Noise in Quantum Information Processors Travis L Scholten - - PowerPoint PPT Presentation

Learning Noise in Quantum Information Processors Travis L Scholten @Travis_Sch Center for Quantum Information and Control, Center for Computing Research, Sandia National Laboratories, Albuquerque, USA University of New Mexico, Albuquerque, USA


slide-1
SLIDE 1

Sandia National Laboratories is a multimission laboratory managed and operated by National Technology and Engineering Solutions of Sandia LLC, a wholly owned subsidiary of Honeywell International Inc. for the U.S. Department of Energy’s National Nuclear Security Administration under contract DE-NA0003525.

CCR

Center for Computing Research

Learning Noise in Quantum Information Processors

Travis L Scholten @Travis_Sch

Center for Quantum Information and Control, University of New Mexico, Albuquerque, USA

Center for Computing Research, Sandia National Laboratories, Albuquerque, USA

QTML 2017

slide-2
SLIDE 2

quantum machine learning annealing

quantum annealing quantum gibbs sampling quantum topological algorithms

quantum rejection sampling / HHL

Quantum ODE solvers

control and metrology

reinforcement learning tomography quantum control phase estimation hamiltonian learning quantum perceptron quantum BM simulated annealing markov chain monte-carlo

neural nets

feed forward neural net quantum PCA quantum SVM quantum NN classification quantum clustering quantum data fitting

machine learning quantum information processing

Biamonte, et. al, arXiv: 1611.09347

There are lots of applications at the intersection

  • f QI/QC and ML…

26

slide-3
SLIDE 3

quantum machine learning annealing

quantum annealing quantum gibbs sampling quantum topological algorithms

quantum rejection sampling / HHL

Quantum ODE solvers

control and metrology

reinforcement learning tomography quantum control phase estimation hamiltonian learning quantum perceptron quantum BM simulated annealing markov chain monte-carlo

neural nets

feed forward neural net quantum PCA quantum SVM quantum NN classification quantum clustering quantum data fitting

machine learning quantum information processing

Biamonte, et. al, arXiv: 1611.09347

…I want to focus on how ML can improve characterization of quantum hardware. 25

slide-4
SLIDE 4

Quantum device characterization (QCVV) techniques arranged by amount learned and time required. Speed of learning

Fast Full Slow

Amount learned

Limited

24

slide-5
SLIDE 5

Tomography is very informative, but time-consuming!

Gate set tomography State tomography

Speed of learning

Fast Slow

Amount learned

Process tomography

Full Limited

23

slide-6
SLIDE 6

Randomized benchmarking is fast, but yields limited information.

Randomized Benchmarking Gate set tomography State tomography

Speed of learning

Fast Slow

Amount learned

Process tomography

(Several variants, leads to different kinds of information learned.)

Full Limited

22

slide-7
SLIDE 7

Depending on how much we want to learn, and how quickly, machine learning could be useful.

Machine Learning Randomized Benchmarking Gate set tomography State tomography

Speed of learning

Fast Slow

Amount learned

Process tomography

Full Limited

Caveat: “Speed” doesn’t include *training time*

21

slide-8
SLIDE 8

Depending on how much we want to learn, and how quickly, machine learning could be useful.

Machine Learning Randomized Benchmarking Gate set tomography State tomography

Speed of learning

Fast Slow

Amount learned

Process tomography

Full Limited

Can machine learning extract information about noise affecting near and medium-term quantum hardware?

20

slide-9
SLIDE 9

Noise affects the

  • utcome probabilities of

quantum circuits. How can we learn about noise using the data we get from running quantum circuits?

19

slide-10
SLIDE 10

|0i Yπ/2 Noise in quantum hardware affects the

  • utcome probabilities of circuits.

(Noise affects outcome probability)

18

Example: over-rotation error of a single-qubit gate (The circuit we write down)

Pr(0) = Tr(|0ih0|E(|0ih0|)) = 1

2(1 sin θ)

slide-11
SLIDE 11

Gate set tomography (GST) provides a set of structured circuits we can use for learning.

GST assumes the device is a black box, described by a gate set. GST prescribes certain circuits to run that collectively amplify all types of noise. Standard use: Outcome probabilities are analyzed by pyGSTi software to estimate the noisy gates

M ρ G1 G2

. . .

Blume-Kohout, et. al, arXiv 1605.07674

|0i Yπ/2

|0i Yπ/2 Yπ/2

|0i Xπ/2 Yπ/2 Zπ/2

17

slide-12
SLIDE 12

|0i Yπ/2 Yπ/2 Yπ/2 Yπ/2

Gate set tomography (GST) provides a set of structured circuits we can use for learning.

GST prescribes certain circuits to run that collectively amplify all types of noise.

|0i Yπ/2

|0i Yπ/2 Yπ/2 Circuits have varying length, up to some maximum length L.

l = 1, 2, 4, · · · , L

l = 1

l = 2

Why? Longer circuits are more sensitive to noise.

16

l = 4

l = 1

l = 2

slide-13
SLIDE 13

To do machine learning on GST data sets, embed them in a feature space.

## Columns = minus count, plus count {} 100 0 Gx 44 56 Gy 45 55 GxGx 9 91 GxGxGx 68 32 GyGyGy 70 30

(GST data set)

f = (f1, f2, · · · ) ∈ Rd

The dimension of the feature space grows with L because more circuits are added.

        .56 .55 .91 .32 .3        

15

slide-14
SLIDE 14

Noise changes some components of the feature vectors.

How can we identify the “signature” of a noise process using GST feature vectors?

14

slide-15
SLIDE 15

Principal component analysis (PCA) is a useful tool for understanding the structure of GST feature vectors.

PCA finds a low-dimensional representation of data by looking for directions of maximum variance. Compute covariance matrix & diagonalize Defines a map:

f → PK

j=1(f · σj)σj

C = PK

j=1 σjσjσT j

σ1 ≥ σ2 · · · ≥ σK

(d = K)

13

slide-16
SLIDE 16

Projection onto a 2-dimensional PCA subspace reveals a structure to GST feature vectors.

Different noise types and noise strengths tend to cluster!

(PCA performed on entire dataset, then individual feature vectors transformed.) 4.5% depolarizing 1% coherent 1% depolarizing 4.5% coherent

12

slide-17
SLIDE 17

Longer GST circuits amplify noise, making the clusters more distinguishable.

Adding longer circuits makes the clusters more distinguishable.

We can use this structure to do classification!

(An independent PCA was done for each L.)

11

slide-18
SLIDE 18

Classification is possible because the data sets cluster based on noise type and strength!

Project feature vectors based on PCA

10

Label feature vectors based on noise Train a soft- margin, linear support vector machine (SVM)

slide-19
SLIDE 19

Classification is possible because the data sets cluster based on noise type and strength!

Project feature vectors based on PCA

9

Label feature vectors based on noise Train a soft- margin, linear support vector machine (SVM)

96% accuracy?? Cross-validation required!

slide-20
SLIDE 20

Under cross-validation, the SVM has reasonably low inaccuracy. 8

SVM is fairly accurate - largest inaccuracy ~2% 20-fold shuffle-split cross-validation (25% withheld for testing)

slide-21
SLIDE 21

7

20-fold shuffle-split cross-validation scheme used, with 25% of the data withheld for testing on each split. A “one-versus-one” multi-class classification scheme was used.

The accuracy of the SVM is affected by the number of components and maximum sequence length.

slide-22
SLIDE 22

Can a classifier learn the difference between arbitrary stochastic and arbitrary coherent noise? 6

Coherent Noise Ideal Stochastic Noise ˙ ρ = −i[H0, ρ]

˙ ρ = −i[H0, ρ] + AρA† − 1

2{A†A, ρ}

˙ ρ = −i[H0, ρ] − i[e, ρ]

E = Λ G0 E = V G0 E = G0

V V T = I ΛΛT 6= I

slide-23
SLIDE 23

Classification in a 2-dimensional subspace is harder, due to structure of PCA-projected feature vectors. 5

“Radio dish” type structure Linear classifier infeasible with

  • nly 2 PCA

components

slide-24
SLIDE 24

Preliminary results indicate a linear, soft-margin SVM can classify these two noise types in higher dimensions. 4

20-fold shuffle-split cross-validation scheme used, with 25% of the data withheld for testing on each split. A “one-verus-one” multi-class classification scheme was used.

For each L:

  • 10 values of noise strength in [10**-4, 10**-1]
  • 260 random instances

Gap goes away if noise <= 10**-2 removed from data

slide-25
SLIDE 25

Machine Learning

Specific machine learning tools can analyze GST circuits and learn about noise.

Randomized Benchmarking Gate set tomography State tomography

Speed of learning

Fast Slow

Amount learned

Process tomography

Full Limited

SVMs & PCA + GST Circuits

3

slide-26
SLIDE 26

Machine Learning Randomized Benchmarking Gate set tomography State tomography

Speed of learning

Fast Slow

Amount learned

Process tomography

Full Limited What else can we learn?? What circuits do we need??

SVMs & PCA + GST Circuits

2 Specific machine learning tools can analyze GST circuits and learn about noise.

slide-27
SLIDE 27

quantum machine learning annealing

quantum annealing quantum gibbs sampling quantum topological algorithms

quantum rejection sampling / HHL

Quantum ODE solvers

control and metrology

reinforcement learning tomography quantum control phase estimation hamiltonian learning quantum perceptron quantum BM simulated annealing markov chain monte-carlo

neural nets

feed forward neural net quantum PCA quantum SVM quantum NN classification quantum clustering quantum data fitting

machine learning quantum information processing

Biamonte, et. al, arXiv: 1611.09347

1

There are lots of problems at the intersection of device characterization and machine learning!

slide-28
SLIDE 28

Thank you!

@Travis_Sch