quantum algorithms for systems of linear equations
play

Quantum Algorithms for Systems of Linear Equations Rolando Somma - PowerPoint PPT Presentation

Quantum Algorithms for Systems of Linear Equations Rolando Somma Theoretical Division Los Alamos National Laboratory Joint work with Andrew Childs Robin Kothari Yigit Subasi Davide Orsucci Maryland Microsoft Vienna Los Alamos Workshop


  1. The HHL Algorithm for the QLSP [5] [HHL08] There exists a quantum algorithm that solves the QLSP with complexity ˜ O [  ( T b + C A (  / ✏ , ✏ /  ))] Considering that many Hamiltonians can be simulated efficiently on quantum computers, the complexity dependence on the dimension is small (e.g., logarithmic) [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  2. Improvements of the HHL Algorithm [6] [HHL08] There exists a quantum algorithm that solves the QLSP with complexity ˜ O [  ( T b + C A (  / ✏ , ✏ /  ))] Considering that many Hamiltonians can be simulated efficiently on quantum computers, the complexity dependence on the dimension is small (e.g., logarithmic) Further improvements by Ambainis (Variable Time Amplitude Amplification or VTAA): [6] There exists a quantum algorithm that solves the QLSP with complexity ˜  T b + C A (  / ✏ 3 , ✏ ) ⇥ ⇤ O [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009) [6] A. Ambainis, STACS 14, 636 (2012)

  3. Improvements of the HHL Algorithm [6] [HHL08] There exists a quantum algorithm that solves the QLSP with complexity ˜ O [  ( T b + C A (  / ✏ , ✏ /  ))] Considering that many Hamiltonians can be simulated efficiently on quantum computers, the complexity dependence on the dimension is small (e.g., logarithmic) Further improvements by Ambainis (Variable Time Amplitude Amplification or VTAA): [6] There exists a quantum algorithm that solves the QLSP with complexity ˜ Almost linear in 𝜆 !  T b + C A (  / ✏ 3 , ✏ ) ⇥ ⇤ O Note that the best Hamiltonian simulation methods have query and gate • complexities almost linear in evolution time and logarithmic in precision [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009) [6] A. Ambainis, STACS 14, 636 (2012)

  4. A quick view of the HHL algorithm and VTAA [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  5. A quick view of the HHL algorithm and VTAA [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  6. A quick view of the HHL algorithm and VTAA N − 1 X c j | v j i I | ˜ | b i ! λ j i E j =0 [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009) [6] A. Ambainis, STACS 14, 636 (2012)

  7. A quick view of the HHL algorithm and VTAA N − 1 X c j | v j i I | ˜ | b i ! λ j i E j =0 This register contains the eigenvalue estimate (superposition): It suffices to have the estimate with relative precision 𝜗 • Order log( 𝜆 / 𝜗 ) ancillary qubits • [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  8. A quick view of the HHL algorithm and VTAA N − 1 X c j | v j i I | ˜ | b i ! λ j i E j =0 ! s 1 1 Then we implement the conditional rotation: | ˜ λ j i E ! | ˜ λ j i E | 0 i O + 1 � | 1 i O κ ˜ κ 2 ˜ λ 2 λ j j [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  9. A quick view of the HHL algorithm and VTAA N − 1 X c j | v j i I | ˜ | b i ! λ j i E j =0 ! s 1 1 Then we implement the conditional rotation: | ˜ λ j i E ! | ˜ λ j i E | 0 i O + 1 � | 1 i O κ ˜ κ 2 ˜ λ 2 λ j j Undo phase estimation [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  10. A quick view of the HHL algorithm and VTAA N − 1 X c j | v j i I | ˜ | b i ! λ j i E j =0 ! s 1 1 Then we implement the conditional rotation: | ˜ λ j i E ! | ˜ λ j i E | 0 i O + 1 � | 1 i O κ ˜ κ 2 ˜ λ 2 λ j j Undo phase estimation Amplitude amplification for amplifying the amplitude of the | 0 i O state [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  11. A quick view of the HHL algorithm and VTAA Roughly, the scaling of the HHL algorithm can be analyzed from the worst case: p | b i = (1 / κ ) | v 1 / κ i + 1 � 1 / κ 2 | v 1 i [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  12. A quick view of the HHL algorithm and VTAA Roughly, the scaling of the HHL algorithm can be analyzed from the worst case: p | b i = (1 / κ ) | v 1 / κ i + 1 � 1 / κ 2 | v 1 i The action of 1/ A will roughly create the equal superposition state, so both are equally important [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  13. A quick view of the HHL algorithm and VTAA Roughly, the scaling of the HHL algorithm can be analyzed from the worst case: p | b i = (1 / κ ) | v 1 / κ i + 1 � 1 / κ 2 | v 1 i The action of 1/ A will roughly create the equal superposition state, so both are equally important For the desired precision we need to evolve with A for time of order 𝜆 / 𝜗 [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  14. A quick view of the HHL algorithm and VTAA Roughly, the scaling of the HHL algorithm can be analyzed from the worst case: p | b i = (1 / κ ) | v 1 / κ i + 1 � 1 / κ 2 | v 1 i The action of 1/ A will roughly create the equal superposition state, so both are equally important For the desired precision we need to evolve with A for time of order 𝜆 / 𝜗 The action of 1/ ( 𝜆 A ) on the state reduces its amplitude by order 1/ 𝜆 and order 𝜆 amplitude amplification rounds are needed [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  15. A quick view of the HHL algorithm and VTAA Roughly, the scaling of the HHL algorithm can be analyzed from the worst case: p | b i = (1 / κ ) | v 1 / κ i + 1 � 1 / κ 2 | v 1 i The action of 1/ A will roughly create the equal superposition state, so both are equally important For the desired precision we need to evolve with A for time of order 𝜆 / 𝜗 The action of 1/ ( 𝜆 A ) on the state reduces its amplitude by order 1/ 𝜆 and order 𝜆 amplitude amplification rounds are needed From here we see that we need to evolve with A for time that is, at least, order 𝜆 2 / 𝜗 [5] Harrow, Hassidim, Lloyd, PRL 103, 150502 (2009)

  16. A quick view of the HHL algorithm and VTAA How can we improve this time complexity to something that is almost linear in the condition number? [6] A. Ambainis, STACS 14, 636 (2012)

  17. A quick view of the HHL algorithm and VTAA How can we improve this time complexity to something that is almost linear in the condition number? One answer is via Variable Time Amplitude Amplification (VTAA) [6] [6] A. Ambainis, STACS 14, 636 (2012)

  18. A quick view of the HHL algorithm and VTAA How can we improve this time complexity to something that is almost linear in the condition number? One answer is via Variable Time Amplitude Amplification (VTAA) [6] The rough idea is as follows (again, considering the worst case): p | b i = (1 / κ ) | v 1 / κ i + 1 � 1 / κ 2 | v 1 i First we do a bad-precision phase estimation to distinguish large from small • eigenvalues. This may be done evolving with A for time independent of 𝜆 Then we implement a rough approximation of 1/ 𝜆 A to eigenstates of large • eigenvalue We need order 𝜆 amplitude amplification steps • We implement an accurate approximation of 1/ 𝜆 A to eigenstates of small • eigenvalue Amplitude amplification for order 1 steps • Undo phase estimation or apply the Fourier transform • [6] A. Ambainis, STACS 14, 636 (2012)

  19. A quick view of the HHL algorithm and VTAA How can we improve this time complexity to something that is almost linear in the condition number? One answer is via Variable Time Amplitude Amplification (VTAA) [6] The rough idea is as follows (again, considering the worst case): p | b i = (1 / κ ) | v 1 / κ i + 1 � 1 / κ 2 | v 1 i First we do a bad-precision phase estimation to distinguish large from small • eigenvalues. This may be done evolving with A for time independent of 𝜆 Then we implement a rough approximation of 1/ 𝜆 A to eigenstates of large • eigenvalue We need order 𝜆 amplitude amplification steps • We implement an accurate approximation of 1/ 𝜆 A to eigenstates of small • eigenvalue Amplitude amplification for order 1 steps • Undo phase estimation or apply the Fourier transform • The complexity of VTAA in terms of precision is worse than that of HHL [6] A. Ambainis, STACS 14, 636 (2012)

  20. This talk: two quantum algorithms for the QSLP • I will present two quantum algorithms for the QLSP that improve previous results in different ways: [7] There exists a quantum algorithm that solves the QLSP with complexity ˜ O [  ( T b + C A (  log(  / ✏ , ✏ /  ))] [7] A. Childs, R. Kothari, RDS , SIAM J. Comp. 46 , 1920 (2017).

  21. This talk: two quantum algorithms for the QSLP • I will present two quantum algorithms for the QLSP that improve previous results in different ways: [7] There exists a quantum algorithm that solves the QLSP with complexity ˜ O [  ( T b + C A (  log(  / ✏ , ✏ /  ))] This results in an exponential improvement on the precision parameter • [7] A. Childs, R. Kothari, RDS , SIAM J. Comp. 46 , 1920 (2017).

  22. This talk: two quantum algorithms for the QSLP • I will present two quantum algorithms for the QLSP that improve previous results in different ways: [7] There exists a quantum algorithm that solves the QLSP with complexity ˜ O [  ( T b + C A (  log(  / ✏ , ✏ /  ))] This results in an exponential improvement on the precision parameter • It can be improved using a version of VTAA to: • [7] There exists a quantum algorithm that solves the QLSP with complexity ˜ O [  T b + C A (  log(  / ✏ , ✏ ))] [7] A. Childs, R. Kothari, RDS , SIAM J. Comp. 46 , 1920 (2017).

  23. This talk: two quantum algorithms for the QSLP Why these improvements are important? • The previous result allowed us to prove a polynomial quantum speedup for hitting time estimation in terms of the spectral gap of a Markov chain and precision (A. Chowdhury, R.D. Somma, QIC 17, 0041 (2017)). • Having a small complexity dependence on precision is important for, e.g., computing expectation values of observables at the quantum metrology limit.

  24. This talk: two quantum algorithms for the QSLP • I will present two quantum algorithms for the QLSP that improve previous results in different ways: [8] There exists a quantum algorithm that solves the QLSP by evolving with Hamiltonians that are linear combinations of (products of) A , the projector in the ˜ initial state, and Pauli matrices. The overall evolution time is O (  / ✏ ) [8] Y. Subasi, RDS , D. Orsucci, arXiv:1805.10549 (2018).

  25. This talk: two quantum algorithms for the QSLP • I will present two quantum algorithms for the QLSP that improve previous results in different ways: [8] There exists a quantum algorithm that solves the QLSP by evolving with Hamiltonians that are linear combinations of (products of) A , the projector in the ˜ initial state, and Pauli matrices. The overall evolution time is O (  / ✏ ) Using Hamiltonian simulation, this transfers to complexity ˜ O (  T b / ✏ + C A (  / ✏ , ✏ )) [8] Y. Subasi, RDS , D. Orsucci, arXiv:1805.10549 (2018).

  26. This talk: two quantum algorithms for the QSLP • I will present two quantum algorithms for the QLSP that improve previous results in different ways: [8] There exists a quantum algorithm that solves the QLSP by evolving with Hamiltonians that are linear combinations of (products of) A , the projector in the ˜ initial state, and Pauli matrices. The overall evolution time is O (  / ✏ ) Using Hamiltonian simulation, this transfers to complexity ˜ O (  T b / ✏ + C A (  / ✏ , ✏ )) The method is very different and based on adiabatic evolutions. It does not • require of complicated subroutines such as phase estimation and variable time amplitude amplification, therefore reducing the number of ancillary qubits substantially. [8] Y. Subasi, RDS , D. Orsucci, arXiv:1805.10549 (2018).

  27. This talk: two quantum algorithms for the QSLP Why this improvement is important? • Phase estimation and VTAA require several ancillary qubits (beyond those needed for Hamiltonian simulation) • Within two weeks of posting our result, a group implemented our algorithm in NMR, claiming that it is the largest simulated instance so far (8x8) [9] [9] J. Wen, et.al., arXiv:1806.0329 (2018)

  28. First algorithm: A Fourier approach for solving the QSLP

  29. First algorithm: A Fourier approach for solving the QSLP

  30. First algorithm: A Fourier approach for solving the QSLP 1/ A is not unitary and we need to find a unitary implementation for it. We then • go through a sequence of approximations:

  31. First algorithm: A Fourier approach for solving the QSLP 1/ A is not unitary and we need to find a unitary implementation for it. We then • go through a sequence of approximations: we are getting closer: Linear combination of unitaries

  32. First algorithm: A Fourier approach for solving the QSLP

  33. First algorithm: A Fourier approach for solving the QSLP 20 10 � 1.0 � 0.5 0.5 1.0 � 10 � 20

  34. First algorithm: A Fourier approach for solving the QSLP 20 10 � 1.0 � 0.5 0.5 1.0 � 10 � 20

  35. First algorithm: A Fourier approach for solving the QSLP 20 10 � 1.0 � 0.5 0.5 1.0 � 10 � 20 The maximum “evolution time” under A in the approximation of 1/ A is

  36. First algorithm: A Fourier approach for solving the QSLP • So far we approximated 1/ A , within the desired accuracy, by a finite linear combination of unitaries. Each unitary corresponds to evolving with A for certain time, and the max evolution time is almost linear in the condition number

  37. First algorithm: A Fourier approach for solving the QSLP • So far we approximated 1/ A , within the desired accuracy, by a finite linear combination of unitaries. Each unitary corresponds to evolving with A for certain time, and the max evolution time is almost linear in the condition number QLSP Hamiltonian simulation

  38. Implementing a linear combination of unitaries Suppose we want to implement the operator λ 1 U 1 + λ 2 U 2 to some state | ψ i where λ i ≥ 0, λ 1 + λ 2 = 1, and U i unitary

  39. Implementing a linear combination of unitaries Suppose we want to implement the operator λ 1 U 1 + λ 2 U 2 to some state | ψ i where λ i ≥ 0, λ 1 + λ 2 = 1, and U i unitary p p λ 1 | 0 i + λ 2 | 1 i | 0 i V † V p ( λ 1 U 1 + λ 2 U 2 ) | ψ i | 0 i + λ 1 λ 2 ( U 1 � U 2 ) | ψ i | 1 i U i | ψ i

  40. Implementing a linear combination of unitaries Suppose we want to implement the operator λ 1 U 1 + λ 2 U 2 to some state | ψ i where λ i ≥ 0, λ 1 + λ 2 = 1, and U i unitary p p λ 1 | 0 i + λ 2 | 1 i | 0 i V † V p ( λ 1 U 1 + λ 2 U 2 ) | ψ i | 0 i + λ 1 λ 2 ( U 1 � U 2 ) | ψ i | 1 i U i | ψ i Incorrect part Correct part (orthogonal)

  41. Implementing a linear combination of unitaries Suppose we want to implement the operator λ 1 U 1 + λ 2 U 2 to some state | ψ i where λ i ≥ 0, λ 1 + λ 2 = 1, and U i unitary p p λ 1 | 0 i + λ 2 | 1 i | 0 i V † V p ( λ 1 U 1 + λ 2 U 2 ) | ψ i | 0 i + λ 1 λ 2 ( U 1 � U 2 ) | ψ i | 1 i U i | ψ i Incorrect part Correct part (orthogonal) M − 1 X This idea can be generalized to the case where the goal is to implement λ i U i i =0

  42. Implementing a linear combination of unitaries Suppose we want to implement the operator λ 1 U 1 + λ 2 U 2 to some state | ψ i where λ i ≥ 0, λ 1 + λ 2 = 1, and U i unitary p p λ 1 | 0 i + λ 2 | 1 i | 0 i V † V p ( λ 1 U 1 + λ 2 U 2 ) | ψ i | 0 i + λ 1 λ 2 ( U 1 � U 2 ) | ψ i | 1 i U i | ψ i Incorrect part Correct part (orthogonal) M − 1 X This idea can be generalized to the case where the goal is to implement λ i U i i =0 M − 1 1 X λ i U i ) | ψ i | 0 . . . 0 i + | ξ ⊥ i λ ( i =0

  43. Implementing a linear combination of unitaries Suppose we want to implement the operator λ 1 U 1 + λ 2 U 2 to some state | ψ i where λ i ≥ 0, λ 1 + λ 2 = 1, and U i unitary p p λ 1 | 0 i + λ 2 | 1 i | 0 i V † V p ( λ 1 U 1 + λ 2 U 2 ) | ψ i | 0 i + λ 1 λ 2 ( U 1 � U 2 ) | ψ i | 1 i U i | ψ i Incorrect part Correct part (orthogonal) M − 1 X This idea can be generalized to the case where the goal is to implement λ i U i i =0 M − 1 1 Amplitude amplification to X λ i U i ) | ψ i | 0 . . . 0 i + | ξ ⊥ i λ ( obtain the correct part i =0

  44. First algorithm: A Fourier approach for solving the QSLP We use the LCU approach to implement the Fourier approximation of 1/ A •

  45. First algorithm: A Fourier approach for solving the QSLP We use the LCU approach to implement the Fourier approximation of 1/ A • Note: We assume that the gate complexity of the operation V is small with • respect to other complexities Adding up all the coefficients in the linear combination of unitaries, we obtain • λ = ˜ O ( κ )

  46. First algorithm: A Fourier approach for solving the QSLP We use the LCU approach to implement the Fourier approximation of 1/ A • Note: We assume that the gate complexity of the operation V is small with • respect to other complexities Adding up all the coefficients in the linear combination of unitaries, we obtain • λ = ˜ O ( κ ) This is also the number of amplitude amplification rounds •

  47. First algorithm: A Fourier approach for solving the QSLP We use the LCU approach to implement the Fourier approximation of 1/A • Note: We assume that the gate complexity of the operation V is small with • respect to other complexities Adding up all the coefficients in the linear combination of unitaries, we obtain • λ = ˜ O ( κ ) This is also the number of amplitude amplification rounds • Then, the overall complexity of this approach is • ˜ O [  ( T b + C A (  log(  / ✏ , ✏ /  ))]

  48. First algorithm: A Fourier approach for solving the QSLP We use the LCU approach to implement the Fourier approximation of 1/A • Note: We assume that the gate complexity of the operation V is small with • respect to other complexities Adding up all the coefficients in the linear combination of unitaries, we obtain • λ = ˜ O ( κ ) This is also the number of amplitude amplification rounds • Then, the overall complexity of this approach is • ˜ O [  ( T b + C A (  log(  / ✏ , ✏ /  ))] This is almost quadratic in the condition number. To improve it to almost linear we use a version of VTAA that doesn’t ruin the logarithmic scaling in precision

  49. First algorithm: A Fourier approach for solving the QSLP VTAA for HHL relies heavily on phase estimation, bringing a prohibitive • complexity dependence on precision But in our case we only need to distinguish the regions for the eigenvalues with • high confidence, so the scaling in precision is logarithmic

  50. First algorithm: A Fourier approach for solving the QSLP VTAA for HHL relies heavily on phase estimation, bringing a prohibitive • complexity dependence on precision But in our case we only need to distinguish the regions for the eigenvalues with • high confidence, so the scaling in precision is logarithmic The final algorithm is VTAA applied to another algorithm that is built upon a • sequence of steps. At each step we do the following: i) We determine the region of the eigenvalue • with high confidence. ii) We apply 1/A within the necessary precision for that region (replacing the condition number)

  51. First algorithm: A Fourier approach for solving the QSLP VTAA for HHL relies heavily on phase estimation, bringing a prohibitive • complexity dependence on precision But in our case we only need to distinguish the regions for the eigenvalues with • high confidence, so the scaling in precision is logarithmic The final algorithm is VTAA applied to another algorithm that is built upon a • sequence of steps. At each step we do the following: i) We determine the region of the eigenvalue • with high confidence. ii) We apply 1/A within the necessary precision for that region (replacing the condition number) The overall complexity of this approach is • ˜ O [  T b + C A (  log(  / ✏ , ✏ ))]

  52. First algorithm: A Fourier approach for solving the QSLP VTAA for HHL relies heavily on phase estimation, bringing a prohibitive • complexity dependence on precision But in our case we only need to distinguish the regions for the eigenvalues with • high confidence, so the scaling in precision is logarithmic The final algorithm is VTAA applied to another algorithm that is built upon a • sequence of steps. At each step we do the following: i) We determine the region of the eigenvalue • with high confidence. ii) We apply 1/A within the necessary precision for that region (replacing the condition number) The overall complexity of this approach is • ˜ O [  T b + C A (  log(  / ✏ , ✏ ))] Using the best Hamiltonian simulation methods, this is almost linear in the condition number and polylog in inverse of precision

  53. Second algorithm: An “adiabatic” approach for the QSLP

  54. Second algorithm: An “adiabatic” approach for the QSLP The idea here is to prepare the eigenstate of a Hamiltonian by preparing a • sequence continuously related eigenstates of a family of Hamiltonians

  55. Second algorithm: An “adiabatic” approach for the QSLP The idea here is to prepare the eigenstate of a Hamiltonian by preparing a • sequence continuously related eigenstates of a family of Hamiltonians We want the eigenstate to be the desired quantum state (after tracing out • ancillary systems)

  56. Second algorithm: An “adiabatic” approach for the QSLP The idea here is to prepare the eigenstate of a Hamiltonian by preparing a • sequence continuously related eigenstates of a family of Hamiltonians We want the eigenstate to be the desired quantum state (after tracing out • ancillary systems) b ~ P ⊥ x = P ⊥ b A. ~ b = 0 B † B | x i = 0 , B = P ⊥ b .A

  57. Second algorithm: An “adiabatic” approach for the QSLP The idea here is to prepare the eigenstate of a Hamiltonian by preparing a • sequence continuously related eigenstates of a family of Hamiltonians We want the eigenstate to be the desired quantum state (after tracing out • ancillary systems) b ~ P ⊥ x = P ⊥ b A. ~ b = 0 H B † B | x i = 0 , B = P ⊥ b .A The following properties can be proven: The desired state is the unique ground state of H • The eigenvalue gap is order 1/ 𝜆 2 •

  58. Second algorithm: An “adiabatic” approach for the QSLP The idea here is to prepare the eigenstate of a Hamiltonian by preparing a • sequence continuously related eigenstates of a family of Hamiltonians We want the eigenstate to be the desired quantum state (after tracing out • ancillary systems) b ~ P ⊥ x = P ⊥ b A. ~ b = 0 H B † B | x i = 0 , B = P ⊥ b .A The following properties can be proven: The desired state is the unique ground state of H • The eigenvalue gap is order 1/ 𝜆 2 • We now seek the family of interpolating Hamiltonians •

  59. Second algorithm: An “adiabatic” approach for the QSLP We assume for the moment that A >1/ 𝜆 •

  60. Second algorithm: An “adiabatic” approach for the QSLP We assume for the moment that A >1/ 𝜆 • We define the interpolating matrix A ( s ) = (1 − s ) I + sA , 0 ≤ s ≤ 1 •

  61. Second algorithm: An “adiabatic” approach for the QSLP We assume for the moment that A >1/ 𝜆 • We define the interpolating matrix • A ( s ) = (1 − s ) I + sA , 0 ≤ s ≤ 1 Similarly, we define • H ( s ) = B † ( s ) B ( s ) , B ( s ) = P ⊥ b A ( s )

  62. Second algorithm: An “adiabatic” approach for the QSLP We assume for the moment that A >1/ 𝜆 • We define the interpolating matrix • A ( s ) = (1 − s ) I + sA , 0 ≤ s ≤ 1 Similarly, we define • H ( s ) = B † ( s ) B ( s ) , B ( s ) = P ⊥ b A ( s ) This is like solving an increasingly difficult system of linear equations! •

  63. Second algorithm: An “adiabatic” approach for the QSLP We assume for the moment that A >1/ 𝜆 • We define the interpolating matrix • A ( s ) = (1 − s ) I + sA , 0 ≤ s ≤ 1 Similarly, we define • H ( s ) = B † ( s ) B ( s ) , B ( s ) = P ⊥ b A ( s ) This is like solving an increasingly difficult system of linear equations! • | x i s = 1 H (1) = A.P ⊥ b .A | b i s = 0 H (0) = P ⊥ b

  64. Second algorithm: An “adiabatic” approach for the QSLP We assume for the moment that A >1/ 𝜆 • We define the interpolating matrix • A ( s ) = (1 − s ) I + sA , 0 ≤ s ≤ 1 Similarly, we define • H ( s ) = B † ( s ) B ( s ) , B ( s ) = P ⊥ b A ( s ) This is like solving an increasingly difficult system of linear equations! • | x i L s = 1 H (1) = A.P ⊥ b .A | b i s = 0 H (0) = P ⊥ b The minimum eigenvalue gap is order 1/ 𝜆 2 and the length of the path L is log( 𝜆 ) •

  65. The randomization method to prepare eigenstates [] s q | x i s 2 H (1) = A.P ⊥ b .A s 1 | b i s = 0 H (0) = P ⊥ b By performing a sequence of projective measurements at sufficiently close • points, we can prepare the related eigenstates with high probability

  66. The randomization method to prepare eigenstates [] s q | x i s 2 H (1) = A.P ⊥ b .A s 1 | b i s = 0 H (0) = P ⊥ b By performing a sequence of projective measurements at sufficiently close • points, we can prepare the related eigenstates with high probability Each measurement can be simulated by evolving with the corresponding • Hamiltonian for random time. This reduces coherences between eigenstates

  67. The randomization method to prepare eigenstates [] s q | x i s 2 H (1) = A.P ⊥ b .A s 1 | b i s = 0 H (0) = P ⊥ b By performing a sequence of projective measurements at sufficiently close • points, we can prepare the related eigenstates with high probability Each measurement can be simulated by evolving with the corresponding • Hamiltonian for random time. This reduces coherences between eigenstates The expected evolution time with the Hamiltonians in the randomization • ✓ L 2 method satisfies ◆ T RM = O ✏ ∆

  68. The randomization method to prepare eigenstates [] s q | x i s 2 H (1) = A.P ⊥ b .A s 1 | b i s = 0 H (0) = P ⊥ b By performing a sequence of projective measurements at sufficiently close • points, we can prepare the related eigenstates with high probability Each measurement can be simulated by evolving with the corresponding • Hamiltonian for random time. This reduces coherences between eigenstates The expected evolution time with the Hamiltonians in the randomization • ✓ L 2 method satisfies L is the path length ◆ T RM = O 𝛦 is the min gap ✏ ∆ 𝜗 is the error (trace norm)

  69. The randomization method for the QLSP [] s q | x i s 2 H (1) = A.P ⊥ b .A s 1 | b i s = 0 H (0) = P ⊥ b By performing a sequence of projective measurements at sufficiently close • points, we can prepare the related eigenstates with high probability Each measurement can be simulated by evolving with the corresponding • Hamiltonian for random time. This reduces coherences between eigenstates The expected evolution time with the Hamiltonians in the randomization • method satisfies ✓  2 log 2 (  ) ◆ T RM = O ✏

  70. The randomization method for the QLSP [] The strong dependence of the evolution time with the spectral gap suggests • one to consider other Hamiltonians that have the same eigenstate but a larger eigenvalue gap

  71. The randomization method for the QLSP [] The strong dependence of the evolution time with the spectral gap suggests • one to consider other Hamiltonians that have the same eigenstate but a larger eigenvalue gap For this problem, spectral gap amplification [10] is useful: • ✓ ◆ 0 B ( s ) H ( s ) → H 0 ( s ) = B † ( s ) ⊗ σ � + B ( s ) ⊗ σ + = B † ( s ) 0 [10] R.D. Somma and S. Boixo, SIAM J. Comp. 593 (2013)

  72. The randomization method for the QLSP [] The strong dependence of the evolution time with the spectral gap suggests • one to consider other Hamiltonians that have the same eigenstate but a larger eigenvalue gap For this problem, spectral gap amplification [10] is useful: • ✓ ◆ 0 B ( s ) H ( s ) → H 0 ( s ) = B † ( s ) ⊗ σ � + B ( s ) ⊗ σ + = B † ( s ) 0 Some results: • Let | x ( s ) i be the eigenstate of 0-eigenvalue of H ( s ). Then, | x ( s ) i | 1 i is an eigenstate of 0-eigenvalue of H 0 ( s ). p This eigenstate is separated from others by an eigenvalue gap ∆ ( s ) [10] R.D. Somma and S. Boixo, SIAM J. Comp. 593 (2013)

  73. The randomization method for the QLSP [] The strong dependence of the evolution time with the spectral gap suggests • one to consider other Hamiltonians that have the same eigenstate but a larger eigenvalue gap For this problem, spectral gap amplification is useful: • ✓ ◆ 0 B ( s ) H ( s ) → H 0 ( s ) = B † ( s ) ⊗ σ � + B ( s ) ⊗ σ + = B † ( s ) 0 Some results: • Let | x ( s ) i be the eigenstate of 0-eigenvalue of H ( s ). Then, | x ( s ) i | 1 i is an eigenstate of 0-eigenvalue of H 0 ( s ). p This eigenstate is separated from others by an eigenvalue gap ∆ ( s ) Note that the path length did not change. The only change for the RM is the • use of a different Hamiltonian.

  74. The randomization method for the QLSP [] Using the randomization method with the new Hamiltonian, the expected • evolution time is ✓  log 2 (  ) ◆ T RM = O ✏

  75. The randomization method for the QLSP [] Using the randomization method with the new Hamiltonian, the expected • evolution time is ✓  log 2 (  ) ◆ T RM = O ✏ The case of non-positive matrix A can be analyzed similarly using • A ( s ) = (1 − s )( σ anc ⊗ I ) + s ( σ anc ⊗ A ) z x

  76. The randomization method, the QLSP, and the gate model For A > 0, the Hamiltonian is H 0 ( s ) = ( I � | b ih b | )((1 � s ) I + sA ) σ � + H.c.

  77. The randomization method, the QLSP, and the gate model For A > 0, the Hamiltonian is H 0 ( s ) = ( I � | b ih b | )((1 � s ) I + sA ) σ � + H.c. If | b i is sparse and A is sparse, then H 0 ( s ) is also sparse

  78. The randomization method, the QLSP, and the gate model For A > 0, the Hamiltonian is H 0 ( s ) = ( I � | b ih b | )((1 � s ) I + sA ) σ � + H.c. If | b i is sparse and A is sparse, then H 0 ( s ) is also sparse We can use a Hamiltonian simulation method to build a quantum circuit that simulates the evolution. The quantum circuit will use queries. ˜ The complexity in terms of queries for is • | b ih b | O (  T b / ✏ ) The complexity in terms of queries for A is almost order • C A (  / ✏ , ✏ )

  79. The randomization method, the QLSP, and the gate model For A > 0, the Hamiltonian is H 0 ( s ) = ( I � | b ih b | )((1 � s ) I + sA ) σ � + H.c. If | b i is sparse and A is sparse, then H 0 ( s ) is also sparse We can use a Hamiltonian simulation method to build a quantum circuit that simulates the evolution. The quantum circuit will use queries. ˜ The complexity in terms of queries for is • | b ih b | O (  T b / ✏ ) The complexity in terms of queries for A is almost order • C A (  / ✏ , ✏ ) The scaling in the precision parameter can be done polyligarithmic by using • faster methods for eigenpath traversal [11] [11] S. Boixo, E. Knill, and R.D. Somma, arXiv:1005.3034 (2010)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend