Experimental Deep Reinforcement Learning for Error-Robust Gate-Set Design on a Superconducting Quantum Computer

被引:79
作者
Baum, Yuval [1 ,2 ]
Amico, Mirko [1 ,2 ]
Howell, Sean [1 ,2 ]
Hush, Michael [1 ,2 ]
Liuzzi, Maggie [1 ,2 ]
Mundada, Pranav [1 ,2 ]
Merkh, Thomas [1 ,2 ]
Carvalho, Andre R. R. [1 ,2 ]
Biercuk, Michael J. [1 ,2 ,3 ]
机构
[1] Q CTRL, Sydney, NSW, Australia
[2] Q CTRL, Los Angeles, CA 90013 USA
[3] Univ Sydney, ARC Ctr Engn Quantum Syst, Sydney, NSW, Australia
来源
PRX QUANTUM | 2021年 / 2卷 / 04期
关键词
DECOHERENCE; ALGORITHM;
D O I
10.1103/PRXQuantum.2.040324
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Quantum computers promise tremendous impact across applications-and have shown great strides in hardware engineering-but remain notoriously error prone. Careful design of low-level controls has been shown to compensate for the processes that induce hardware errors, leveraging techniques from optimal and robust control. However, these techniques rely heavily on the availability of highly accurate and detailed physical models, which generally achieve only sufficient representative fidelity for the most simple operations and generic noise modes. In this work, we use deep reinforcement learning to design a universal set of error-robust quantum logic gates in runtime on a superconducting quantum computer, without requiring knowledge of a specific Hamiltonian model of the system, its controls, or its underlying error processes. We experimentally demonstrate that a fully autonomous deep-reinforcement-learning agent can design single qubit gates up to 3x faster than default DRAG operations without additional leakage error, and exhibiting robustness against calibration drifts over weeks. We then show that ZX (-pi/2) operations implemented using the cross-resonance interaction can outperform hardware default gates by over 2x and equivalently exhibit superior calibration-free performance up to 25 days post optimization. We benchmark the performance of deep-reinforcement-learning-derived gates against other black-box optimization techniques, showing that deep reinforcement learning can achieve comparable or marginally superior performance, even with limited hardware access.
引用
收藏
页数:12
相关论文
共 68 条
[1]   Qiskit pulse: programming quantum computers through the cloud with pulses [J].
Alexander, Thomas ;
Kanazawa, Naoki ;
Egger, Daniel J. ;
Capelluto, Lauren ;
Wood, Christopher J. ;
Javadi-Abhari, Ali ;
McKay, David C. .
QUANTUM SCIENCE AND TECHNOLOGY, 2020, 5 (04)
[2]   Quantum optimal control of multilevel dissipative quantum systems with reinforcement learning [J].
An, Zheng ;
Song, Hai-Jing ;
He, Qi-Kai ;
Zhou, D. L. .
PHYSICAL REVIEW A, 2021, 103 (01)
[3]   Deep reinforcement learning for quantum gate control [J].
An, Zheng ;
Zhou, D. L. .
EPL, 2019, 126 (06)
[4]  
[Anonymous], 2020, MEASUREMENT ERROR MI
[5]   Quantum supremacy using a programmable superconducting processor [J].
Arute, Frank ;
Arya, Kunal ;
Babbush, Ryan ;
Bacon, Dave ;
Bardin, Joseph C. ;
Barends, Rami ;
Biswas, Rupak ;
Boixo, Sergio ;
Brandao, Fernando G. S. L. ;
Buell, David A. ;
Burkett, Brian ;
Chen, Yu ;
Chen, Zijun ;
Chiaro, Ben ;
Collins, Roberto ;
Courtney, William ;
Dunsworth, Andrew ;
Farhi, Edward ;
Foxen, Brooks ;
Fowler, Austin ;
Gidney, Craig ;
Giustina, Marissa ;
Graff, Rob ;
Guerin, Keith ;
Habegger, Steve ;
Harrigan, Matthew P. ;
Hartmann, Michael J. ;
Ho, Alan ;
Hoffmann, Markus ;
Huang, Trent ;
Humble, Travis S. ;
Isakov, Sergei V. ;
Jeffrey, Evan ;
Jiang, Zhang ;
Kafri, Dvir ;
Kechedzhi, Kostyantyn ;
Kelly, Julian ;
Klimov, Paul V. ;
Knysh, Sergey ;
Korotkov, Alexander ;
Kostritsa, Fedor ;
Landhuis, David ;
Lindmark, Mike ;
Lucero, Erik ;
Lyakh, Dmitry ;
Mandra, Salvatore ;
McClean, Jarrod R. ;
McEwen, Matthew ;
Megrant, Anthony ;
Mi, Xiao .
NATURE, 2019, 574 (7779) :505-+
[6]  
Ball H., 2020, ARXIV200104060
[7]   Effect of noise correlations on randomized benchmarking [J].
Ball, Harrison ;
Stace, Thomas M. ;
Flammia, Steven T. ;
Biercuk, Michael J. .
PHYSICAL REVIEW A, 2016, 93 (02)
[8]   Optimized dynamical decoupling in a model quantum memory [J].
Biercuk, Michael J. ;
Uys, Hermann ;
VanDevender, Aaron P. ;
Shiga, Nobuyasu ;
Itano, Wayne M. ;
Bollinger, John J. .
NATURE, 2009, 458 (7241) :996-1000
[9]  
Borah S., 2021, ARXIV210411856
[10]   Reinforcement Learning in Different Phases of Quantum Control [J].
Bukov, Marin ;
Day, Alexandre G. R. ;
Sels, Dries ;
Weinberg, Phillip ;
Polkovnikov, Anatoli ;
Mehta, Pankaj .
PHYSICAL REVIEW X, 2018, 8 (03)