Unleashing the Potential of LLMs for Quantum Computing: A Study in Quantum Architecture Design

Large Language Models (LLMs) contribute significantly to the development of conversational AI and has great potentials to assist the scientific research in various areas. This paper attempts to address the following questions: What opportunities do the current generation of generative pre-trained transformers (GPTs) offer for the developments of noisy intermediate-scale quantum (NISQ) technologies? Additionally, what potentials does the forthcoming generation of GPTs possess to push the frontier of research in fault-tolerant quantum computing (FTQC)? In this paper, we implement a QGAS model, which can rapidly propose promising ansatz architectures and evaluate them with application benchmarks including quantum chemistry and quantum finance tasks. Our results demonstrate that after a limited number of prompt guidelines and iterations, we can obtain a high-performance ansatz which is able to produce comparable results that are achieved by state-of-the-art quantum architecture search methods. This study provides a simple overview of GPT's capabilities in supporting quantum computing research while highlighting the limitations of the current GPT at the same time. Additionally, we discuss futuristic applications for LLM in quantum research.

[1]  P. Warburton,et al.  Quantum computing and materials science: A practical guide to applying quantum annealing to the configurational analysis of materials , 2023, Journal of Applied Physics.

[2]  Yiyu Shi,et al.  On the Viability of Using LLMs for SW/HW Co-Design: An Example in Designing CiM DNN Accelerators , 2023, 2023 IEEE 36th International System-on-Chip Conference (SOCC).

[3]  Leibny Paola Garcia,et al.  PQLM - Multilingual Decentralized Portable Quantum Language Model , 2023, ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[4]  Shouvanik Chakrabarti,et al.  Alignment between Initial State and Mixer Improves QAOA Performance for Constrained Portfolio Optimization , 2023, ArXiv.

[5]  Linyang Li,et al.  Improving Contrastive Learning of Sentence Embeddings from AI Feedback , 2023, ACL.

[6]  Fei Wang,et al.  Can GPT-4 Perform Neural Architecture Search? , 2023, ArXiv.

[7]  P. Kogge,et al.  Towards Advantages of Parameterized Quantum Pulses , 2023, 2304.09253.

[8]  P. Schwaller,et al.  ChemCrow: Augmenting large-language models with chemistry tools , 2023, 2304.05376.

[9]  Richard van Noorden,et al.  What ChatGPT and generative AI mean for science , 2023, Nature.

[10]  E. A. V. van Dis,et al.  ChatGPT: five priorities for research , 2023, Nature.

[11]  Jie Pan Large language model for molecular chemistry , 2023, Nature Computational Science.

[12]  M. Hillery,et al.  Discrete outcome quantum sensor networks , 2022, Physical Review A.

[13]  Yiyu Shi,et al.  TorchQuantum Case Study for Robust Quantum Circuits , 2022, International Conference on Computer Aided Design.

[14]  Betis Baheri,et al.  Pinpointing the System Reliability Degradation in NISQ Machines , 2022, International Conference on Quantum Computing and Engineering.

[15]  Youzuo Lin,et al.  Quantum Neural Network Compression , 2022, 2022 IEEE/ACM International Conference On Computer Aided Design (ICCAD).

[16]  Chao-Han Huck Yang,et al.  Theoretical error performance analysis for variational quantum circuit based functional regression , 2022, npj Quantum Information.

[17]  R. Wille,et al.  Adaptive Compilation of Multi-Level Quantum Operations , 2022, 2022 IEEE International Conference on Quantum Computing and Engineering (QCE).

[18]  Xuehai Qian,et al.  Variational Quantum Pulse Learning , 2022, 2022 IEEE International Conference on Quantum Computing and Engineering (QCE).

[19]  Ryan J. Lowe,et al.  Training language models to follow instructions with human feedback , 2022, NeurIPS.

[20]  Lingpeng Kong,et al.  ZeroGen: Efficient Zero-shot Learning via Dataset Generation , 2022, EMNLP.

[21]  B. Höcker,et al.  Controllable protein design with language models , 2022, Nature Machine Intelligence.

[22]  Jinjun Xiong,et al.  Exploration of Quantum Neural Architecture by Mixing Quantum Neuron Designs: (Invited Paper) , 2021, 2021 IEEE/ACM International Conference On Computer Aided Design (ICCAD).

[23]  David Z. Pan,et al.  QuantumNAS: Noise-Adaptive Search for Robust Quantum Circuits , 2021, 2022 IEEE International Symposium on High-Performance Computer Architecture (HPCA).

[24]  Diane M. Korngiebel,et al.  Considering the possibilities and pitfalls of Generative Pre-trained Transformer 3 (GPT-3) in healthcare delivery , 2021, npj Digital Medicine.

[25]  Frederic T. Chong,et al.  Virtual Logical Qubits: A Compact Architecture for Fault-Tolerant Quantum Computing , 2021, IEEE Micro.

[26]  C. Gogolin,et al.  Local, expressive, quantum-number-preserving VQE ansätze for fermionic systems , 2021, New Journal of Physics.

[27]  M. Hutson Robo-writers: the rise and risks of language-generating AI , 2021, Nature.

[28]  L. Cincio,et al.  Layer VQE: A Variational Approach for Combinatorial Optimization on Noisy Quantum Computers , 2021, IEEE Transactions on Quantum Engineering.

[29]  Min-Hsiu Hsieh,et al.  Quantum circuit architecture search for variational quantum algorithms , 2020, npj Quantum Information.

[30]  Mark Chen,et al.  Language Models are Few-Shot Learners , 2020, NeurIPS.

[31]  J. Biamonte,et al.  Variational Quantum Eigensolver for Frustrated Quantum Systems , 2020, ArXiv.

[32]  Xuehai Qian,et al.  AccQOC: Accelerating Quantum Optimal Control Based Pulse Generation , 2020, 2020 ACM/IEEE 47th Annual International Symposium on Computer Architecture (ISCA).

[33]  S. Lloyd,et al.  Quantum embeddings for machine learning , 2020, 2001.03622.

[34]  Harper R. Grimsley,et al.  Is the Trotterized UCCSD Ansatz Chemically Well-Defined? , 2019, Journal of chemical theory and computation.

[35]  Arthur G. Rattew,et al.  A Domain-agnostic, Noise-resistant, Hardware-efficient Evolutionary Variational Quantum Eigensolver , 2019, 1910.09694.

[36]  Giacomo Nannicini,et al.  Improving Variational Quantum Optimization using CVaR , 2019, Quantum.

[37]  Maxwell Henderson,et al.  Quanvolutional neural networks: powering image recognition with quantum circuits , 2019, Quantum Machine Intelligence.

[38]  Thomas Alexander,et al.  Qiskit Backend Specifications for OpenQASM and OpenPulse Experiments , 2018, ArXiv.

[39]  Ryan Babbush,et al.  Barren plateaus in quantum neural network training landscapes , 2018, Nature Communications.

[40]  Hartmut Neven,et al.  Classification with Quantum Neural Networks on Near Term Processors , 2018, 1802.06002.

[41]  Shane Legg,et al.  Deep Reinforcement Learning from Human Preferences , 2017, NIPS.

[42]  J. Gambetta,et al.  Hardware-efficient variational quantum eigensolver for small molecules and quantum magnets , 2017, Nature.

[43]  Joanna Bryson,et al.  Semantics derived automatically from language corpora contain human-like biases , 2016, Science.

[44]  P. Coveney,et al.  Scalable Quantum Simulation of Molecular Energies , 2015, 1512.06860.

[45]  Alán Aspuru-Guzik,et al.  A variational eigenvalue solver on a photonic quantum processor , 2013, Nature Communications.

[46]  Xuehai Qian,et al.  PAN: Pulse Ansatz on NISQ Machines , 2022, ArXiv.

[47]  K. Brown,et al.  Stable Turnkey Laser System for a Yb/Ba Trapped-Ion Quantum Computer , 2022, IEEE Transactions on Quantum Engineering.

[48]  Himanshu Gupta,et al.  Transmitter Localization using Quantum Sensor Networks , 2022 .

[49]  Song Han,et al.  On-chip QNN: Towards Efficient On-Chip Training of Quantum Neural Networks , 2022, ArXiv.