{"id":89173,"date":"2024-09-30T09:00:00","date_gmt":"2024-09-30T16:00:00","guid":{"rendered":"https:\/\/developer.nvidia.com\/blog\/?p=89173"},"modified":"2025-07-22T16:33:12","modified_gmt":"2025-07-22T23:33:12","slug":"advancing-quantum-algorithm-design-with-gpt","status":"publish","type":"post","link":"https:\/\/developer.nvidia.com\/blog\/advancing-quantum-algorithm-design-with-gpt\/","title":{"rendered":"Advancing Quantum Algorithm Design with GPTs"},"content":{"rendered":"\n<p>AI techniques like large language models (LLMs) are rapidly transforming many scientific disciplines. <a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/quantum-computing\/\">Quantum computing<\/a> is no exception. A collaboration between NVIDIA, the University of Toronto, and Saint Jude Children\u2019s Research Hospital is bringing generative pre-trained transformers (GPTs) to the <a href=\"https:\/\/arxiv.org\/abs\/2401.09253\">design of new quantum algorithms<\/a>, including the <a href=\"https:\/\/www.nvidia.com\/en-us\/on-demand\/session\/gtc24-s62497\/\">Generative Quantum Eigensolver (GQE)<\/a> technique.&nbsp;<\/p>\n\n\n\n<p>The GQE technique is the latest in a wave of so-called <a href=\"https:\/\/developer.nvidia.com\/blog\/enabling-quantum-computing-with-ai\/\">AI for Quantum<\/a> techniques. Developed with the NVIDIA CUDA-Q platform, GQE is the first method enabling you to use your own GPT model for creating complex quantum circuits.&nbsp;<\/p>\n\n\n\n<p>The CUDA-Q platform has been instrumental in developing GQE. Training and using GPT models in quantum computing requires hybrid access to CPUs, GPUs, and QPUs. The CUDA-Q focus on <a href=\"https:\/\/developer.nvidia.com\/blog\/an-introduction-to-quantum-accelerated-supercomputing\/\">accelerated quantum supercomputing<\/a> makes it a fully hybrid computing environment perfectly suited for GQE.&nbsp;<\/p>\n\n\n\n<p>According to GQE co-author <a href=\"https:\/\/www.nvidia.com\/en-us\/on-demand\/session\/gtc24-ep64025\/?playlistId=playList-d63e5c44-196c-4246-bdca-d2a7bd2e00c1\">Alan Aspuru-Guzik<\/a>, these abilities position CUDA-Q as a scalable standard.<\/p>\n\n\n\n<h2 id=\"learning_the_grammar_of_quantum_circuits\"  class=\"wp-block-heading\">Learning the grammar of quantum circuits<a href=\"#learning_the_grammar_of_quantum_circuits\" class=\"heading-anchor-link\"><i class=\"fas fa-link\"><\/i><\/a><\/h2>\n\n\n\n<p><a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/large-language-models\/#:~:text=Large%20language%20models%20largely%20represent,the%20words%20in%20this%20sentence.\">Conventional LLMs<\/a> can be a useful analogy for understanding GQE. In general, the goal of an LLM is to take a vocabulary of many words; train a transformer model with text samples to understand things like meaning, context, and grammar; and then sample the trained model to produce words, which are then strung together to generate a new document.<\/p>\n\n\n\n<p>Where LLMs deal with words, GQE deals with quantum circuit operations. GQE takes a pool of unitary operations (vocabulary) and trains a transformer model to generate a sequence of indices corresponding to unitary operations (words) that define a resulting quantum circuit (document). The grammar for generating these indices is a set of rules trained by minimizing a cost function, which is evaluated by computing expectation values using previously generated circuits.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69efddc8875e7&quot;}\" data-wp-interactive=\"core\/image\" class=\"aligncenter size-full-page-width wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"322\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on-async--click=\"actions.showLightbox\" data-wp-on-async--load=\"callbacks.setButtonStyles\" data-wp-on-async-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-1024x322.png\" alt=\"Diagram shows generated outputs from the GQE (quantum circuit) and an LLM (sentence).\" class=\"wp-image-89187\" srcset=\"https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-1024x322.png 1024w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-300x94.png 300w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-625x197.png 625w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-179x56.png 179w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-768x241.png 768w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-1536x483.png 1536w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-645x203.png 645w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-500x157.png 500w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-160x50.png 160w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-362x114.png 362w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-350x110.png 350w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm-960x302.png 960w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-llm.png 1832w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on-async--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><figcaption class=\"wp-element-caption\"><em>Figure 1. Comparing GQE to a LLM<\/em><\/figcaption><\/figure><\/div>\n\n\n<p>Figure 1 shows that GQE is analogous to a LLM. Instead of adding individual words to construct a sentence, unitary operations are added to generate a quantum circuit.<\/p>\n\n\n\n<h2 id=\"gqe-enabled_algorithms\"  class=\"wp-block-heading\">GQE-enabled algorithms.<a href=\"#gqe-enabled_algorithms\" class=\"heading-anchor-link\"><i class=\"fas fa-link\"><\/i><\/a><\/h2>\n\n\n\n<p>In the era of noisy, small-scale quantum (NISQ) computers, quantum algorithms are limited by several hardware constraints. This has motivated the development of hybrid quantum-classical algorithms like the Variational Quantum Eigensolver (VQE), which attempts to circumvent these limitations by offloading onerous tasks to a conventional computer.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69efddc8880ff&quot;}\" data-wp-interactive=\"core\/image\" class=\"aligncenter size-full-page-width wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"675\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on-async--click=\"actions.showLightbox\" data-wp-on-async--load=\"callbacks.setButtonStyles\" data-wp-on-async-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-1024x675.png\" alt=\"For GQE, the parameters are only in the GPT model. For VQE, the variational parameters are in the quantum circuit.\" class=\"wp-image-89188\" srcset=\"https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-1024x675.png 1024w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-300x198.png 300w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-625x412.png 625w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-174x115.png 174w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-768x506.png 768w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-645x425.png 645w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-455x300.png 455w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-137x90.png 137w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-362x239.png 362w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-167x110.png 167w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe-819x540.png 819w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/gqe-vs-vqe.png 1438w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on-async--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><figcaption class=\"wp-element-caption\"><em>Figure 2. Comparison between the GQE and VQE<\/em><\/figcaption><\/figure><\/div>\n\n\n<p>All optimized parameters are handled classically in the GPT model and are updated based on the expected values of the generated circuits. This enables optimization to occur in a more favorable deep neural network landscape and offers a potential route to avoiding the <a href=\"https:\/\/www.nature.com\/articles\/s41467-018-07090-4\">barren plateaus<\/a> that impede variational algorithms. This also eliminates the need for the many intermediate circuit evaluations required in techniques like reinforcement learning.<\/p>\n\n\n\n<p>The GQE method is the first hybrid quantum-classical algorithm leveraging the power of AI to accelerate NISQ applications. GQE extends NISQ algorithms in several ways:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Ease of optimization:<\/strong> GQE builds quantum circuits without quantum variational parameters (Figure 2).&nbsp;<\/li>\n\n\n\n<li><strong>Quantum resource efficiency:<\/strong> By replacing quantum gradient evaluation with sampling and backpropagation, GQE is expected to provide greater utility with fewer quantum circuit evaluations.<\/li>\n\n\n\n<li><strong>Customizability:<\/strong> The GQE is very flexible and can be modified to incorporate <em>a priori<\/em> domain knowledge, or applied to target applications outside of chemistry.<\/li>\n\n\n\n<li><strong>Pretrainability:<\/strong> The GQE transformer can be pretrained, eliminating the need for additional quantum circuit evaluations. We discuss this later in this post.<\/li>\n<\/ul>\n\n\n\n<h2 id=\"results_from_gpt-qe\"  class=\"wp-block-heading\">Results from GPT-QE<a href=\"#results_from_gpt-qe\" class=\"heading-anchor-link\"><i class=\"fas fa-link\"><\/i><\/a><\/h2>\n\n\n\n<p>For the inaugural application of GQE, the authors built a specific model inspired by GPT-2 (referred to explicitly as GPT-QE) and used it to estimate the ground state energies of a set of small molecules.<\/p>\n\n\n\n<p>The operator pool of vocabulary was built from chemically inspired operations such as excitations and time evolution steps that were derived from a standard <a href=\"https:\/\/arxiv.org\/abs\/2402.15879\">ansatz<\/a> known as \u2018unitary coupled-clusters with single and double excitations\u2019 (UCCSD). An <em>ansatz<\/em> is an approach to parameterizing quantum circuits.<\/p>\n\n\n\n<p>Variational algorithms must be started with a \u2018best guess\u2019 initial state, generated with existing classical methods. To demonstrate GPT-QE, the authors generated an initial state using the Hartree-Fock method with an STO-3G basis set. The GPT model used in this work was identical to OpenAI\u2019s GPT-2 model, including 12 attention layers, 12 attention heads, and 768 embedding dimensions. For more information and a comprehensive technical explanation of the training process, see <em>2.2. GPT Quantum Eigensolver<\/em> in <a href=\"https:\/\/arxiv.org\/pdf\/2401.09253.pdf\">The generative quantum eigensolver (GQE) and its application for ground state search<\/a>.&nbsp;<\/p>\n\n\n\n<p>A great advantage of this technique is that it is highly parallelizable, both in terms of using GPU acceleration for the classical component and in using multiple QPUs for the quantum calculations. Since the publication of the paper, the workflow has been accelerated by parallelizing the expectation value computations of the GPT-QE sampled circuits using the NVIDIA CUDA-Q multi-QPU backend, mqpu.<\/p>\n\n\n\n<p>The <code>mqpu<\/code> backend is designed for parallel and asynchronous quantum co-processing, enabling multiple GPUs to simulate multiple QPUs. As the availability of physical quantum hardware increases, these backends can trivially be replaced with access to multiple instances of potentially varying QPU hardware.<\/p>\n\n\n\n<p>Figure 3 shows the speedup realized by using the <code>nvidia-mqpu<\/code> backend on a much larger 18-qubit CO<sub>2 <\/sub>GQE experiment. Baseline CPU computations were obtained by calculating the expectation value of 48 sampled circuits on a 56-core Intel Xeon Platinum 8480CL E5.<\/p>\n\n\n\n<p>Using a single NVIDIA H100 GPU instead of the CPU provided a 40x speedup. The CUDA-Q <code>mqpu<\/code> backend provides an additional 8x speedup by enabling asynchronous computation of the expectation values across eight GPUs using an NVIDIA DGX-H100 system.&nbsp;<\/p>\n\n\n\n<p>The authors also trained a 30-qubit CO<sub>2 <\/sub>GQE experiment for which the CPU failed. The model trained in 173 hours on a single NVIDIA H100 GPU, which was reduced to 3.5 hours when parallelized across 48 H100 GPUs.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69efddc88919d&quot;}\" data-wp-interactive=\"core\/image\" class=\"aligncenter size-full-page-width wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"643\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on-async--click=\"actions.showLightbox\" data-wp-on-async--load=\"callbacks.setButtonStyles\" data-wp-on-async-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-1024x643.png\" alt=\"The bar chart shows a 40x speedup for an NVIDIA H100 GPU and 320x for an NVIDIA DGX-H100 system.\" class=\"wp-image-89189\" srcset=\"https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-1024x643.png 1024w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-300x188.png 300w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-625x393.png 625w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-179x112.png 179w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-768x482.png 768w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-645x405.png 645w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-478x300.png 478w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-143x90.png 143w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-362x227.png 362w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-175x110.png 175w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe-860x540.png 860w, https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/expectation-value-computation-gqe.png 1280w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on-async--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><figcaption class=\"wp-element-caption\"><em>Figure 3. Expectation value computation for GQE circuit samples<\/em><\/figcaption><\/figure><\/div>\n\n\n<p>Figure 3 shows GQE circuit samples accelerated with a single NVIDIA H100 GPU or asynchronous evaluation across multiple GPUs using an NVIDIA DGX-H100.<\/p>\n\n\n\n<p>As the scale of quantum computations continues to increase, the ability to parallelize simulation workloads across multiple GPUs, and eventually QPUs, will become increasingly important.&nbsp;<\/p>\n\n\n\n<p>Beyond access to these hardware capabilities, implementing GPT-QE using CUDA-Q provided additional benefits like interoperability with GPU-accelerated libraries such as PyTorch to accelerate the classical parts of the algorithm. This is a huge benefit of the CUDA-Q platform, which also has access to the world&#8217;s fastest implementations of conventional mathematical operations through the GPU-accelerated <a href=\"https:\/\/www.nvidia.com\/en-us\/technologies\/cuda-x\">CUDA-X<\/a> libraries.&nbsp;<\/p>\n\n\n\n<p>The CUDA-Q QPU agnosticism is also key in enabling future experiments on multiple physical QPUs. Most importantly, by embodying hybrid quantum computing and offloading gradient calculations to classical processors, large-scale systems can be explored and open the door to useful quantum computing applications enabled by AI.<\/p>\n\n\n\n<h2 id=\"opportunities_to_extend_the_gqe_framework\"  class=\"wp-block-heading\">Opportunities to extend the GQE framework<a href=\"#opportunities_to_extend_the_gqe_framework\" class=\"heading-anchor-link\"><i class=\"fas fa-link\"><\/i><\/a><\/h2>\n\n\n\n<p>This collaboration is a first step towards understanding the broad opportunities for how GPT models can enable quantum supercomputing applications.<\/p>\n\n\n\n<p>Future research will hone exploring different operator pools for GQE and optimal strategies for training. This includes a focus on <em>pretraining<\/em>, a process where existing datasets can be used to either make the transformer training more efficient or aid in the convergence of the training process. This is possible if there is a sufficiently large data set available containing generated circuits and their associated expectation values. Pretrained models can also provide a warm start for training other similar models.<\/p>\n\n\n\n<p>For example, the output from a prior run would create a database of circuits and their associated ground state energies. Poorly performing circuits can be thrown away and the transformer can be trained using only the better-performing circuits, without the need for a quantum computer or simulator. This pretrained transformer can then be used as the initialization point for further training, which is expected to converge quicker and exhibit better performance.<\/p>\n\n\n\n<p>There is also a huge scope for applications using GQE outside of quantum chemistry. A collaboration between NVIDIA and Los Alamos National Lab is exploring using the ideas of GQE for geometric quantum machine learning.<\/p>\n\n\n\n<p>For more information about the GQE code, including examples, see the <a href=\"https:\/\/github.com\/cudaq-libraries\/cudaqlib\/tree\/main\/examples\/python\">GQE<\/a> GitHub repo.<\/p>\n\n\n\n<h2 id=\"explore_nvidia_tools_for_quantum_research\"  class=\"wp-block-heading\">Explore NVIDIA tools for quantum research<a href=\"#explore_nvidia_tools_for_quantum_research\" class=\"heading-anchor-link\"><i class=\"fas fa-link\"><\/i><\/a><\/h2>\n\n\n\n<p>The GQE is a novel example of how GPT models and AI in general can be used to enable many aspects of quantum computing.<\/p>\n\n\n\n<p>NVIDIA is developing hardware and software tools such as CUDA-Q to ensure scalability and acceleration of both the classical and quantum parts of hybrid workflows. For more information about NVIDIA\u2019s quantum efforts, visit the <a href=\"https:\/\/www.nvidia.com\/en-us\/solutions\/quantum-computing\/\">Quantum Computing<\/a> page.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>AI techniques like large language models (LLMs) are rapidly transforming many scientific disciplines. Quantum computing is no exception. A collaboration between NVIDIA, the University of Toronto, and Saint Jude Children\u2019s Research Hospital is bringing generative pre-trained transformers (GPTs) to the design of new quantum algorithms, including the Generative Quantum Eigensolver (GQE) technique.&nbsp; The GQE technique &hellip; <a href=\"https:\/\/developer.nvidia.com\/blog\/advancing-quantum-algorithm-design-with-gpt\/\">Continued<\/a><\/p>\n","protected":false},"author":1950,"featured_media":89473,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"publish_to_discourse":"","publish_post_category":"318","wpdc_auto_publish_overridden":"1","wpdc_topic_tags":"","wpdc_pin_topic":"","wpdc_pin_until":"","discourse_post_id":"1493930","discourse_permalink":"https:\/\/forums.developer.nvidia.com\/t\/advancing-quantum-algorithm-design-with-gpts\/308355","wpdc_publishing_response":"success","wpdc_publishing_error":"","nv_subtitle":"","ai_post_summary":"<ul><li>The Generative Quantum Eigensolver (GQE) technique uses generative pre-trained transformers (GPTs) to design new quantum algorithms, leveraging the NVIDIA CUDA-Q platform for hybrid access to CPUs, GPUs, and QPUs.<\/li><li>GQE treats quantum circuit operations as a vocabulary and generates a sequence of indices corresponding to unitary operations to define a resulting quantum circuit, optimizing parameters classically to avoid barren plateaus in variational algorithms.<\/li><li>By using the NVIDIA CUDA-Q platform, GQE achieves significant speedup, with a 40x speedup using a single NVIDIA H100 GPU and an additional 8x speedup when parallelized across eight GPUs using an NVIDIA DGX-H100 system.<\/li><\/ul>","footnotes":"","_links_to":"","_links_to_target":""},"categories":[3110],"tags":[453,2735,1914],"coauthors":[3645,4040,4050,4051],"class_list":["post-89173","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-generative-ai","tag-featured","tag-quantum-computing","tag-cluster-supercomputing","tagify_workload-generative-ai","tagify_workload-data-science","tagify_workload-simulation-modeling-design"],"acf":{"post_industry":["HPC \/ Scientific Computing"],"post_products":["CUDA"],"post_learning_levels":["Intermediate Technical"],"post_content_types":["Deep dive"],"post_collections":""},"jetpack_featured_media_url":"https:\/\/developer-blogs.nvidia.com\/wp-content\/uploads\/2024\/09\/quantum-algorithm-gpt-featured-2.png","primary_category":{"category":"Simulation \/ Modeling \/ Design","link":"https:\/\/developer.nvidia.com\/blog\/category\/simulation-modeling-design\/","id":503,"data_source":""},"nv_translations":[{"language":"zh_CN","title":"\u5229\u7528 GPT \u63a8\u8fdb\u91cf\u5b50\u7b97\u6cd5\u8bbe\u8ba1","post_id":11498}],"jetpack_shortlink":"https:\/\/wp.me\/pcCQAL-nch","jetpack_likes_enabled":true,"jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/posts\/89173","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/users\/1950"}],"replies":[{"embeddable":true,"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/comments?post=89173"}],"version-history":[{"count":4,"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/posts\/89173\/revisions"}],"predecessor-version":[{"id":89743,"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/posts\/89173\/revisions\/89743"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/media\/89473"}],"wp:attachment":[{"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/media?parent=89173"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/categories?post=89173"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/tags?post=89173"},{"taxonomy":"author","embeddable":true,"href":"https:\/\/developer-blogs.nvidia.com\/wp-json\/wp\/v2\/coauthors?post=89173"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}