Quantum Computing

Assembly Line

Booz Allen Ventures Invests in Quantum Hardware Innovation for National Security

📅 Date:

🔖 Topics: Quantum Computing, Funding Event

🏢 Organizations: Booz Allen, SEEQC


Booz Allen Hamilton’s corporate venture capital arm, Booz Allen Ventures, has made a strategic investment in SEEQC, a quantum computing company. This collaboration will accelerate the development and deployment of quantum computing technology to support government clients. Quantum computing has the potential to transform areas such as drug discovery, financial modeling, and logistics optimization. SEEQC is developing innovative approaches to quantum hardware to enable the rapid scaling of quantum computers, addressing barriers such as latency, energy, accuracy, and cost. The partnership will drive forward hardware and software improvements essential for scalable quantum computing, helping unlock new levels of problem-solving power for national security, civil, and commercial missions.

Read more at Business Wire

Meet Willow, our state-of-the-art quantum chip

📅 Date:

✍️ Author: Hartmut Neven

🔖 Topics: Quantum Computing

🏭 Vertical: Semiconductor

🏢 Organizations: Google


Published in Nature, our results show that the more qubits we use in Willow, the more we reduce errors, and the more quantum the system becomes. We tested ever-larger arrays of physical qubits, scaling up from a grid of 3x3 encoded qubits, to a grid of 5x5, to a grid of 7x7 — and each time, using our latest advances in quantum error correction, we were able to cut the error rate in half. In other words, we achieved an exponential reduction in the error rate. This historic accomplishment is known in the field as “below threshold” — being able to drive errors down while scaling up the number of qubits. You must demonstrate being below threshold to show real progress on error correction, and this has been an outstanding challenge since quantum error correction was introduced by Peter Shor in 1995.

Read more at Google Blog

NVIDIA Accelerates Google Quantum AI Processor Design With Simulation of Quantum Device Physics

📅 Date:

🔖 Topics: Partnership, Quantum Computing

🏢 Organizations: NVIDIA, Google


NVIDIA announced it is working with Google Quantum AI to accelerate the design of its next-generation quantum computing devices using simulations powered by the NVIDIA CUDA-Q™ platform. Google Quantum AI is using the hybrid quantum-classical computing platform and the NVIDIA Eos supercomputer to simulate the physics of its quantum processors. This will help overcome the current limitations of quantum computing hardware, which can only run a certain number of quantum operations before computations must cease, due to what researchers call “noise.”

Understanding noise in quantum hardware designs requires complex dynamical simulations capable of fully capturing how qubits within a quantum processor interact with their environment. These simulations have traditionally been prohibitively computationally expensive to pursue. Using the CUDA-Q platform, however, Google can employ 1,024 NVIDIA H100 Tensor Core GPUs at the NVIDIA Eos supercomputer to perform one of the world’s largest and fastest dynamical simulation of quantum devices — at a fraction of the cost.

Read more at NVIDIA News

Quantum error correction below the surface code threshold

📅 Date:

🔖 Topics: Quantum Computing

🏢 Organizations: Google


Quantum error correction provides a path to reach practical quantum computing by combining multiple physical qubits into a logical qubit, where the logical error rate is suppressed exponentially as more qubits are added. However, this exponential suppression only occurs if the physical error rate is below a critical threshold. In this work, we present two surface code memories operating below this threshold: a distance-7 code and a distance-5 code integrated with a real-time decoder. The logical error rate of our larger quantum memory is suppressed by a factor of Λ = 2.14 ± 0.02 when increasing the code distance by two, culminating in a 101-qubit distance-7 code with 0.143% ± 0.003% error per cycle of error correction. This logical memory is also beyond break-even, exceeding its best physical qubit’s lifetime by a factor of 2.4 ± 0.3. We maintain below-threshold performance when decoding in real time, achieving an average decoder latency of 63 μs at distance-5 up to a million cycles, with a cycle time of 1.1 μs. To probe the limits of our error-correction performance, we run repetition codes up to distance-29 and find that logical performance is limited by rare correlated error events occurring approximately once every hour, or 3 × 109 cycles. Our results present device performance that, if scaled, could realize the operational requirements of large scale fault-tolerant quantum algorithms.

Read more at arXiv

High-threshold and low-overhead fault-tolerant quantum memory

📅 Date:

✍️ Authors: Sergey Bravyi, Andrew Cross, Jay Gambettav

🔖 Topics: Quantum Computing, Quantum Error Correction

🏢 Organizations: IBM


The accumulation of physical errors prevents the execution of large-scale algorithms in current quantum computers. Quantum error correction4 promises a solution by encoding k logical qubits onto a larger number n of physical qubits, such that the physical errors are suppressed enough to allow running a desired computation with tolerable fidelity. Quantum error correction becomes practically realizable once the physical error rate is below a threshold value that depends on the choice of quantum code, syndrome measurement circuit and decoding algorithm5. We present an end-to-end quantum error correction protocol that implements fault-tolerant memory on the basis of a family of low-density parity-check codes6. Our approach achieves an error threshold of 0.7% for the standard circuit-based noise model, on par with the surface code7,8,9,10 that for 20 years was the leading code in terms of error threshold. The syndrome measurement cycle for a length-n code in our family requires n ancillary qubits and a depth-8 circuit with CNOT gates, qubit initializations and measurements. The required qubit connectivity is a degree-6 graph composed of two edge-disjoint planar subgraphs. In particular, we show that 12 logical qubits can be preserved for nearly 1 million syndrome cycles using 288 physical qubits in total, assuming the physical error rate of 0.1%, whereas the surface code would require nearly 3,000 physical qubits to achieve said performance. Our findings bring demonstrations of a low-overhead fault-tolerant quantum memory within the reach of near-term quantum processors.

Read more at nature

Intelligent synthesis of magnetic nanographenes via chemist-intuited atomic robotic probe

📅 Date:

🔖 Topics: Quantum Computing

🏢 Organizations: National University of Singapore


Atomic-scale manufacturing of carbon-based quantum materials with single-bond precision holds immense potential in advancing tailor-made quantum materials with unconventional properties, which are crucial in developing next-generation spintronic devices and quantum information technologies. On-surface chemistry approaches, including surface-assisted synthesis and probe-assisted manipulation, are impeded by challenges in reaction selectivity control or restricted by scalability and production efficiency. Here we demonstrate the concept of the chemist-intuited atomic robotic probe by integrating probe chemistry knowledge and artificial intelligence, allowing for atomically precise single-molecule manipulation to fabricate single-molecule quantum π-magnets with single-bond precision. Our deep neural networks not only transform complex probe chemistry into machine-understandable tasks but also provide chemist intuition to elusive reaction mechanisms by extracting the critical chemical information within the data. A joint experimental and theoretical investigation demonstrates that a voltage-controlled two-electron-assisted electronic excitation enables synchronous six-bond transformations to extend the zigzag edge topology of single-molecule quantum π-magnets, triggered by phenyl C(sp2)–H bond activation, which aligns with initial conjectures given by the deep neural models. Our work represents a transition from autonomous fabrication to intelligent synthesis with levels of selectivity and precision beyond current synthetic tools for improved synthesis of organic quantum materials towards on-chip integration.

Read more at Nature

🧠🤖 Optimising Intralogistics with AI

📅 Date:

🔖 Topics: Intralogistics, Automated Guided Vehicle, AutoQML, Quantum Computing

🏢 Organizations: KEB Automation, Fraunhofer IAO


In its production facilities in Barntrup, KEB operates the in-house transport system AGILOX, which is designed specifically for intralogistics tasks. The AGILOX system is comprised of a swarm (union) of smart automated guided vehicles (AGVs), working collaboratively to transport items throughout KEB’s warehouses.

In AutoQML – a project that develops solution approaches for linking quantum computing and machine learning – KEBs primary objective is to devise a machine learning solution capable of monitoring vehicle status and predicting potential failures. This aligns with KEBs larger objective of facilitating the broader transition to quantum computing in the future, by supporting research institutes with practical, real-world applications.

Read more at KEB Automation Blog

Bosch’s new partnership aims to explore quantum digital twins

📅 Date:

🔖 Topics: Quantum Computing, Simulation

🏢 Organizations: Bosch, Multiverse Computing


Industrial giant Bosch has partnered with Multiverse Computing, a Spanish quantum software platform, to integrate quantum algorithms into digital twin simulation workflows. Bosch already has an extensive industrial simulation practice that provides insights across various business units. This new collaboration will explore ways quantum-inspired algorithms and computers could help scale these simulations more efficiently.

One of the most promising use cases for the new quantum algorithms is creating better machine learning models more quickly. Hernández Caballer said quantum computing shows tremendous promise in use cases with many combinations of parameters and materials. This early research could give Bosch a leg up in taking advantage of these new systems to improve machine learning and simulation.

Read more at VentureBeat