RIKEN Center for Quantum Computing
Quantum compilers that reduce the number of T gates are essential for minimizing the overhead of fault-tolerant quantum computation. To achieve effective T-count reduction, it is necessary to identify equivalent circuit transformation rules that are not yet utilized in existing tools. In this paper, we rewrite any given Clifford+T circuit using a Clifford block followed by a sequential Pauli-based computation, and introduce a nontrivial and ancilla-free equivalent transformation rule, the multi-product commutation relation (MCR). This rule constructs gate sequences based on specific commutation properties among multi-Pauli operators, yielding seemingly non-commutative instances that can be commuted. To evaluate whether existing compilers account for this commutation rule, we create a benchmark circuit dataset using quantum circuit unoptimization. This technique intentionally adds redundancy to the circuit while keeping its equivalence. By leveraging the known structure of the original circuit before unoptimization, this method enables a quantitative evaluation of compiler performance by measuring how closely the optimized circuit matches the original one. Our numerical experiments reveal that the transformation rule based on MCR is not yet incorporated into current compilers. This finding suggests an untapped potential for further T-count reduction by integrating MCR-aware transformations, paving the way for improvements in quantum compilers.
An architecture for measurement-based fault-tolerant quantum computation is introduced, designed for high-connectivity devices to enable megaquop to gigaquop scale operations using moderate physical qubit counts. It leverages Knill's error-correcting teleportation with self-dual CSS codes, offering a resource-efficient path to early fault-tolerant quantum computing.
To simulate plasma phenomena, large-scale computational resources have been employed in developing high-precision and high-resolution plasma simulations. One of the main obstacles in plasma simulations is the requirement of computational resources that scale polynomially with the number of spatial grids, which poses a significant challenge for large-scale modeling. To address this issue, this study presents a quantum algorithm for simulating the nonlinear electromagnetic fluid dynamics that govern space plasmas. We map it, by applying Koopman-von Neumann linearization, to the Schrödinger equation and evolve the system using Hamiltonian simulation via quantum singular value transformation. Our algorithm scales O(sNxpolylog(Nx)T)O \left(s N_x \, \mathrm{polylog} \left( N_x \right) T \right) in time complexity with ss, NxN_x, and TT being the spatial dimension, the number of spatial grid points per dimension, and the evolution time, respectively. Comparing the scaling O(sNxs(T5/4+TNx))O \left( s N_x^s \left(T^{5/4}+T N_x\right) \right) for the classical method with the finite volume scheme, this algorithm achieves polynomial speedup in NxN_x. The space complexity of this algorithm is exponentially reduced from O(sNxs)O\left( s N_x^s \right) to O(spolylog(Nx))O\left( s \, \mathrm{polylog} \left( N_x \right) \right). Numerical experiments validate that accurate solutions are attainable with smaller mm than theoretically anticipated and with practical values of mm and RR, underscoring the feasibility of the approach. As a practical demonstration, the method accurately reproduces the Kelvin-Helmholtz instability, underscoring its capability to tackle more intricate nonlinear dynamics. These results suggest that quantum computing can offer a viable pathway to overcome the computational barriers of multiscale plasma modeling.
We propose a single-shot conditional displacement gate between a trapped atom as the control qubit and a traveling light pulse as the target oscillator, mediated by an optical cavity. Classical driving of the atom synchronized with the light reflection off the cavity realizes the single-shot implementation of the crucial gate for the universal control of hybrid systems. We further derive a concise gate model incorporating cavity loss and atomic decay, facilitating the evaluation and optimization of the gate performance. This proposal establishes a key practical tool for coherently linking stationary atoms with itinerant light, a capability essential for realizing hybrid quantum information processing.
Researchers from Fujitsu and Osaka University developed a 'decoder switching' framework for real-time quantum error correction that resolves the inherent speed-accuracy tradeoff. The system combines a fast, soft-output weak decoder with a slower, accurate strong decoder, achieving logical error rates comparable to or better than the strong decoder alone while maintaining the average decoding speed of the weak decoder. The switching rate to the strong decoder was found to decay exponentially with increasing code distance.
Recent advances in combining Clifford circuits with tensor network (TN) states have shown that classically simulable disentanglers can significantly reduce entanglement, mitigating the bond-dimension bottleneck in TN simulations. In this work, we develop a variational TN framework based on Grassmann tensor networks, which natively encode fermionic statistics while preserving locality. By incorporating locally defined Clifford circuits within the fermionic formalism, we simulate benchmark models including the tight-binding and tt-VV models. Our results show that Clifford disentangling removes the classically simulable component of entanglement, leading to a reduced bond dimension and improved accuracy in ground-state energy estimates. Interestingly, imposing the natural Grassmann-evenness constraint on the Clifford circuits significantly reduces the number of disentangling gates, from 720 to just 32, yielding a far more efficient implementation. These findings highlight the potential of Clifford-augmented Grassmann TNs as a scalable and accurate tool for studying strongly correlated fermionic systems, particularly in higher dimensions.
How the detailed structure of quantum complexity emerges from quantum dynamics remains a fundamental challenge highlighted by advances in quantum simulators and information processing. The celebrated Small-Incremental-Entangling (SIE) theorem provides a universal constraint on the rate of entanglement generation, yet it leaves open the problem of fully characterizing fine entanglement structures. Here we introduce the concept of Spectral-Entangling strength, which captures the structural entangling power of an operator, and establish a spectral SIE theorem: a universal speed limit for R'enyi entanglement growth at α1/2\alpha \ge 1/2, revealing a robust 1/s21/s^2 decay threshold in the entanglement spectrum. Remarkably, our bound at α=1/2\alpha=1/2 is both qualitatively and quantitatively optimal, defining the universal threshold beyond which entanglement growth becomes unbounded. This exposes the detailed structure of Schmidt coefficients and enables rigorous truncation-based error control, linking entanglement structure to computational complexity. Building on this, we derive a generalized entanglement area law under an adiabatic-path condition, extending a central principle of quantum many-body physics to general interactions. As a concrete application, we show that one-dimensional long-range interacting systems admit polynomial bond-dimension approximations for ground, time-evolved, and thermal states, thereby closing the long-standing quasi-polynomial gap and demonstrating that such systems can be simulated efficiently with tensor-network methods. By explicitly controlling R'enyi entanglement, we obtain a rigorous, a priori error guarantee for the time-dependent density-matrix renormalization-group algorithm. Overall, our results extend the SIE theorem to the spectral domain and establish a unified framework that unveils the detailed and universal structure underlying quantum complexity.
Keisuke Fujii introduces Out-of-Time-Order Correlator (OTOC) spectroscopy, an algorithmic interpretation of higher-order OTOCs using Quantum Signal Processing (QSP). This framework establishes OTOCs as measurements of specific Fourier components of singular value distributions, enabling frequency-selective probing of quantum scrambling and diverse many-body dynamics.
A full-stack analog optical quantum computer has been developed, featuring 100 inputs and operating at a 100 MHz clock frequency. This system demonstrated robust 101-mode, 1000-step parallel quantum teleportation and performed quantum state sorting for a hundred input modes through a cloud-accessible Python SDK.
We propose the quasi-Monte Carlo method for linear combination of unitaries via classical post-processing (LCU-CPP) on quantum applications. The LCU-CPP framework has been proposed as an approach to reduce hardware resources, expressing a general target operator F(A)F(A) as F(A)=Vf(t)G(A,t)dtF(A) = \int_V f(t) G(A, t)dt, where each G(A,t)G(A, t) is proportional to a unitary operator. On a quantum device, Re[Tr(G(A,t)ρ)]Re[Tr(G(A, t)\rho)] can be estimated using the Hadamard test and then combined through classical integration, allowing for the realization of nonunitary functions with reduced circuit depth. While previous studies have employed the Monte Carlo method or the trapezoid rule to evaluate the integral in LCU-CPP, we show that the quasi-Monte Carlo method can achieve even lower errors. In two numerical experiments, ground state property estimation and Green's function estimation, the quasi-Monte Carlo method achieves the lowest errors with a number of Hadamard test shots per unitary that is practical for real hardware implementations. These results indicate that quasi-Monte Carlo is an effective integration strategy within the LCU-CPP framework.
High-rate and large-distance quantum codes are expected to make fault-tolerant quantum computing more efficient, but most of them lack efficient fault-tolerant encoded-state preparation methods. We propose such a fault-tolerant encoder for a [[30, 6, 5]] symplectic double code. The advantage of this code is its compactness, in addition to its high encoding rate, allowing for early experimental realization. Detecting crucial errors during encoding with as few auxiliary qubits as possible, our encoder can reduce resource overheads while keeping low logical error rates, compared to more naive methods. Numerical simulations with a circuit-level noise model demonstrate the reliability and effectiveness of the proposed method. We also develop an arbitrary-state encoder that enables the injection of arbitrary quantum states into the code space. Combined with basic fault-tolerant operations, this supports universal quantum computation. We thus demonstrate that efficient and reliable logical state preparation is achievable even for a compact and high-rate code, offering a potential step toward efficient fault-tolerant quantum computing suitable for near-term experiments.
Bosonic codes, leveraging infinite-dimensional Hilbert spaces for redundancy, offer great potential for encoding quantum information. However, the realization of a practical continuous-variable bosonic code that can simultaneously correct both single-photon loss and dephasing errors remains elusive, primarily due to the absence of exactly orthogonal codewords and the lack of an experiment-friendly state preparation scheme. Here, we propose a code based on the superposition of squeezed Fock states with an error-correcting capability that scales as exp(7r)\propto\exp(-7r), where rr is the squeezing level. The codewords remain orthogonal at all squeezing levels. The Pauli-X operator acts as a rotation in phase space is an error-transparent gate, preventing correctable errors from propagating outside the code space during logical operations. In particular, this code achieves high-precision error correction for both single-photon loss and dephasing, even at moderate squeezing levels. Building on this code, we develop quantum error correction schemes that exceed the break-even threshold, supported by analytical derivations of all necessary quantum gates. Our code offers a competitive alternative to previous encodings for quantum computation using continuous bosonic qubits.
This study investigates the use of spiral geometry in superconducting resonators to achieve high intrinsic quality factors, crucial for applications in quantum computation and quantum sensing. We fabricated Archimedean Spiral Resonators (ASRs) using domain-matched epitaxially grown titanium nitride (TiN) on silicon wafers, achieving intrinsic quality factors of $Q_\mathrm{i} = (9.6 \pm 1.5) \times 10^6atthesinglephotonleveland at the single-photon level and Q_\mathrm{i} = (9.91 \pm 0.39) \times 10^7$ at high power, significantly outperforming traditional coplanar waveguide (CPW) resonators. We conducted a comprehensive numerical analysis using COMSOL to calculate surface participation ratios (PRs) at critical interfaces: metal-air, metal-substrate, and substrate-air. Our findings reveal that ASRs have lower PRs than CPWs, explaining their superior quality factors and reduced coupling to two-level systems (TLSs).
Researchers at the RIKEN Center for Quantum Computing developed a quantum algorithm for thermal state preparation that leverages optimized generalized statistical ensembles within the Quantum Singular Value Transformation (QSVT) framework. This approach substantially reduces the computational cost, achieving up to two orders of magnitude fewer queries to the Hamiltonian block-encoding for N=50 systems compared to conventional canonical ensemble methods.
Runtime optimization of the quantum computing within a given computational resource is important to achieve practical quantum advantage. In this paper, we propose a runtime reduction protocol for the lattice surgery, which utilizes the soft information corresponding to the logical measurement error. Our proposal is a simple two-step protocol: operating the lattice surgery with the small number of syndrome measurement cycles, and reexecuting it with full syndrome measurement cycles in cases where the time-like soft information catches logical error symptoms. We firstly discuss basic features of the time-like complementary gap as the concrete example of the time-like soft information based on numerical results. Then, we show that our protocol surpasses the existing runtime reduction protocol called temporally encoded lattice surgery (TELS) for the most cases. In addition, we confirm that the combination of our protocol and the TELS protocol can reduce the runtime further, over 50% in comparison to the naive serial execution of the lattice surgery. The proposed protocol in this paper can be applied to any quantum computing architecture based on the lattice surgery, and we expect that this will be one of the fundamental building blocks of runtime optimization to achieve practical scale quantum computing.
A quantum computer with low-error, high-speed quantum operations and capability for interconnections is required for useful quantum computations. A logical qubit called Gottesman-Kitaev-Preskill (GKP) qubit in a single Bosonic harmonic oscillator is efficient for mitigating errors in a quantum computer. The particularly intriguing prospect of GKP qubits is that entangling gates as well as syndrome measurements for quantum error correction only require efficient, noise-robust linear operations. To date, however, GKP qubits have been only demonstrated at mechanical and microwave frequency in a highly nonlinear physical system. The physical platform that naturally provides the scalable linear toolbox is optics, including near-ideal loss-free beam splitters and near-unit efficiency homodyne detectors that allow to obtain the complete analog syndrome for optimized quantum error correction. Additional optical linear amplifiers and specifically designed GKP qubit states are then all that is needed for universal quantum computing. In this work, we realize a GKP state in propagating light at the telecommunication wavelength and demonstrate homodyne meausurements on the GKP states for the first time without any loss corrections. Our GKP states do not only show non-classicality and non-Gaussianity at room temperature and atmospheric pressure, but unlike the existing schemes with stationary qubits, they are realizable in a propagating wave system. This property permits large-scale quantum computation and interconnections, with strong compatibility to optical fibers and 5G telecommunication technology.
Quantum error-correcting codes with high encoding rate are good candidates for large-scale quantum computers as they use physical qubits more efficiently than codes of the same distance that encode only a few logical qubits. Some logical gate of a high-rate code can be fault-tolerantly implemented using transversal physical gates, but its logical operation may depend on the choice of a symplectic basis that defines logical Pauli operators of the code. In this work, we focus on [ ⁣[n,k,d] ⁣][\![n,k,d]\!] self-dual Calderbank-Shor-Steane (CSS) codes with k1k \geq 1 and prove necessary and sufficient conditions for the code to have a symplectic basis such that (1) transversal logical Hadamard gates j=1kHˉj\bigotimes_{j=1}^{k} \bar{H}_j can be implemented by transversal physical Hadamard gates i=1nHi\bigotimes_{i=1}^{n} H_i, and (2) for any (a1,,ak){1,1}k(a_1,\dots,a_k)\in\{-1,1\}^k, transversal logical phase gates j=1kSˉjaj\bigotimes_{j=1}^{k} \bar{S}_j^{a_j} can be implemented by transversal physical phase gates i=1nSibi\bigotimes_{i=1}^{n} S_i^{b_i} for some (b1,,bn){1,1}n(b_1,\dots,b_n)\in\{-1,1\}^n. Self-dual CSS codes satisfying the conditions include any codes with odd nn. We also generalize the idea to concatenated self-dual CSS codes and show that certain logical Clifford gates have multiple transversal implementations, each by logical gates at a different level of concatenation. Several applications of our results for fault-tolerant quantum computation with low overhead are also provided.
Counterdiabatic (CD) protocols enable fast driving of quantum states by invoking an auxiliary adiabatic gauge potential (AGP) that suppresses transitions to excited states throughout the driving process. Usually, the full spectrum of the original unassisted Hamiltonian is a prerequisite for constructing the exact AGP, which implies that CD protocols are extremely difficult for many-body systems. Here, we apply a variational CD protocol recently proposed by P. W. Claeys et al. [Phys. Rev. Lett. 123, 090602 (2019)] to a two-component fermionic Hubbard model in one spatial dimension. This protocol engages an approximated AGP expressed as a series of nested commutators. We show that the optimal variational parameters in the approximated AGP satisfy a set of linear equations whose coefficients are given by the squared Frobenius norms of these commutators. We devise an exact algorithm that escapes the formidable iterative matrix-vector multiplications and evaluates the nested commutators and the CD Hamiltonian in analytic representations. We then examine the CD driving of the one-dimensional Hubbard model up to L=14L = 14 sites with driving order l3l \leqslant 3. Our results demonstrate the usefulness of the variational CD protocol to the Hubbard model and permit a possible route towards fast ground-state preparation for many-body systems.
Cluster states are a class of multi-qubit entangled states with broad applications such as quantum metrology and one-way quantum computing. Here, we present a protocol to generate frequency-bin-encoded dual-rail cluster states using a superconducting circuit consisting of a fixed-frequency transmon qubit, a resonator and a Purcell filter. We implement time-frequency multiplexing by sequentially emitting co-propagating microwave photons of distinct frequencies. The frequency-bin dual-rail encoding enables erasure detection based on photon occupancy. We characterize the state fidelity using quantum tomography and quantify the multipartite entanglement using the metric of localizable entanglement. Our implementation achieves a state fidelity exceeding 50%\% for a cluster state consisting of up to four logical qubits. The localizable entanglement remains across chains of up to seven logical qubits. After discarding the erasure errors, the fidelity exceeds 50% for states with up to eight logical qubits, and the entanglement persists across chains of up to eleven qubits. These results highlight the improved robustness of frequency-bin dual-rail encoding against photon loss compared to conventional single-rail schemes. This work provides a scalable pathway toward high-dimensional entangled state generation and photonic quantum information processing in the microwave domain.
We propose a pulse and continuous wave (CW) hybrid architecture of continuous-variable measurement-based optical quantum computation utilizing the strengths of both pulsed and CW light. In this architecture, input and ancillary non-Gaussian quantum states necessary for fault-tolerance and universality of quantum computing are generated with pulsed light, whereas quantum processors including continuous-variable cluster states and homodyne measurement systems are operated with CW light. This architecture is expected to enable both generation of quantum states with shorter optical wavepackets and low-loss manipulation and measurement of these states, thus is compatible with ultrafast and low-loss quantum information processing. In this study, as a proof-of-principle, an ultrafast homodyne measurement using CW local oscillator was performed on single-photon states generated with pulsed light. The measured single-photon state's temporal width was around 70 ps and the value of the Wigner function at the origin was W(0,0) = -0.153 +/- 0.003, which is highly non-classical. This will be a core technology for realizing high-speed optical quantum information processing.
There are no more papers matching your filters at the moment.