No Arabic abstract
Boson sampling is a well-defined task that is strongly believed to be intractable for classical computers, but can be efficiently solved by a specific quantum simulator. However, an outstanding problem for large-scale experimental boson sampling is the scalability. Here we report an experiment on boson sampling with photon loss, and demonstrate that boson sampling with a few photons lost can increase the sampling rate. Our experiment uses a quantum-dot-micropillar single-photon source demultiplexed into up to seven input ports of a 16*16 mode ultra-low-loss photonic circuit, and we detect three-, four- and five-fold coincidence counts. We implement and validate lossy boson sampling with one and two photons lost, and obtain sampling rates of 187 kHz, 13.6 kHz, and 0.78 kHz for five-, six- and seven-photon boson sampling with two photons lost, which is 9.4, 13.9, and 18.0 times faster than the standard boson sampling, respectively. Our experiment shows an approach to significantly enhance the sampling rate of multiphoton boson sampling.
Entangled photon sources with simultaneously near-unity heralding efficiency and indistinguishability are the fundamental elements for scalable photonic quantum technologies. We design and realize a degenerate entangled-photon source from an ultrafast pulsed laser pumped spontaneous parametric down-conversion (SPDC), which show simultaneously ~97% heralding efficiency and ~96% indistinguishability between independent single photons. Such a high-efficiency and frequency-uncorrelated SPDC source allows generation of the first 12-photon genuine entanglement with a state fidelity of 0.572(24). We further demonstrate a blueprint of scalable scattershot boson sampling using 12 SPDC sources and a 12*12-modes interferometer for three-, four-, and five-boson sampling, which yields count rates more than four orders of magnitudes higher than all previous SPDC experiments. Our work immediately enables high-efficiency implementations of multiplexing, scattershot boson sampling, and heralded creation of remotely entangled photons, opening up a promising pathway to scalable photonic quantum technologies.
Boson sampling is considered as a strong candidate to demonstrate the quantum computational supremacy over classical computers. However, previous proof-of-principle experiments suffered from small photon number and low sampling rates owing to the inefficiencies of the single-photon sources and multi-port optical interferometers. Here, we develop two central components for high-performance boson sampling: robust multi-photon interferometers with 0.99 transmission rate, and actively demultiplexed single-photon sources from a quantum-dot-micropillar with simultaneously high efficiency, purity and indistinguishability. We implement and validate 3-, 4-, and 5-photon boson sampling, and achieve sampling rates of 4.96 kHz, 151 Hz, and 4 Hz, respectively, which are over 24,000 times faster than the previous experiments, and over 220 times faster than obtaining one sample through calculating the matrices permanent using the first electronic computer (ENIAC) and transistorized computer (TRADIC) in the human history. Our architecture is feasible to be scaled up to larger number of photons and with higher rate to race against classical computers, and might provide experimental evidence against the Extended Church-Turing Thesis.
Universal quantum computers promise a dramatic speed-up over classical computers but a full-size realization remains challenging. However, intermediate quantum computational models have been proposed that are not universal, but can solve problems that are strongly believed to be classically hard. Aaronson and Arkhipov have shown that interference of single photons in random optical networks can solve the hard problem of sampling the bosonic output distribution which is directly connected to computing matrix permanents. Remarkably, this computation does not require measurement-based interactions or adaptive feed-forward techniques. Here we demonstrate this model of computation using high--quality laser--written integrated quantum networks that were designed to implement random unitary matrix transformations. We experimentally characterize the integrated devices using an in--situ reconstruction method and observe three-photon interference that leads to the boson-sampling output distribution. Our results set a benchmark for quantum computers, that hold the potential of outperforming conventional ones using only a few dozen photons and linear-optical elements.
Sampling the distribution of bosons that have undergone a random unitary evolution is strongly believed to be a computationally hard problem. Key to outperforming classical simulations of this task is to increase both the number of input photons and the size of the network. We propose driven boson sampling, in which photons are input within the network itself, as a means to approach this goal. When using heralded single-photon sources based on parametric down-conversion, this approach offers an $sim e$-fold enhancement in the input state generation rate over scattershot boson sampling, reaching the scaling limit for such sources. More significantly, this approach offers a dramatic increase in the signal-to-noise ratio with respect to higher-order photon generation from such probabilistic sources, which removes the need for photon number resolution during the heralding process as the size of the system increases.
Boson Sampling has emerged as a tool to explore the advantages of quantum over classical computers as it does not require a universal control over the quantum system, which favours current photonic experimental platforms.Here, we introduce Gaussian Boson Sampling, a classically hard-to-solve problem that uses squeezed states as a non-classical resource. We relate the probability to measure specific photon patterns from a general Gaussian state in the Fock basis to a matrix function called the hafnian, which answers the last remaining question of sampling from Gaussian states. Based on this result, we design Gaussian Boson Sampling, a #P hard problem, using squeezed states. This approach leads to a more efficient photonic boson sampler with significant advantages in generation probability and measurement time over currently existing protocols.