Do you want to publish a course? Click here

Experimentally detecting a quantum change point via Bayesian inference

154   0   0.0 ( 0 )
 Added by Gael Sent\\'is
 Publication date 2018
  fields Physics
and research's language is English




Ask ChatGPT about the research

Detecting a change point is a crucial task in statistics that has been recently extended to the quantum realm. A source state generator that emits a series of single photons in a default state suffers an alteration at some point and starts to emit photons in a mutated state. The problem consists in identifying the point where the change took place. In this work, we consider a learning agent that applies Bayesian inference on experimental data to solve this problem. This learning machine adjusts the measurement over each photon according to the past experimental results finds the change position in an online fashion. Our results show that the local-detection success probability can be largely improved by using such a machine learning technique. This protocol provides a tool for improvement in many applications where a sequence of identical quantum states is required.



rate research

Read More

Claude Shannon proved in 1949 that information-theoretic-secure encryption is possible if the encryption key is used only once, is random, and is at least as long as the message itself. Notwithstanding, when information is encoded in a quantum system, the phenomenon of quantum data locking allows one to encrypt a message with a shorter key and still provide information-theoretic security. We present one of the first feasible experimental demonstrations of quantum data locking for direct communication and propose a scheme for a quantum enigma machine that encrypts 6 bits per photon (containing messages, new encryption keys, and forward error correction bits) with less than 6 bits per photon of encryption key while remaining information-theoretically secure.
Quantum key distribution (QKD) enables unconditionally secure communication guaranteed by the laws of physics. The last decades have seen tremendous efforts in making this technology feasible under real-life conditions, with implementations bridging ever longer distances and creating ever higher secure key rates. Readily deployed glass fiber connections are a natural choice for distributing the single photons necessary for QKD both in intra- and intercity links. Any fiber-based implementation however experiences chromatic dispersion which deteriorates temporal detection precision. This ultimately limits maximum distance and achievable key rate of such QKD systems. In this work, we address this limitation to both maximum distance and key rate and present an effective and easy-to-implement method to overcome chromatic dispersion effects. By exploiting the entangled photons frequency correlations, we make use of nonlocal dispersion compensation to improve the photons temporal correlations. Our experiment is the first implementation utilizing the inherently quantum-mechanical effect of nonlocal dispersion compensation for QKD in this way. We experimentally show an increase in key rate from 6.1 to 228.3 bits/s over 6.46 km of telecom fiber. Our approach is extendable to arbitrary fiber lengths and dispersion values, resulting in substantially increased key rates and even enabling QKD in the first place where strong dispersion would otherwise frustrate key extraction at all.
Quantum coherence, which quantifies the superposition properties of a quantum state, plays an indispensable role in quantum resource theory. A recent theoretical work [Phys. Rev. Lett. textbf{116}, 070402 (2016)] studied the manipulation of quantum coherence in bipartite or multipartite systems under the protocol Local Incoherent Operation and Classical Communication (LQICC). Here we present the first experimental realization of obtaining maximal coherence in assisted distillation protocol based on linear optical system. The results of our work show that the optimal distillable coherence rate can be reached even in one-copy scenario when the overall bipartite qubit state is pure. Moreover, the experiments for mixed states showed that distillable coherence can be increased with less demand than entanglement distillation. Our work might be helpful in the remote quantum information processing and quantum control.
Many experiments in the field of quantum foundations seek to adjudicate between quantum theory and speculative alternatives to it. This requires one to analyze the experimental data in a manner that does not presume the correctness of the quantum formalism. The mathematical framework of generalized probabilistic theories (GPTs) provides a means of doing so. We present a scheme for determining which GPTs are consistent with a given set of experimental data. It proceeds by performing tomography on the preparations and measurements in a self-consistent manner, i.e., without presuming a prior characterization of either. We illustrate the scheme by analyzing experimental data for a large set of preparations and measurements on the polarization degree of freedom of a single photon. We find that the smallest and largest GPT state spaces consistent with our data are a pair of polytopes, each approximating the shape of the Bloch Sphere and having a volume ratio of $0.977 pm 0.001$, which provides a quantitative bound on the scope for deviations from quantum theory. We also demonstrate how our scheme can be used to bound the extent to which nature might be more nonlocal than quantum theory predicts, as well as the extent to which it might be more or less contextual. Specifically, we find that the maximal violation of the CHSH inequality can be at most $1.3% pm 0.1$ greater than the quantum prediction, and the maximal violation of a particular inequality for universal noncontextuality can not differ from the quantum prediction by more than this factor on either side. The most significant loophole in this sort of analysis is that the set of preparations and measurements one implements might fail to be tomographically complete for the system of interest.
Probabilistic approaches for tensor factorization aim to extract meaningful structure from incomplete data by postulating low rank constraints. Recently, variational Bayesian (VB) inference techniques have successfully been applied to large scale models. This paper presents full Bayesian inference via VB on both single and coupled tensor factorization models. Our method can be run even for very large models and is easily implemented. It exhibits better prediction performance than existing approaches based on maximum likelihood on several real-world datasets for missing link prediction problem.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا