Abstract

Gaussian Boson Sampling is a model of photonic quantum computing where single-mode squeezed states are sent through linear-optical interferometers and measured using single-photon detectors. In this work, we employ a recent exact sampling algorithm for GBS with threshold detectors to perform classical simulations on the Titan supercomputer. We determine the time and memory resources as well as the amount of computational nodes required to produce samples for different numbers of modes and detector clicks. It is possible to simulate a system with 800 optical modes postselected on outputs with 20 detector clicks, producing a single sample in roughly two hours using $40\%$ of the available nodes of Titan. Additionally, we benchmark the performance of GBS when applied to dense subgraph identification, even in the presence of photon loss. We perform sampling for several graphs containing as many as 200 vertices. Our findings indicate that large losses can be tolerated and that the use of threshold detectors is preferable over using photon-number-resolving detectors postselected on collision-free outputs.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call