Speaker
Description
Being able to efficiently represent mixed quantum states is essential in order to describe the effects of dissipation, such as those arising in Open Quantum Systems, or in order to represent the noisy outcome of a circuit executed on a present-day Quantum Computer.
The challenges in the description of such objects arise from the exponential growth of the Hilbert space and from the need to enforce the positive-definiteness of the resulting matrix.
A compact, physical representation of density matrices in terms of Neural Networks was originally proposed by Torlai and coworkers in 2018, based on the purification of a Restricted Boltzmann Machine, but that approach was limited to shallow networks.
In this talk we will discuss the Gram-Hadamard Density Operator (GHDO), a new deep neural-network architecture that can encode positive semi-definite density operators of exponential rank with poly- nomial resources.
We will then show how to embed an autoregressive structure in the GHDO to allow direct sampling of the probability distribution.