I am a postdoc in the Julia Lab at the Massachusetts Institute of Technology (MIT). Previously: PhD in physics in the Bruder group within the “Quantum Computing and Quantum Technology” PhD school at the University of Basel.
Automatic differentiation (AD) is great: use gradients to optimize, sample faster, or just for fun! But what about coin flips? Agent-based models? Nope, these aren’t differentiable... or are they? StochasticAD.jl is an open-source research package for AD of stochastic programs, implementing AD algorithms for handling programs that can contain discrete randomness.
In this talk, we present continuous-adjoint sensitivity methods for hybrid differential equations (i.e., ordinary or stochastic differential equations with callbacks) modeling explicit and implicit events. The methods are implemented in the SciMLSensitivity.jl package. As a concrete example, we consider the sensitivity analysis of dosing times in pharmacokinetic models. We discuss different options for the automatic differentiation backend.
Feedback control policies for quantum systems often lack performance targets and certificates of optimality. Here, we will show how bounds on the best possible control performance are readily computable for a wide range of quantum control problems by means of convex optimization using Julia's optimization ecosystem. We discuss how these bounds provide targets and certificates to improve the design of quantum feedback controllers.