Will Harvey

I am a fifth-year PhD student at the University of British Columbia, and member of the Programming Languages for Artificial Intelligence group led by Frank Wood. Previously I studied for an MEng in Engineering Science at the University of Oxford.

The broad theme of my research is the intersection of deep learning and Bayesian inference. Recently I've focused on the diffusion modelling framework, with applications both in the video domain and for problems with more explicit structural information.

Email  /  Google Scholar  /  CV

profile photo
Refereed Publications
Flexible Diffusion Modeling of Long Videos
William Harvey, , , ,
NeurIPS 2022

We present a diffusion model for video. Given a few frames on context it can generate videos over an hour long.

Conditional Image Generation by Conditioning Variational Auto-Encoders
William Harvey, ,
ICLR 2022

We present conditional VAEs which can be quickly trained by leveraging existing unconditional VAEs. The resulting models provide a more faithful representation of uncertainty than GAN-based approaches with similar training times.

Planning as Inference in Epidemiological Models
, , , , , William Harvey, , , , ,
Frontiers in Artificial Intelligence | Medicine and Public Health 2022

We demonstrate how existing software tools can be used to automate parts of infectious disease-control policy-making via performing inference in existing epidemiological dynamics models.

Our new architecture for amortized inference.

Attention for Inference Compilation
William Harvey*, *, , ,

We present a transformer-based architecture for improved amortized inference in probabilistic programs with complex and stochastic control flow.

Near-Optimal Glimpse Sequences for Improved Hard Attention Neural Network Training
William Harvey, ,
IJCNN 2022

Bayesian experimental design can be used to find near-optimal attention locations for a hard attention mechanism. These can be used to speed up the later training of hard attention mechanisms.

Assisting the Adversary to Improve GAN Training
, William Harvey,
IJCNN 2021

We improve image quality by training a GAN generator in a way that accounts for a sub-optimal discriminator.

Structured Conditional Continuous Normalizing Flows for Efficient Amortized Inference in Graphical Models
, , William Harvey,

We use knowledge about the structure of a generative model to automatically select a good normalizing flow architecture.

End-to-end Training of Differentiable Pipelines Across Machine Learning Frameworks
, , , William Harvey, , ,
NIPS Autodiff Workshop 2017

We present an interface for gradient-based training of pipelines of machine learning primitives. This allows joint training of machine learning modules written in different languages, making it useful for automated machine learning (AutoML).

Graphically Structured Diffusion Models
, William Harvey,

We present a framework for constructing diffusion models that incorporate problem structure. Experiments on problems including matrix factorisation suggest an improved scaling relationship between problem dimension and model performance.

Website source forked from Jon Barron.