Andreas Kirsch

Bio & research interests

Department of Computer Science

University of Oxford

Wolfson Building

Oxford, OX1 3QD

I obtained my a PhD (“DPhil”) with Prof Yarin Gal in the OATML group at the University of Oxford and as a student in the AIMS CDT program.

During my DPhil, my interests were in information theory and its applications: information bottlenecks and active learning using Bayesian deep learning, and uncertainty quantification. I also enjoyed thinking about AI ethics and AI safety: in particular, the ML safety course by the Center of AI Safety was a lot of fun.

Originally from Romania, I grew up in Southern Germany. After studying Computer Science and Mathematics at the Technical University in Munich (i.a. reading machine learning under Jürgen Schmidhuber 🎉), I spent a couple of years in Zurich as a software engineer at Google (YouTube Monetization) and worked as a performance research engineer at DeepMind for a year in 2016/17 before spending a gap year as a fellow at Newspeak House. I began my DPhil in September 2018 and submitted my thesis in April 2023.

selected publications

  1. CVPR 2023
    Highlight
    Deterministic Neural Networks with Appropriate Inductive Biases Capture Epistemic and Aleatoric Uncertainty
    Conference on Computer Vision and Pattern Recognition 2023
  2. Preprint
    Black-Box Batch Active Learning for Regression
    Kirsch, Andreas
    arXiv 2023
  3. AISTATS 2023
    Prediction-Oriented Bayesian Active Learning
    Bickford Smith*, Freddie, Kirsch*, Andreas, Farquhar, Sebastian, Gal, Yarin, Foster, Adam, and Rainforth, Tom
    26th International Conference on Artificial Intelligence and Statistics 2023
  4. TMLR
    A Note on ”Assessing Generalization of SGD via Disagreement”
    Kirsch, Andreas, and Gal, Yarin
    Transactions on Machine Learning Research 2022
  5. TMLR
    Unifying Approaches in Active Learning and Active Sampling via Fisher Information and Information-Theoretic Quantities
    Kirsch, Andreas, and Gal, Yarin
    Transactions on Machine Learning Research 2022
  6. ICML 2022
    Prioritized Training on Points that are Learnable, Worth Learning, and not yet Learnt
    Mindermann*, Sören, Brauner*, Jan M, Razzak*, Muhammed T, Sharma*, Mrinank, Kirsch, Andreas, Xu, Winnie, Höltgen, Benedikt, Gomez, Aidan N, Morisot, Adrien, Farquhar, Sebastian, and Gal, Yarin
    In Proceedings of the 39th International Conference on Machine Learning 2022
  7. NeurIPS 2021
    Causal-BALD: Deep Bayesian Active Learning of Outcomes to Infer Treatment-Effects from Observational Data
    Jesson, Andrew, Tigas, Panagiotis, Amersfoort, Joost, Kirsch, Andreas, Shalit, Uri, and Gal, Yarin
    In Advances in Neural Information Processing Systems 2021
  8. UDL 2020
    Learning CIFAR-10 with a Simple Entropy Estimator Using Information Bottleneck Objectives
    Kirsch, Andreas, Lyle, Clare, and Gal, Yarin
    In Uncertainty & Robustness in Deep Learning at Int. Conf. on Machine Learning (ICML Workshop) 2020
  9. Preprint
    Unpacking Information Bottlenecks: Unifying Information-Theoretic Objectives in Deep Learning
    Kirsch, Andreas, Lyle, Clare, and Gal, Yarin
    arXiv Preprint 2020
  10. NeurIPS 2019
    BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning
    Kirsch*, Andreas, van Amersfoort*, Joost, and Gal, Yarin
    NeurIPS 2019

news

Jul 24, 2021

Seven workshop papers at ICML 2021 (out of which five are first author submissions):

Uncertainty & Robustness in Deep Learning

Two papers and posters at the Uncertainty & Robustness in Deep Learning workshop:

SubSetML: Subset Selection in Machine Learning: From Theory to Practice

Four papers (posters, one spotlight) at the SubSetML: Subset Selection in Machine Learning: From Theory to Practice workshop:

Neglected Assumptions In Causal Inference

One paper (poster) at the Neglected Assumptions In Causal Inference workshop:

Feb 23, 2021

Lecture on “Bayesian Deep Learning, Information Theory and Active Learning” for Oxford Global Exchanges. You can download the slides here.

Feb 21, 2021

Deterministic Neural Networks with Appropriate Inductive Biases Capture Epistemic and Aleatoric Uncertainty has been uploaded to arXiv as pre-print. Joint work with Jishnu Mukhoti, and together with Joost van Amersfoort, Philip H.S. Torr, Yarin Gal. We show that a single softmax neural net with minimal changes can beat the uncertainty predictions of Deep Ensembles and other more complex single-forward-pass uncertainty approaches.

Dec 10, 2020

Unpacking Information Bottlenecks: Unifying Information-Theoretic Objectives in Deep Learning was also presented as a poster at the “NeurIPS Europe meetup on Bayesian Deep Learning”.

You can find the poster below (click to open):

Image version

or as PDF version to download.

Jul 17, 2020

Two workshop papers have been accepted to Uncertainty & Robustness in Deep Learning Workshop at ICML 2020:

  1. Scalable Training with Information Bottleneck Objectives, and
  2. Learning CIFAR-10 with a Simple Entropy Estimator Using Information Bottleneck Objectives

both together with Clare Lyle and Yarin Gal. They are based on Unpacking Information Bottlenecks: Unifying Information-Theoretic Objectives in Deep Learning for the former, and an application of the UIB framework for the latter: we can use it to train models that perform well on CIFAR-10 without using a cross-entropy loss at all.

Mar 27, 2020

Unpacking Information Bottlenecks: Unifying Information-Theoretic Objectives in Deep Learning, together with Clare Lyle and Yarin Gal, has been uploaded as pre-print to arXiv. It examines and unifies different Information Bottleneck objectives and shows that we can introduce simple yet effective surrogate objectives without complex derivations.

Sep 4, 2019

BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning got accepted into NeurIPS 2019. See you all in Vancouver!

Jun 24, 2019

BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning has been uploaded to arXiv, and we have also published an extensive blog post about it on OATML with the code available on GitHub.

Follow me on Twitter @blackhc