Frank Schneider

Postdoc | University of Tübingen | Methods of Machine Learning (MoML)

/profile.jpg

Hi, I’m Frank! I’m a postdoctoral researcher at the University of Tübingen in the Methods of Machine Learning group, led by Philipp Hennig. I’m also a chair of the Algorithms working group at MLCommons.

My research focuses on efficient and user-friendly training methods for machine learning. I’m particularly interested in eliminating tedious hyperparameters (e.g., learning rates, schedules) to automate neural network training and make deep learning more accessible. A key aspect of my work is designing rigorous and meaningful benchmarks for training methods, such as AlgoPerf.

Previously, I earned my PhD in Computer Science from the University of Tübingen, supervised by Philipp Hennig, as part of the IMPRS-IS (International Max Planck Research School for Intelligent Systems). Before that, I studied Simulation Technology (B.Sc., M.Sc.) at the University of Stuttgart and Industrial and Applied Mathematics (M.Sc.) at TU/e Eindhoven. My master’s thesis, supervised by Maxim Pisarenco and Michiel Hochstenbach, explored novel preconditioners for structured Toeplitz matrices. This thesis was conducted at ASML (Eindhoven), a company specializing in lithography systems for the semiconductor industry.

News

Mar 2025 📝 Our paper “Accelerating Neural Network Training: An Analysis of the AlgoPerf Competition” has been accepted at ICLR 2025! See you in Singapore!
Feb 2025 🤝 We are organizing the first AlgoPerf Workshop at Meta HQ in Menlo Park on February 11th & 12th.
Feb 2024 📝 Our paper “Kronecker-Factored Approximate Curvature for Modern Neural Network Architectures” received a spotlight presentation at NeurIPS 2023!
Jun 2023 📝 We published the first paper of our MLCommons’ Algorithms working group on arXiv titled “Benchmarking Neural Network Training Algorithms”.
Jun 2022 🎓 I succesfully defended my Ph.D. thesis with the title “Understanding Deep Learning Optimization via Benchmarking and Debugging”!

Selected Publications

2025

  1. Accelerating neural network training: An analysis of the AlgoPerf competition
    Priya Kasimbeg*Frank Schneider*, Runa Eschenhagen, and 11 more authors
    In International Conference on Learning Representations (ICLR), 2025
    We analyze the results of the inaugural AlgoPerf competition

2023

  1. Kronecker-Factored Approximate Curvature for Modern Neural Network Architectures
    Runa Eschenhagen, Alexander Immer, Richard Turner, and 2 more authors
    In Neural Information Processing Systems (NeurIPS), 2023
    We extend Kronecker-Factored Approximate Curvature to generic modern neural network architectures

2021

  1. Cockpit: A Practical Debugging Tool for the Training of Deep Neural Networks
    Frank Schneider*, Felix Dangel*, and Philipp Hennig
    In Neural Information Processing Systems (NeurIPS), 2021
    We introduce a visual and statistical debugger specifically designed for deep learning helping to understand the dynamics of neural network training
  2. Descending through a Crowded Valley - Benchmarking Deep Learning Optimizers
    Robin M. Schmidt*Frank Schneider*, and Philipp Hennig
    In International Conference on Machine Learning (ICML), 2021
    We empirically compared fifteen popular deep learning optimizers

2019

  1. DeepOBS: A Deep Learning Optimizer Benchmark Suite
    Frank Schneider, Lukas Balles, and Philipp Hennig
    In International Conference on Learning Representations (ICLR), 2019
    We provide a software package that drastically simplifies, automates, and improves the evaluation of deep learning optimizers