Jonas Hübotter
PhD Student at ETH Zurich. I work on Test-Time Training and Reinforcement Learning.
I am a PhD student in the Learning and Adaptive Systems Group at ETH Zurich working with Andreas Krause. My research aims to leverage foundation models for solving hard tasks through specialization and reinforcement learning.
From March 2026 – May 2026 I am a research visitor at Stanford working with Carlos Guestrin. Prior to this, I obtained a Master’s degree in Theoretical Computer Science and Machine Learning from ETH Zurich and a Bachelor’s degree in Computer Science and Mathematics from the Technical University of Munich. I am a recipient of the ETH Medal.
Always feel free to reach out to me with things you find exciting.
Contacts: jhuebotter@ethz.ch Google Scholar GitHub Linkedin
News
| Mar, 2026 | ICLR 2026 Workshops: We will present our work on self-distillation (1, 2) with oral presentations at the LLA workshop and TTU workshop, and as spotlight at the RSI workshop. We also present majority voting for code generation at the TTU workshop. |
|---|---|
| Feb, 2026 | We release three papers on self-distillation: [1] on self-distillation from demonstrations enabling continual learning, [2] on reinforcement learning via self-distillation, and [3] on online self-distillation from raw user interactions. Read more. |
| Jan, 2026 | ICLR 2026: Specialization after Generalization: Towards Understanding Test-Time Training in Foundation Models has been accepted! |
| Nov, 2025 | Gave an invited lecture on test-time training at EPFL. Slides are here. |
| Sep, 2025 | NeurIPS 2025: DISCOVER: Automated Curricula for Sparse-Reward Reinforcement Learning has been accepted! We will also present our work towards understanding the effectiveness of test-time training in foundation models with an oral presentation at the CCFM workshop. |
Selected Publications
Latest Talks
→ all talks| Apr 27, 2026 Contributed Talk | Test-Time Self-Distillation ICLR Workshop on Test-Time Updates, Rio de Janeiro |
|---|---|
| Mar 26, 2026 Interview | Interview on Self-Distillation with Idan Shenfeld Yacine Mahdid |
| February-April 2026 Invited Talk | Rethinking Post-Training via Self-Distillation 🎥 📝 ELLIS Sardine Seminars, Amazon AGI, Cohere, Engram, Goodfire, Stanford HAI |
| Jan 13, 2026 Invited Talk | Test-Time Training Agents for Deep Exploration 🎥 📝 BLISS Speaker Series, Berlin |
| Dec 7, 2025 Contributed Talk | Specialization after Generalization: Towards Understanding Test-Time Training in Foundation Models 🎥 📝 NeurIPS Workshop on Continual and Compatible Foundation Model Updates, San Diego |
Supervision
I have had the privilege of advising several BSc and MSc students during their theses and semester projects. Some of these projects have led to publications.
- Lejs Behric (MSc): Reinforcement Learning via Self-Distillation
- Tim Launer (MSc): Evaluating Training Paradigms and Consensus Mechanisms For Learning To Reason (with Marco Bagatella and Ido Hakimi, TTU@ICLR'26)
- Dennis Jüni (MSc): Meta Test-Time Training for Image Classification (with Frederike Lübeck, ICLR'26)
- Matthias Otth (MSc): Efficient Fine-Tuning and Test-Time Training of Large Language Models for Reasoning Tasks (with Ido Hakimi, SCALR@COLM'25)
- Leander Diaz-Bone (MSc): Directed Goal-Conditioned Reinforcement Learning (with Marco Bagatella, NeurIPS'25)
- Ryo Bertolissi (BSc): Test-Time Model Merging for Mixture of Local Experts (with Ido Hakimi, COLM'25)
- Nicolas Menet (MSc): Efficiently Estimating Gaussian Probability of Maximality (with Parnian Kassraie, AISTATS'25)
- Sascha Bongni (BSc): Active Fine-Tuning of Large Language Models (ICLR'25)
- Pablo Lahmann (MSc): Safe Control as Inference (with Yarden As)
- Anh Duc Nguyen (BSc): Safe Bayesian Optimization without Regret