Hi. I’m Louis Kirsch.
I am a PhD with Jürgen Schmidhuber at IDSIA (The Swiss AI Lab) working on meta reinforcement learning. Previously, I completed my Master of Research at University College London. My long-term research goal is to create RL agents that learn their own learning algorithm, making them truly general in the AGI sense. They should be able to design their own abstractions for prediction, planning, and learning, and continuously improve across a wide range of environments.
- 2022-2023: Research Intern at Google
- Summer 2021: Research Scientist Intern at DeepMind
- Currently PhD student with Jürgen Schmidhuber at IDSIA (The Swiss AI Lab).
- Graduated as the best student of class 2018 from University College London supervised by David Barber (MRes Computational Statistics and Machine Learning).
- Graduated as the best student of class 2017 from Hasso Plattner Institute
HPI is ranked 1st for computer science in most categories in Germany (CHE ranking 2015)
- Self-employed software developer during high school and my undergraduate studies [Project selection]
12/2022 When neural networks implement general-purpose in-context learners
Emergent in-context learning with Transformers is exciting! But what is necessary to make neural nets implement general-purpose in-context learning? 2^14 tasks, a large model + memory, and initial memorization to aid generalization.— Louis Kirsch (@LouisKirschAI) December 9, 2022
Full paper https://t.co/yyp9467WgF
7/2022 My invited talk at ICML DARL 2022 covers how we ca learn how to learn without any human-engineered meta-optimization
What would it take for algorithms to invent our future reinforcement learning algorithms?— Louis Kirsch (@LouisKirschAI) July 22, 2022
In my invited talk at @darl_icml, I take you on a journey of meta-learning general-purpose learning algorithms.
At #ICML2022 Hall G 6:10pm
Work w/ @SchmidhuberAI @DeepMind and many others pic.twitter.com/dRh8uZD7em
02/2022 Our work at DeepMind on Introducing Symmetries to Black Box Meta Reinforcement Learning will appear at AAAI 2022!
10/2021 My work on Variable Shared Meta Learning (VSML) will appear at NeurIPS 2021!
Excited to share our preprint 'Meta Learning Backpropagation And Improving It'— Louis Kirsch (@LouisKirschAI) February 26, 2021
VS-ML meta learns general purpose learning algos that start to rival backprop in their performance. This is achieved purely by weight sharing LSTMs - no grads required. https://t.co/IKSG0TVAsO
09/2021 I collaborated with some great people at DeepMind on general-purpose Meta Learning during an internship.
Our paper on Introducing Symmetries to Black Box Meta Reinforcement Learning.
How to meta-learn general-purpose RL algorithms that don't need any gradients?— Louis Kirsch (@LouisKirschAI) September 23, 2021
Black-box methods (eg RL^2) tend to overfit.
We identify symmetries in backpropagation-based meta RL and apply them to black-box meta RL.
This improves generalization.https://t.co/7JLgB8zsLH
Excited about my first invited talk @ Meta Learn NeurIPS 2020 on #MetaLearning general-purpose Learning Algorithms.— Louis Kirsch (@LouisKirschAI) December 11, 2020
Includes #MetaGenRL and new work on #VSML. Unifies learned learning rules and learning in activations. Can learn to backprop.
10/2020 I have been awarded a total of 550 thousand GPU compute hours on the Swiss National Supercomputer.
Huge thanks to CSCS for making exciting new Meta Learning research possible!
12/2019 My first work on meta-learning RL algorithms has been accepted at ICLR 2020 including a spotlight talk!
Arxiv link: Improving Generalization in Meta Reinforcement Learning using Learned Objectives
Read more about it in my blog post
A complete list can be found on Google scholar.
General-Purpose In-Context Learning by Meta-Learning Transformers [ArXiv]
Preprint. Internship project at Google with James Harrison, Jascha Sohl-Dickstein, and Luke Metz
Eliminating Meta Optimization Through Self-Referential Meta Learning [ArXiv]
Workshop paper at ICML and AutoML 2022
Introducing Symmetries to Black Box Meta Reinforcement Learning [ArXiv]
Conference paper at AAAI 2022
Internship project at DeepMind with Sebastian Flennerhag, Hado van Hasselt, Abram Friesen, Junhyuk Oh, Yutian Chen
Framework for Exploring and Understanding Multivariate Correlations
Demo track paper at ECML PKDD 2017 (Kirsch et al. 2017)
Recent blog posts
Transformers and other black-box models can exhibit in-context learning-to-learn that generalizes to significantly different datasets while undergoing multiple phase transitions in terms of their learning behavior. [Continue reading]
We investigate self-referential meta learning systems that modify themselves without the need for explicit meta optimization. [Continue reading]
subscribe via RSSFollow @LouisKirschAI