This job is part of an Advanced Research + Invention Agency-funded project.
Dovetail is an agent foundations research group. We’ve recently received an ARIA grant to fund more team members over the next year. This application is for a 10-week fellowship, which may then be extended to 1 year full-time.1 We’re especially looking for people based in the UK. This application will be open until Monday 15th September 2025.
You can read more about agent foundations and our research agenda on our website.
Alfred and Alex will be leading a group of roughly 4 to 6 other people all engaged in mathematical AI safety research. We’re open to experience levels from undergraduates who are enthusiastic about math to post-docs who are looking for a way to transition into AI safety.2 (We also welcome applicants outside academia!)
Some group members might work together, while others might do solo projects. All group members will have regular one-on-one check-ins with us, and we’ll also hold regular group meetings. We’ll spend our time finding useful formal definitions of relevant concepts, formulating and proving theorems about them, and communicating our ideas and results. We’ll find and share relevant papers, host read-through meetings, and discuss ideas with external researchers.
Here are some of the basic parameters of the fellowship.
As with a lot of research in agent foundations, it’s quite difficult to concisely communicate what exactly we work on. Probably the best way to tell if you will be interested in our research problems is to read our research wiki or our write-ups on LessWrong, and then have a conversation with us about it.
All our research is purely mathematical,3 rather than experimental or empirical. None of it involves machine learning per se, but the resulting theorems should apply to ML systems.
The domains of math that we’ve been focusing on include: dynamical systems of all kinds, probability theory, information theory, algorithmic information theory, measure theory, ergodic theory. Things we’re interested in but less knowledgeable about include: singular learning theory, computational mechanics, abstract algebra, category theory, reinforcement learning theory.
Here are some more concrete examples of projects you could work on.
If there is something else you are excited to work on, we are also open to hearing other project proposals, provided that they are within the scope of our research program.
If you’re interested, fill out this application form! You’re also welcome to email us with any questions. If you are unsure whether to submit an application, we encourage you to err on the side of applying. After that, the rest of the application steps are;
After this, we should have a pretty good sense of whether we would work well together, and we’ll make a decision about whether to offer you the 10-week fellowship (or whatever else we may have negotiated). If you are successful, we expect that the time between the first interview and the job starting will be two weeks.
We’re flexible with the specifics, such as if you would prefer to align the start and end dates with an academic calendar, or if you have existing obligations on your time. ↩
To give some more detail on these bounds; we expect applicants to have mastery of some mathematical topic beyond the standard STEM classes (multivariable/vector calc, differential equations, linear algebra, and probability theory) but it doesn’t matter what, specifically. No one is an expert in agent foundations, so we’re really seeking what is sometimes called “mathematical maturity”. On the other side, if you have far more technical expertise than either of us, we may be willing to include you in the program if it seems mutually beneficial. ↩
More specifically, the desired results are mathematical. The ideas are almost all “pre-mathematical”, in that the first part will be to translate the ideas into the appropriate formalisms. ↩
A. A. Brudno, Entropy and the complexity of the trajectories of a dynamical system (1983) ↩