Charbel-Raphael Segerie
Glen M. Taggart
And other projects aimed at producing pareto improvements in SAEs (at higher funding levels)
Louis S. Berman
AI-Risk Education for Politicians
Karpov
I plan to investigate what realistic RL training conditions might lead to LLMs developing steganographic capabilities.
Remmelt Ellen
Kabir Kumar
Apart Research
Incubate AI safety research and develop the next generation of global AI safety talent via research sprints and research fellowships
Alexander Pan
Kunvar Thaman
Zhonghao He
Surveying neuroscience for tools to analyze and understand neural networks and building a natural science of deep learning
Ryan Kidd
Help us support more research scholars!
Dusan D Nesic
Free/Subsidized/Cheap office space outside of EU but in good timezones with favorable visa policies (especially for Chinese/Russian but also other citizens).
Lawrence Chan
3 month
Luan Rafael Marques de Oliveira
Support to translate BlueDot Impact’s AI alignment curriculum into (Br) Portuguese to be used in university study groups and an online course
Cadenza Labs
We're a team of SERI-MATS alumni working on interpretability, seeking funding to continue our research after our LTFF grant ended.
Robert Krzyzanowski
Compute and infrastructure costs
Apollo Research
Hire 3 additional AI safety research engineers / scientists
Jesse Hoogland
6-month funding for a team of researchers to assess a novel AI alignment research agenda that studies how structure forms in neural networks
Jaeson Booker
Skilling up on interpretability and multi-agent alignment
Allison Duettmann
focused on 1. bci and wbe for safe ai, 2. cryptography and security for safe ai, and 3. safe multipolar ai