Jash Vira
Does harmful fine-tuning data cause broad misalignment only when the model already recognises the target behaviour as a norm violation?
APEX Advocacy
Strengthening the capacity of under-resourced B.I.P.O.C.-led animal rescues.
Adam Lin
344 MIT rules merged into Microsoft Agent Governance Toolkit, Cisco AI Defense, MISP, OWASP. Microsoft Copilot SWE Agent uses ATR for CVE triage.
Mohamed Mohamed Ramadan
AI infrastructure for turning natural language hardware ideas into manufacturable devices
Dev D. Goyal
Small API budget to test whether semantic reporting failures in an OpenAI-validated VCG market simulation replicate across Claude and Gemini.
Christian Lysenstøen
Building an AI research agent that can propose, test, and write up small ML findings
Pedro Bentancour Garin
Building early AI governance and safety systems focused on alignment, oversight and risk reduction before more capable AI arrives.
Emma Humphrey
$5,000 USD to bring 16 vetted academics and policy leads to NZ's first AI Safety Conference, ensuring national representation and cross-sector collaboration
Avyukt Jindal
Marcus Abramovitch
A new fund for animal welfare, AI-pilled
Leticia Prados
Designing liability, insurance and fiduciary mechanisms for frontier AI using commercial space law as a structurally precise comparative framework
Joseph A Wecker
First-principles theory of the adaptive feedback loop — one structural account governing Kalman filters, RL agents, software teams, and frontier LLMs alike.
Lucas Kempe
Make "which tool did the agent pick, and why?" an inspectable artifact instead of a vector lookup.
Salvatore Barbera
Building the missing public-mobilisation layer for AI safety in Italy and Southern Europe, starting with autonomous weapons and youth AI literacy.
Tom Bibby
Social media content across YouTube, Instagram, and TikTok to grow AI x-risk awareness and build political momentum for a global pause.
Nathan Thornhill
An ORCID-gated submission pipeline where a multi-model AI panel plus quality-control layer delivers rigorous peer review without institutional gatekeeping.
Sean Peters
An early-stage AI safety research group based in Sydney, Australia
Modeling Cooperation
Software tools and research to quantify coordination failures and inform policy decisions.
Joshua Michael Sparks
Stage 0 bridge for a focused-ultrasound research program targeting the neural architecture that maintains chronic suffering.
Fatika Umar Ibrahim
The first AI safety evaluation benchmark for Nigerian indigenous livestock systems testing whether frontier models are safe to deploy in African food systems.