Vanisa C Koringura
Fund for living expenses to build a two-layer inference-time safety architecture on Llama-3 8B - geometric trajectory bounding paired with activation steering.
Vlad M.
Model-agnostic R-Cycle / 4-voice / Oath-Lock protocol that reduces LLM reactivity at the orchestration layer. DOI-published. Working code. Solo founder.
Ahmed Abdelhamed
Fatika Umar Ibrahim
The first AI safety evaluation benchmark for Nigerian indigenous livestock systems testing whether frontier models are safe to deploy in African food systems.
Atmadeep Ghoshal
Requesting funding for ICML 2026 spotlight position paper on ML safety for combating intimate partner violence
Detecting the EC–EpC gap in deployed LLMs: when AI systems sound credible but misrepresent their own knowledge boundaries.
Marcus Abramovitch
A new fund for animal welfare, AI-pilled
Sankalp Gilda
Two co-authored workshop papers (LLM reasoning, agentic-AI accountability), presented April 2026 in Rio. Asking partial trip reimbursement.
Alex Kwon
If your reward model is an LLM, you cannot tell whether the policy is gaming the reward or actually getting better. We built a simulator instead.
Matthew A Cator
Funding the open-source launch of a working claim-state system and the local firewall bridge that carries verification before voice into governed agent action.
Tom Bibby
Social media content across YouTube, Instagram, and TikTok to grow AI x-risk awareness and build political momentum for a global pause.
José Wheeler
Identifying and auditing reasoning circuits in LLMs within Algoverse 2026 using Sparse Autoencoders (SAEs).
Kumari Neha Priya
Urgent funding needed by May 8 for graduate policy training focused on AI governance
Modeling Cooperation
Software tools and research to quantify coordination failures and inform policy decisions.
Developing enforceable architectural constraints, safety mechanisms, and certification criteria to keep advanced AI systems aligned and non-conscious
Sean Peters
An early-stage AI safety research group based in Sydney, Australia
Aldan Creo
Putting explainability at the forefront of AI text detection
AI Understanding
Building the first browser-based digital laboratory for interactive AI Safety education and failure-mode discovery.
Sardor Razikov
First quantitative framework for measuring when LLMs surrender independent reasoning under authority pressure
Karsten Brensing
Limited Legal Personhood as a Reversible Safety Instrument