Manifund foxManifund
Home
Login
About
People
Categories
Newsletter
HomeAboutPeopleCategoriesLoginCreate
0

Human Inductive Bias Project

ACX Grants 2025
UtilityHotbar avatar

Ka Wang Pang

ActiveGrant
$10,000raised
$15,000funding goal

Donate

Sign in to donate

Description of proposed project

This project is concerned with the intersection between human learning and machine learning. Based on the Free Energy Principle framework developed by Karl Friston, we want to examine what connections deep learning based AI systems have with human brains and human learning processes. This is important for alignment work since humans have many properties like trust, honesty, self-maintenance, and corrigibility, which we want future AI systems to also possess. We are also concerned with the AI safety properties of non-LLM brain-like AI models which have been proposed by different parties, and want to proactively consider what it would take to develop an ‘alignment moonshot’ based on a coherent theory of learning which applies to both humans and AI systems.

Why are you qualified to work on this?

The current branch of the project (up to end of Apr 2026) has received funding from the ARIA Mathematics for Safeguarded AI opportunity space, under the opportunity seeds program.

Other links

Project overview - https://docs.google.com/document/d/1fl7LE8AN7mLJ6uFcPuFCzatp0zCIYvjRIjQRgHPAkSE

What would you do if not funded?

This Manifund project has been set up to receive the ACX grant for 2025. Any further funding will go towards recouping living expenses accumulated in the period before the project started, or extension research after the current project branch ends in May 2026.

How much money do you need?

N/A

Supporting documents

I am worried about near-term non-LLM AI developments (LW post explaining rationale behind project.

Comments2Donations1Similar4
charbel-raphael avatar

Charbel-Raphael Segerie

Investigating constructability as a safer approach to machine-learning

Science & technologyTechnical AI safety
7
6
$0 raised
is-sky-a-sea avatar

Aditya Raj

6-month research funding to challenge current AI safety methods

Current LLM safety methods—treat harmful knowledge as removable chunks. This is controlling a model and it does not work.

Technical AI safetyGlobal catastrophic risks
2
0
$0 raised
LawrenceC avatar

Lawrence Chan

Exploring novel research directions in prosaic AI alignment

3 month

Technical AI safety
5
9
$30K raised
🐯

Scott Viteri

Attention-Guided-RL for Human-Like LMs

Compute Funding

Technical AI safety
4
2
$3.1K raised