Manifund foxManifund
Home
Login
About
People
Categories
Newsletter
HomeAboutPeopleCategoriesLoginCreate
0

Human Inductive Bias Project

ACX Grants 2025
UtilityHotbar avatar

Ka Wang Pang

ActiveGrant
$10,000raised
$15,000funding goal

Donate

Sign in to donate

Description of proposed project

This project is concerned with the intersection between human learning and machine learning. Based on the Free Energy Principle framework developed by Karl Friston, we want to examine what connections deep learning based AI systems have with human brains and human learning processes. This is important for alignment work since humans have many properties like trust, honesty, self-maintenance, and corrigibility, which we want future AI systems to also possess. We are also concerned with the AI safety properties of non-LLM brain-like AI models which have been proposed by different parties, and want to proactively consider what it would take to develop an ‘alignment moonshot’ based on a coherent theory of learning which applies to both humans and AI systems.

Why are you qualified to work on this?

The current branch of the project (up to end of Apr 2026) has received funding from the ARIA Mathematics for Safeguarded AI opportunity space, under the opportunity seeds program.

Other links

Project overview - https://docs.google.com/document/d/1fl7LE8AN7mLJ6uFcPuFCzatp0zCIYvjRIjQRgHPAkSE

What would you do if not funded?

This Manifund project has been set up to receive the ACX grant for 2025. Any further funding will go towards recouping living expenses accumulated in the period before the project started, or extension research after the current project branch ends in May 2026.

How much money do you need?

N/A

Supporting documents

I am worried about near-term non-LLM AI developments (LW post explaining rationale behind project.

Comments2Donations1Similar4
Jonas avatar

Jonas Kgomo

1 day ago

Ka Wang is a great multi-disciplinary renaissance man, it would be great to see where this project goes. Reading this project breakdown tells me this project has a potential net positive impact on the general space of human compatible AI reasoning. I am writing without any conflict of interest, this is a comment to support any prospective supporter here.

donated $10,000
acx-grants avatar

ACX Grants

about 1 month ago

Grant from ACX Grants 2025