I will join the chorus of people who are like "oh man nobody is actually trying to, uh, solve the problem of pointing/constraining/steering/limiting vast amounts of optimization power, but instead we are spending resources on many things that are related" (monitoring the situation! finding ways to steer non-vast amounts of optimization power! measuring non-vast amounts of optimization power! building infrastructure for slightly less non-vast amounts of optimization power!). "It surely would be great if more people tried to figure out how to figure out strategies that even in concept work for vast overwhelming optimizers. Or build things that are not that but still useful enough as an alternative." This seems like the kind of event that would cause more thinking on alignment foundations to happen (not necessarily agent foundations, mind you, we might want to leave that line of retreat open).
I found this project so enticing that I considered foregoing tax benefits and donating a prospective 10% of my income to it (which would be fairly little, but what can you do).
I've interacted with Mateusz for about ten hours in total, and with plex for about 1½ hours, and found them to be good thinkers, plausibly better than me, maybe a bit more obscurantist. I feel like I can have a higher confidence in Mateusz than plex given the amount of public to-me-legible-output, but I may be misremembering as I'm shooting this comment from my hip.
I mean finally if it's another $200k to METR or to this I definitely choose this.