Manifund foxManifund
Home
Login
About
People
Categories
Newsletter
HomeAboutPeopleCategoriesLoginCreate
9

Shallow review of AI safety 2024

gleech avatar

Gavin Leech

CompleteGrant
$20,860raised

Project summary

Last year I and a collaborator summarised every live project in AI safety, tried to understand their theories of change, listed outputs, personnel, and funding amounts, and wrote an editorial.

We talked to a couple dozen researchers to check our glosses and get their views. The post was well-received (100 karma on AF, which is very rare) and is e.g. a standard intro resource at 80k. We did it pro bono (or rather, failed to obtain retroactive funding).

We want to update the review for 2024: progress, shutdowns, trends, and our takes.

What are this project's goals? How will you achieve them?

The original goal was to help new researchers orient and know their options, to help everyone understand where things stand, and to help funders see quickly what has already been funded. Simply putting all links in one place was perhaps half of the value.

This iteration: same as above but incorporating last year's feedback and seeking to get sign-off from more than 50% of those covered. Also a professionalised version suitable for policy audiences.

$8K: bare bones update (80 hours). Skim everything, reuse the taxonomy and seek correction in the comments.

$13K: much more effort on verifying details and seeking out consensus, more editorial and synthesis

$17K: section on academic and outgroup efforts. Add a glossy formal report optimised for policy people.

How will this funding be used?

Wages.

Who is on your team? What's your track record on similar projects?

Gavin and Stag did last year's version. Stephen is the source of much of the (limited) descriptive statistics about the field.

We ran this project last year, and it was well-received. Habryka: "I think overall this post did a pretty good job of a lot of different work happening in the field. I don't have a ton more to say, I just think posts like this should come out every few months, and the takes in this one overall seemed pretty good to me."

What are the most likely causes and outcomes if this project fails?

N/A

How much money have you raised in the last 12 months, and from where?

$0 so far.

Comments14Donations8Similar7
Mikolaj-Kniejski avatar

Mikolaj Kniejski

Shallow review of cost-effectivness of technical AI safety orgs

Do ACE-style cost-effectivness analysis of technical AI safety orgs.

Technical AI safetyAI governance
3
10
$0 raised
KabirKumar avatar

Kabir Kumar

AI-Plans.com

Science & technologyTechnical AI safetyAI governance
5
4
$5.37K raised
KabirKumar avatar

Kabir Kumar

AI-Plans.com

Alignment Research Platform

3
1
$0 raised
🍋

Jonas Vollmer

AI forecasting and policy research by the AI 2027 team

AI Futures Project

AI governanceForecasting
7
9
$35.6K raised
StevenK avatar

Steven Kaas

Stampy’s AI Safety Info

Creating an interactive FAQ to explain existential risk from AI to all audiences

8
3
$0 raised
zabrown avatar

Zachary Brown

Create ‘Responsible AI Investing’ recommendations for institutional investors

Four months salary to draft and promote the recommendations, helping investors advocate for specific safety and governance practices at labs and chipmakers.

1
2
$0 raised
wiserhuman avatar

Francesca Gomez

Develop technical framework for human control mechanisms for agentic AI systems

Building a technical mechanism to assess risks, evaluate safeguards, and identify control gaps in agentic AI systems, enabling verifiable human oversight.

Technical AI safetyAI governance
3
5
$10K raised