Manifund foxManifund
Home
Login
About
People
Categories
Newsletter
HomeAboutPeopleCategoriesLoginCreate
1

Modeling Automated AI R&D

Technical AI safety
paulsalmon avatar

Paul Salmon

Not fundedGrant
$0raised

Project summary

At some point programmers will be able to spin up a fleet of thousands of world class automated researchers—the resulting progress accomplished by this researcher cohort could be immense. Future poorly controlled self-improvement is probably the biggest issue in AI development and has the least amount of work on thinking about how to control it. This research project will use agent-based models to model the process of automated AI R&D / uncontrolled self-improvement (USI). The goal is to better understand potential runaway AI R&D processes and how to make these processes more controllable. Paul Salmon is a highly-cited expert on complex systems, human factors, and sociotechnical systems. He has also previously published work on how human factors can be used to manage the risks associated with AGI. Basically for this problem we don’t need extreme ML knowledge; instead we need knowledge of how to model systems and their dynamics and ways they can go awry.

What are this project's goals and how they be achieved?

The goal is to better understand potential runaway AI R&D processes and how to make them more controllable. It will create basic models of the dynamics and interactions between interventions that attempt to make this process more controllable.

How will this funding be used?

This funding will be used to pay graduate students who will help with the modeling.

Who is on the team and what's their track record on similar projects?

Paul Salmon is a professor at the University of the Sunshine Coast. He is a highly-cited expert on complex systems, human factors, and sociotechnical systems (google scholar). He has become interested in AGI risk and has previously published work on how human factors can be used to manage the risks associated with AGI.

What are the most likely causes and outcomes if this project fails? (premortem)

There is a chance that modeling runaway AI R&D processes requires a different skillset or is more complicated/intractable than initially expected. If that happens, the model may produce less interesting results or explain less than one might hope.

What other funding is this person or project getting?

None that I am aware of.

Comments2Similar7
🐸

SaferAI

General support for SaferAI

Support for SaferAI’s technical and governance research and education programs to enable responsible and safe AI.

AI governance
3
1
$100K raised
wiserhuman avatar

Francesca Gomez

Develop technical framework for human control mechanisms for agentic AI systems

Building a technical mechanism to assess risks, evaluate safeguards, and identify control gaps in agentic AI systems, enabling verifiable human oversight.

Technical AI safetyAI governance
3
5
$10K raised
🐶

Alexander Pan

Removing Hazardous Knowledge from AIs

Technical AI safety
1
2
$190K raised
🦁

Peter Salib

Funding a Consortium of Legal Scholars and Academics doing AI Safety

AI governance
1
2
$0 raised
peterwildeford avatar

Peter Wildeford

AI Policy work @ IAPS

AI governance
8
3
$10.1K raised
hendrycks avatar

Dan Hendrycks

Research Staff for AI Safety Research Projects

Technical AI safetyBiosecurity
10
5
$26.7K raised
🐯

Scott Viteri

Attention-Guided-RL for Human-Like LMs

Compute Funding

Technical AI safety
4
2
$3.1K raised