Manifund foxManifund
Home
Login
About
People
Categories
Newsletter
HomeAboutPeopleCategoriesLoginCreate
1

TransformerLens - Bridge Funding

Science & technologyTechnical AI safety
🐧

Bryce Meyer

ActiveGrant
$13,000raised
$30,000funding goal

Donate

Sign in to donate

Problem:

There is a currently a lot of work that goes into adding new models/architectures to TransformerLens. It requires a lot of additional reimplementations and verifications.

Solution:

The project ("TransformerBridge") will allow loading any nn.Module, including the current transformers/HuggingFace models into TransformerLens in a simple way. People will be able to easily use any architectures with TransformerLens, regardless of whether these models exist on HuggingFace by configuring one file. This project is not only designed to support and enhance all current TransformerLens usages, but it also opens the door for interpretability research in closed environments where HuggingFace hosted models may not be the target.

Timeline:

The proof of concept of this is done already, but it will take more time to complete, polish, and test, so that it can be rolled out for real world interpretability research. The next two months will allow us to enter beta, and begin helping people transition to using the new module as opposed to the existing HookedRootModules. Once all of the reasonable use cases of TransformerLens have been tested, we will release this new module into TransformerLens 3.0.

Funding:

  • $10,000 USD for Bryce to work on this for the next 2 months

  • $3,000 USD for Fabian, Bryce's mentee, who has already been making great contributions to TransformerLens over the past year

  • Any additional funds will be used to continue support for TransformerLens. If the full funding goal of $30,000 is met, that will be enough for Bryce to manage TransformerLens through the rest of 2025 with no issue.

Comments1Donations1Similar7
robertzk avatar

Robert Krzyzanowski

Scaling Training Process Transparency

Compute and infrastructure costs

Technical AI safety
3
4
$5.15K raised
kacloud avatar

Alex Cloud

Compute for 4 MATS scholars to rapidly scale promising new method pre-ICLR

Technical AI safety
3
5
$16K raised
redJ avatar

Jared Mantell

Augmentation Lab 2025: Prototyping Human-Aligned Futures

A 10-week Harvard/MIT residency exploring human augmentation via 'Rhizome Futurism' to build an interconnected, beneficial future.

Science & technology
1
0
$0 raised
jesse_hoogland avatar

Jesse Hoogland

Scoping Developmental Interpretability

6-month funding for a team of researchers to assess a novel AI alignment research agenda that studies how structure forms in neural networks

Technical AI safety
13
11
$145K raised
hijohnnylin avatar

Johnny Lin

Neuronpedia - Open Interpretability Platform

Platform for interpretability researchers, especially those creating/using Sparse Autoencoders

Technical AI safety
3
9
$2.5K raised
MatthewClarke avatar

Matthew A. Clarke

Salaries for SAE Co-occurrence Project

Working title - “Compositionality and Ambiguity: Latent Co-occurrence and Interpretable Subspaces”

Science & technologyTechnical AI safety
3
1
$0 raised
🥑

Apollo Research

Apollo Research: Scale up interpretability & behavioral model evals research

Hire 3 additional AI safety research engineers / scientists

Technical AI safety
11
12
$339K raised