This repository contains code for estimating the cost-effectiveness of various field-building programs. It has been built for the Center for AI Safety. This is a fork to estimate the cost effectiveness of Apart Research's programs:
- Alignment Jam hackathons
- Apart Labs program
- Apart YouTube channel (talks)
- Apart's AI safety newsletter (YouTube, newsletter, podcast)
- AI Safety Ideas
Each of these will be evaluated in turn. Earlier points have higher priority.
To get this repository working on your local machine:
- Install Python and your preferred code editor.
- Fork this repository.
- Install the repository's dependencies by executing
pip install -r requirements.txt
in your terminal.
Then, see the examples README for demonstrations of the repository's use.
If you would like assistance with this repo and/or your own evaluations, contact CAIS at [email protected].
src
: Contains source code.models
: Contains cost-effectiveness models, the main logic of this project.parameters
: Contains parameter instances for each program evaluated. (The parameters README describes what we mean by instances.)scripts
: Contains scripts for generating outputs. Organized into subdirectories forexamples
of how to use the repository, and code used to generate content for writtenposts
.utilities
: Contains functions and assumptions that are common across multiple cost-benefit analyses. Organized into subdirectories forassumptions
,defaults
,functions
,plotting
, andsampling
.
output
: Contains data and plot outputs generated fromscripts
.
The scripts
feed parameters
into models
to produce output
s. The utilities
are used at many different stages of the project -- providing functions for specifying parameters, lower-level functions for models, sampling functions for the scripts, plotting functions for the outputs, and more.
Our introduction post lays out our approach to modeling – including our motivations for using models, the benefits and limitations of our key metric, comparisons between programs for students and professionals, and more.
We have two posts evaluating student programs and professional programs respectively.
Finally, for definitions and values of the parameters used in our models, refer to the Parameter Documentation sheet.
- The Trojan Detection Challenge (or ‘TDC’): A prize at a top ML conference.
- The NeurIPS ML Safety Social (or ‘NeurIPS Social’): A social at a top ML conference.
- The NeurIPS ML Safety Workshop (or ‘NeurIPS Workshop’): A workshop at a top ML conference.
- The Atlas Fellowship: A 10-day in-person program providing a scholarship and networking opportunities for select high school students.
- ML Safety Scholars (or 'MLSS'): CAIS’s discontinued summer course, designed to teach undergraduates ML safety.
- Student Group : A high-cost, high-engagement student group at a top university, similar to HAIST, MAIA, or SAIA.
- Undergraduate Stipends: Specifically, the ML Safety Student Scholarship, which provides stipends to undergraduates connected with research opportunities.
Notice that the professional_program
and student_program
:
- Are flexible enough to accommodate a wide range of possible field-building programs, and
- Could be easily repurposed for research areas beyond AI safety.
We hope that the tools in this repository might be used or extended by other organizations. For suggestions of how to go about this, see the examples README.
This project is licensed under the MIT License.