Wholeheartedly agree, and I think it's great that you're doing this.
I'll be very interested in what you learn along the way w.r.t. more/less effective processes.
(Bonus points for referencing the art of game design - one of my favourite books.)
Thanks! Yes, this is very much an experiment, and even if it fails, I expect it to be a productive mistake we can learn from. ;)
I’m opening an incubator called Refine for conceptual alignment research in London, which will be hosted by Conjecture. The program is a three-month fully-paid fellowship for helping aspiring independent researchers find, formulate, and get funding for new conceptual alignment research bets, ideas that are promising enough to try out for a few months to see if they have more potential.
If this sounds like something you’d be interested in, you can apply here!
Why?
I see a gaping hole in the alignment training ecosystem: there are no programs dedicated specifically to creating new independent conceptual researchers and helping them build original research agendas.
The programs that do exist (AI Safety Camp, SERI MATS) tend to focus on an apprenticeship (or “accelerated PhD”) model in which participants work under researchers on already-established research directions. And while there are avenues for independent alignment researchers to get started on their own, it is fraught with many risks, slowing down progress considerably.
So I feel the need for a program geared specifically towards conceptual alignment researchers that are interested in doing their own research and making their own research bets.
Who?
This program is for self-motivated and curious people who want to become independent conceptual alignment researchers and expand the portfolio of alignment bets and research ideas available.
When I look at great conceptual researchers like John Wentworth, Paul Christiano, Evan Hubinger, Steve Byrnes, Vanessa Kosoy, and others, as well as at the good (famous and not) researchers I know from my PhD, they all have the same thing in common: they ask a question and keep looking for the answer. They tolerate confusion, not in the sense that they accept it, but in that they are able to work with it and not hide away behind premature formalization. They don’t give up on the problem; they search for different angles and approaches until it yields. Paul Graham calls this being relentlessly resourceful.
(Relentlessly Resourceful, Paul Graham, 2009)
This is one of the main traits I’m looking for in an applicant — someone who will lead a new research agenda and morph it proactively, as needed.
Another point that matters is being curious about different topics and ideas than the ones traditionally discussed in alignment. As I wrote in a recent post and plan to discuss more in an upcoming sequence, I think we need to be more pluralist in our approach to alignment, and explore far more directions, from novel ideas to old approaches that may have been discarded too soon. And new ideas often come from unexpected places.
As one example, here is what Jesse Schell writes about his experience speaking to a professional juggler who performed tricks no one else could do:
(The Art of Game Design, Jesse Schell, 2008)
As for previous experience with alignment research, it can both be a blessing and a curse. While familiarity with alignment concepts can help bootstrap the learning and idea generation process, it also risks clogging the babble process by constraining “what makes sense”. For those it would be helpful for, the program includes some initial teaching on core alignment ideas (according to me) and the mental moves necessary for good alignment research.
Some concrete details
We plan to invite the first cohort of 4-5 fellows from July/August through September/October (wiggle room depending on some ops details), though exact dates will be determined by their availability. We anticipate that other cohorts will follow, so if you miss the first round but are still interested, please apply.
This is a full-time position in London where fellows will work out of Conjecture’s offices. The program includes:
During the first month of the program, participants will spend their time discussing abstract models of alignment, what the problem is about, and the different research approaches that have been pursued. The focus will be on understanding the assumptions and constraints behind the different takes and research programs, to get a high-level map of the field.
The next ~two months of the program will focus on helping fellows babble new research bets on alignment, refine them, test them, and either throw them away or change them. By the end, the goal is for fellows to narrow in on a research bet that could be further investigated in the following 6 months, and is promising enough to warrant funding.
It’s worth noting that while the incubator is being housed by Conjecture, fellows do not have any constraints imposed by the company. Fellows will not have to work on Conjecture’s research agendas or be obligated to collaborate after the program is over. Similarly, I’m not looking for people to work on my own research ideas, but for new exciting research bets I wouldn’t have thought about.
How can I apply?
We will review applications on a rolling-basis, with a usual delay of 1 week before response and a month before a decision (with a work task in the middle). The application is open now!