Interviews on the strength of the evidence for AI risk claims
This is a collection of interviews with AI risk researchers on the strength of the evidence for AI risk.
Notes on interviewee selection
We contacted AI researchers we knew (or knew of) at some prominent labs and AI safety organizations and in academia
All of the people we interviewed are concerned about AI risk and spend some or all of their work time working to reduce it
On the one hand, this means that they are expert in the topic we’re interested in (the evidence for AI risk claims)
On the other hand, it also means that they have an incentive to interpret the evidence as stronger rather than weaker
The information we’ve gathered from the interview series should therefore be taken as evidence for AI risk claims, rather than an unbiased review of the evidence on AI risk claims overall
The questions we asked
The core questions we asked were:
Can you introduce yourself and the work you do?
What’s your probability that AI causes human extinction by 2100?
What evidence do you find most convincing for existential risk from AI?
How would you summarize the state of the evidence for AI risk overall?
Most of the interview time was spent on question 3. We sought to ask follow-up questions covering misalignment (including specification gaming, goal misgeneralization and deceptive alignment) and power-seeking (including self-preservation and self-improvement).
A list of the interviews
Note that these interviews represent researchers’ personal views and not the views of their employers.
(We are still waiting for permission to publish more interview summaries.)