User Tools

Site Tools


arguments_for_ai_risk:interviews_on_plausibility_of_ai_safety_by_default

Interviews on plausibility of AI safety by default

Published 02 April, 2020; last updated 15 September, 2020

This is a list of interviews on the plausibility of AI safety by default.

Background

AI Impacts conducted interviews with several thinkers on AI safety in 2019 as part of a project exploring arguments for expecting advanced AI to be safe by default. The interviews also covered other AI safety topics, such as timelines to advanced AI, the likelihood of current techniques leading to AGI, and currently promising AI safety interventions.

List

arguments_for_ai_risk/interviews_on_plausibility_of_ai_safety_by_default.txt · Last modified: 2022/09/21 07:37 (external edit)