Survey on AI existential risk scenarios

post by SamClarke, Alexis Carlier, jonasschuett · 2021-06-08T17:12:29.810Z · EA · GW · 5 comments

Contents

  Summary
  Motivation
  The survey
  Key results
    There was considerable disagreement among researchers about which risk scenarios are most likely
    Researchers are uncertain about which risk scenarios are most likely
    Researchers put substantial credence on “other scenarios”
  Key takeaway
  Caveats
  Other notable results
  Full version
  Acknowledgements
None
4 comments

Cross-posted to LessWrong [LW · GW].

Summary

Motivation

It has been argued that AI could pose an existential risk. The original risk scenarios were described by Nick Bostrom and Eliezer Yudkowsky. More recently, these have [? · GW] been criticised, and a [AF · GW] number of [EA · GW] alternative scenarios [AF · GW] have [AF · GW] been [AF · GW] proposed [AF · GW]. There has been some useful work exploring these alternative scenarios, but much of this is informal. Most pieces are only presented as blog posts, with neither the detail of a book, nor the rigour of a peer-reviewed publication. For further discussion of this dynamic, see work by Ben [EA · GW] Garfinkel, Richard [AF · GW] Ngo [? · GW] and Tom Adamczewski.

The result is that it is no longer clear which AI risk scenarios experts find most plausible. We think this state of affairs is unsatisfactory for at least two reasons. First, since many of the proposed scenarios seem underdeveloped, there is room for further work analyzing them in more detail. But this is time-consuming and there are a wide range of scenarios that could be analysed, so knowing which scenarios leading experts find most plausible is useful for prioritising this work. Second, since the views of top researchers will influence the views of the broader AI safety and governance community, it is important to make the full spectrum of views more widely available. The survey is intended to be a first step in this direction.

The survey

We asked researchers to estimate the probability of five AI risk scenarios, conditional on an existential catastrophe due to AI having occurred. There was also a catch-all “other scenarios” option.

These were the five scenarios we asked about, and the descriptions we gave in the survey:

We chose these five scenarios because they have been most prominent in previous discussions about different AI risk scenarios. For more details about the survey, you can find a copy of it at this link.

Key results

There was considerable disagreement among researchers about which risk scenarios are most likely

Researchers are uncertain about which risk scenarios are most likely

The median self-reported confidence level given by respondents was 2, on a seven point Likert scale from 0 to 6, where:

Researchers put substantial credence on “other scenarios”

The “other scenarios” option had the highest median probability, at 20%. Some researchers left free-form comments describing these other scenarios. Most of them have seen no public write-up, and the others have been explored in less detail than the five scenarios we asked about.

Key takeaway

Together, these three results suggest that there is a lot of value in exploring the likelihood of different risk scenarios in more detail. This could look like:

Recent “failure stories” by Andrew Critch [AF · GW] and Paul Christiano [AF · GW] - which seem to have been well-received and appreciated - also suggest that there is value in exploring different risk scenarios in more detail. Likewise, Rohin Shah advocates for this kind of work, and AI Impacts has recently compiled a collection of stories to clarify, explore or appreciate possible future AI scenarios.

Caveats

One important caveat is the tractability of exploring the likelihood of different AI risk scenarios in more detail. The existence of considerable disagreement, despite there having been some attempts to clarify and discuss these issues, could suggest that making progress on this is difficult. However, we think there has been relatively little effort towards this kind of work so far, and that there is still a lot of low-hanging fruit.

Additionally, there were a number of limitations in the survey design, which are summarised in this document. If we were to run the survey again, we would do many things differently. Whilst we think that our main findings stand up to these limitations, we nonetheless advise taking them cautiously, and as just one piece of evidence - among many - about researchers’ views on AI risk.

Other notable results

Most of this community’s discussion about existential risk from AI focuses on scenarios involving one or more powerful, misaligned AI systems that take control of the future. This kind of concern is articulated most prominently in “Superintelligence” and “What failure looks like [AF · GW]”, corresponding to three scenarios in our survey (the “Superintelligence” scenario, part 1 and part 2 of “What failure looks like”). The median respondent’s total (conditional) probability on these three scenarios was 50%, suggesting that this kind of concern about AI risk is still prevalent, but far from the only kind of risk that researchers are concerned about today.

69% of respondents reported that they have lowered their probability estimate in the “Superintelligence” scenario (as described above)[7] since the first year they were involved in AI safety/governance. This may be because they now assign relatively higher probabilities to other risk scenarios happening first, and not necessarily because they think that fast takeoff or other premises of the “Superintelligence” scenario are less plausible than they originally did.

Full version

At this time, we are only publishing this abbreviated version of the results. We have a version of the full results that we may publish at a later date. Please contact one of us if you would like access to this, and include a sentence on why the results would be helpful or what you intend to use them for.

Acknowledgements

We would like to thank all researchers who participated in the survey. We are also grateful for valuable comments and feedback from JJ Hepburn, Richard Ngo, Ben Garfinkel, Max Daniel, Rohin Shah, Jess Whittlestone, Rafe Kennedy, Spencer Greenberg, Linda Linsefors, David Manheim, Ross Gruetzemacher, Adam Shimi, Markus Anderljung, Chris McDonald, David Kreuger, Paolo Bova, Vael Gates, Michael Aird, Lewis Hammond, Alex Holness-Tofts, Nicholas Goldowsky-Dill, the GovAI team, the AI:FAR group, and anyone else we ought to have mentioned here. This project grew out of AISC and FHI SRF. All errors are our own.


  1. We will not look at any responses from now on; this is intended just to show what questions were asked, and in case any readers are interested in thinking through their own responses. ↩︎

  2. AI existential risk scenarios are sometimes called threat models [? · GW]. ↩︎

  3. Bostrom describes many scenarios in the book “Superintelligence”. We think that this scenario is the one that most people remember from the book, but nonetheless, we think it was probably a mistake to refer to this particular scenario by this name. ↩︎

  4. Likewise, the mean responses for the five given scenarios are all between 15% and 18%, and the mean response for “other scenarios” was 25%. ↩︎

  5. Other similar results: 77% of respondents assigned ≤5% (conditional) probability to at least one scenario; 51% of respondents assigned ≤5% (conditional) probability to at least two scenarios. ↩︎

  6. For another way of interpreting this, consider that if respondents were evenly split into six completely “polarised” camps, each of which put 100% probability on one option and 0% on the others, then the mean absolute deviation for each scenario would be ~28%. ↩︎

  7. As per footnote 3, the particular scenario we are referring to here is not the only scenario described in “Superintelligence”. ↩︎

5 comments

Comments sorted by top scores.

comment by RobBensinger · 2021-06-08T19:28:43.639Z · EA(p) · GW(p)

Fascinating results! I really appreciate the level of thought and precision you all put into the survey questions.

Were there any strong correlations between which of the five scenarios respondents considered more likely?

Replies from: SamClarke
comment by SamClarke · 2021-06-11T13:56:05.650Z · EA(p) · GW(p)

Thanks Rob, interesting question. Here are the correlation coefficients between pairs of scenarios (sorted from max to min):

So it looks like there are only weak correlations between some scenarios.

It's worth bearing in mind that we asked respondents not to give an estimate for any scenarios they'd thought about for less than 1 hour. The correlations could be stronger if we didn't have this requirement.

comment by MichaelA · 2021-06-09T06:19:02.186Z · EA(p) · GW(p)

Thanks for sharing these interesting and findings and write-up! This seems like a quite useful survey to have done, for similar reasons  to why I think/hope work like Clarifying some key hypotheses in AI alignment [LW · GW], Crucial questions for longtermists [EA · GW], Database of existential risk estimates [EA · GW], and much of the prior work you cite would be useful.

People could also check out that database for a sense of how these respondents' views compare to some other quantitative estimates that have been publicly made (which is of course only a subset of all relevant views from all relevant people).

I've also now added to my database of existential risk estimates the six estimates indicated by the sentence "If you take the median response for each scenario and compare them, those (conditional) probabilities are fairly similar (between 10% and 12.5% for the five given scenarios, and 20% for “other scenarios”)", in the "Conditional existential-risk-level estimates". 

Replies from: jonasschuett
comment by jonasschuett · 2021-06-11T07:55:25.122Z · EA(p) · GW(p)

Thanks :)