The Rival AI Deployment Problem: a Pre-deployment Agreement as the least-bad response 2022-09-23T09:28:45.352Z
Compute & Antitrust: Regulatory implications of the AI hardware supply chain, from chip design to cloud APIs 2022-08-19T17:20:52.077Z
Effective Enforceability of EU Competition Law Under Different AI Development Scenarios: A Framework for Legal Analysis 2022-08-19T17:20:25.923Z
How Could AI Governance Go Wrong? 2022-05-26T21:29:42.403Z
Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg 2022-05-19T08:42:15.392Z
13 ideas for new Existential Risk Movies & TV Shows – what are your ideas? 2022-04-12T11:47:32.578Z
Free to attend: Cambridge Conference on Catastrophic Risk (19-21 April) 2022-03-21T13:23:09.208Z
CSER Conference 2022: Call for Lightning Talks 2021-10-13T15:30:41.078Z
Response to Phil Torres’ ‘The Case Against Longtermism’ 2021-03-08T18:09:57.419Z
Assessing Climate Change’s Contribution to Global Catastrophic Risk 2021-02-19T16:26:41.595Z
Alternatives to donor lotteries 2021-02-14T18:02:13.887Z
13 Recent Publications on Existential Risk (Jan 2021 update) 2021-02-08T12:42:17.694Z
Centre for the Study of Existential Risk Four Month Report June - September 2020 2020-12-02T18:33:42.374Z
4 Years Later: President Trump and Global Catastrophic Risk 2020-10-25T16:28:00.115Z
Centre for the Study of Existential Risk Newsletter June 2020 2020-07-02T14:03:07.303Z
11 Recent Publications on Existential Risk (June 2020 update) 2020-07-02T13:09:12.935Z
5 Recent Publications on Existential Risk (April 2020 update) 2020-04-29T09:37:40.792Z
Centre for the Study of Existential Risk Four Month Report October 2019 - January 2020 2020-04-08T13:28:13.479Z
19 Recent Publications on Existential Risk (Jan, Feb & Mar 2020 update) 2020-04-08T13:19:55.687Z
16 Recent Publications on Existential Risk (Nov & Dec 2019 update) 2020-01-15T12:07:42.000Z
The Labour leadership election: a high leverage, time-limited opportunity for impact (*1 week left to register for a vote*) 2020-01-12T21:53:25.644Z
21 Recent Publications on Existential Risk (Sep 2019 update) 2019-11-05T14:26:31.698Z
Centre for the Study of Existential Risk Six Month Report April - September 2019 2019-09-30T19:20:24.798Z
Centre for the Study of Existential Risk Six Month Report: November 2018 - April 2019 2019-05-01T15:34:20.425Z
Lecture Videos from Cambridge Conference on Catastrophic Risk 2019-04-23T16:03:21.275Z
CSER Advice to EU High-Level Expert Group on AI 2019-03-08T20:42:10.796Z
CSER and FHI advice to UN High-level Panel on Digital Cooperation 2019-03-08T20:39:29.657Z
Centre for the Study of Existential Risk: Six Month Report May-October 2018 2018-11-30T20:32:01.600Z
CSER Special Issue: 'Futures of Research in Catastrophic and Existential Risk' 2018-10-02T17:18:48.449Z
New Vacancy: Policy & AI at Cambridge University 2017-02-13T19:32:23.538Z
President Trump as a Global Catastrophic Risk 2016-11-18T18:02:46.526Z


Comment by HaydnBelfield on I'm interviewing prolific AI safety researcher Richard Ngo (now at OpenAI and previously DeepMind). What should I ask him? · 2022-09-29T11:05:31.273Z · EA · GW

I think he quit his PhD actually. So you could ask him why, and what factors people should consider when choosing to do a PhD or deciding to change while on it.


<Before that he did a PhD in the Philosophy of Machine Learning at Cambridge, on the topic of "to what extent is the development of artificial intelligence analogous to the biological and cultural evolution of human intelligence?">

Comment by HaydnBelfield on Stress Externalities More in AI Safety Pitches · 2022-09-26T22:14:21.174Z · EA · GW

I'm very pro framing this as an externality. Doesn't just help with left-leaning people, it can also be  helpful for talking to other audiences, such as those immersed in economics or antitrust/competition law.

Comment by HaydnBelfield on NASA will re-direct an asteroid tonight as a test for planetary defence (link-post) · 2022-09-26T22:07:31.281Z · EA · GW

For more on this risk, see this interesting recent book: Dark Skies: Space Expansionism, Planetary Geopolitics, and the Ends of Humanity (Jun. 2020) Daniel Deudney 

Comment by HaydnBelfield on Assessing SERI/CHERI/CERI summer program impact by surveying fellows · 2022-09-26T22:01:42.368Z · EA · GW

This is really fascinating and useful work, thanks for putting it together (and everyone who contributed)!

Comment by HaydnBelfield on EA on nuclear war and expertise · 2022-09-20T14:42:10.170Z · EA · GW

Oof this comment was a shame to read - I downvoted it. Ad hominem attack and no discussion of the content of the paper. 

Also, the paper has ten authors and got through Nature peer-review - seems a stretch to write it off as just two people's ideology.

Comment by HaydnBelfield on EA on nuclear war and expertise · 2022-09-20T14:38:54.906Z · EA · GW

Just to respond to the nuclear winter point.

I actually think the EA world has been pretty good epistemically on winter: appropriately humble and exploratory, mostly funding research to work out how big a problem it is, not basing big claims on (possibly) unsettled science. The argument for serious action on reducing nuclear risk doesn't rely on claims about nuclear winter - though nuclear winter would really underline its importance. The Rethink Priorities report you critique talks at length about the debate over winter, which is great. See also 80,000 Hours profile, which is similarly cautious/hedged. 

The EA world has been the major recent funder of research on nuclear winter: OpenPhil in 2017, 2020, perhaps Longview, and soon FLI. The research has advanced considerably since 2016. Indeed, most of the research ever published on nuclear winter has been published in the last few years, using the latest climate modelling. The most recent papers are getting published in Nature. I would disagree that theres a "reliance on papers that have a number of obvious flaws".

Comment by HaydnBelfield on Thomas Kwa's Shortform · 2022-09-18T12:04:24.535Z · EA · GW

So as I see it the main phenomenon is that there's just much more being posted on the forum. I think there's two factors behind that 1) community growth and 2) strong encouragement to post on the Forum. Eg there's lots of encouragement to post on the forum from: the undergraduate introductory/onboarding fellowships, the AGI/etc 'Fundamentals' courses, the SERI/CERI/etc Summer Fellowships, or this or this (h/t John below).

The main phenomenon is that there is a lot more posted on the forum, mostly from newer/more junior people. It could well be the case that the average quality of posts has gone down. However, I'm not so sure that the quality of the best posts has gone down, and I'm not so sure that there are fewer of the best posts every month. Nevertheless, spotting the signal from the noise has become harder. 

But then the forum serves several purposes. To take two of them: One (which is the one commenters here are most focussed on) is "signal" - producing really high-quality content - and its certainly got harder to find that. But another purpose is more instrumental - its for more junior people to demonstrate their writing/reasoning ability to potential employees. Or its to act as an incentive/endgoal for them to do some research - where the benefit is more that they see whether its a fit for them or not, but they wouldn't actually do the work if it wasn't structured towards writing something public.

So the main thing that those of us who are looking for "signal" need to do is find better/new ways to do so. The curated posts are a postive step in this direction, as are the weekly summaries and the monthly summaries.

Comment by HaydnBelfield on The Pugwash Conferences and the Anti-Ballistic Missile Treaty as a case study of Track II diplomacy · 2022-09-17T12:28:47.798Z · EA · GW

This is really great work! Very clearly structured and written, persuasively argued and (fairly) well supported by the evidence.

I’m currently doing my PhD/DPhil on the history of arms control agreements, and 1972 is one of my four case-studies. So obviously I think its really important and interesting, and that more people should know about it – and I have a lot of views on the subject! So I’ve got a few thoughts on methodology, further literature and possible extensions which I’ll share below. But they’re all to adding to what is excellent work.


Its a bit unclear to me what your claim is for the link between these Track II discussions and the ultimate outcome of the two 1972 agreements. Its not that they were sufficient (needed SALT negotiations, and even then needed Kissinger/Dobrynin backchannel). Is it that the discussions were necessary for the outcome? Or just that they contributed in a positive way? I would be interested in your view.

The limitations section is good. But I think you could have been even clearer on the limits and strengths of a ‘single N’ approach. The limits are how much this can be generalised to the entire ‘universe of cases’. However, single N also has strengths - its most useful for developing and exploring mechanisms. So I think you could frame your contribution as exploring and deepening an analysis of the mechanisms. For example, something like "Two main mechanisms are proposed in the literature, this case study provides strong evidence for mechanism 1 (conveying new conceptions/ideas) and demonstrates how it works".

On another point, I'd be concerned that if you chose this case because it was one of the most successful Track II cases you'd be ‘selecting on the dependent variable’ (apologies for the political science jargon – something like “cherrypicked for having a particular outcome”) . Can you justify your motivation and case-selection differently, for example as one of (the?) biggest and most sustained Track 2 dialogues? e.g. you say: “when the first Pugwash conference happened in 1957, there were either no, or almost no, other opportunities for Soviet and American scientists to have conversations about security policy and nuclear issues”

Further literature

Adler + Schelling are great on the US side of the story. I assume you would be familiar with them, but I don’t see them cited. If you haven’t read them, you’re in for a treat – they’re great, and largely agree with you.

If you want to go down a tangent, you might want to engage with new line of argument that many US nuclear policymakers never accepted the parity of MAD, but continued seeking advantage (Green and Long 2017;  Green 2020; Lieber and Press 2006, 2020; Long and Green 2015).

As a sidenote, I’m curious why so much of the research on the two nuclear 1972 agreements focusses on ABM. ABM is the more intellectually interesting and counterintuitive. But its not clear to me it was *more important* then the limits on offensive weapons though.

Next steps/possible extensions

My impression is your main audiences are funders (and to a lesser extent general researchers and activists) within GCR. However if you wanted to adapt it, this very plausibly could be a paper. Its already a paper length, ~8,000 words. If you wanted to go down that route, there's a few things I'd do:

  • I’d cut most of the personal best guesses (“it seems likely to me” etc).  
  • I think the notes are really great and interesting! If you incorporated some of them in the text of the piece you could deepen some of your claims in section 4, slim down the other sections.
  • Have a paragraph or two placing this piece within wider IR theoretical debates on constructivism, epistemic communities, going against systemic theories to open the 'black box' of the unitary state, etc.

If you wanted to continue this research, you could contrast this case with a similar conference and see what the difference in outcomes was; or try and draw up a list of the whole universe of cases (all major Track II dialogues).

Comment by HaydnBelfield on Future Matters #5: supervolcanoes, AI takeover, and What We Owe the Future · 2022-09-16T18:52:47.985Z · EA · GW

Hmm I strongly read it as focussed on magnitude 7. Eg In the paper they focus on magnitude 7 eruptions, and the 1/6 this century probability: "The last magnitude-7 event was in Tambora, Indonesia, in 1815." / "Given the estimated recurrence rate for a magnitude-7 event, this equates to more than US$1 billion per year." This would be corroborated by their thread, Forum post, and previous work, which emphasise 7 & 1/6.

Sorry to be annoying/pedantic about this. I'm being pernickety as I view a key thrust of their research as distinguishing 7 from 8. We can't just group magnitude 7 (1/6 chance) along with magnitude 8 and write them off as a teeny 1/14,000 chance. We need to distinguish 7 from 8, consider their severity/probability seperately, and prioritise them differently.

Comment by HaydnBelfield on Future Matters #5: supervolcanoes, AI takeover, and What We Owe the Future · 2022-09-14T17:05:21.332Z · EA · GW

Hi Pablo and Matthew, just a quick one:

"Michael Cassidy and Lara Mani warn about the risk from huge volcanic eruptions. Humanity devotes significant resources to managing risk from asteroids, and yet very little into risk from supervolcanic eruptions, despite these being substantially more likely. The absolute numbers are nonetheless low; super-eruptions are expected roughly once every 14,000 years. Interventions proposed by the authors include better monitoring of eruptions, investments in preparedness, and research into geoengineering to mitigate the climatic impacts of large eruptions or (most speculatively) into ways of intervening on volcanoes directly to prevent eruptions."

However, their Nature paper is about magnitude 7 eruptions, which may have a probability this century of 1/6, not supervolcanic eruptions (magnitude 8), which as you point out have a much lower probability. 


I think its a fascinating paper that in a prominent, rigorous and novel way applies importance/neglectedness/tractability to a comparison of two hazards:

"Over the next century, large-scale volcanic eruptions are hundreds of times more likely to occur than are asteroid and comet impacts, put together. The climatic impact of these events is comparable, yet the response is vastly different. ‘Planetary defence’ receives hundreds of millions of dollars in funding each year, and has several global agencies devoted to it. [...] By contrast, there is no coordinated action, nor large-scale investment, to mitigate the global effects of large-magnitude eruptions. This needs to change."

Comment by HaydnBelfield on Larks's Shortform · 2022-08-31T18:07:20.120Z · EA · GW

Thanks for pulling it together over so many years, and reading through a truly mindboggling array of papers & reports!

Comment by HaydnBelfield on Effective altruism is no longer the right name for the movement · 2022-08-31T16:45:49.307Z · EA · GW

Yep totally fair point, my examples were about pieces. However, note that the quote you pulled out referred to 'good work in the segments' (though this is quite a squirmy lawyerly point for me to make). Also, interestingly 2019-era Will was a bit more skeptical of xrisk - or at least wrote a piece exploring that view.

I'm a bit wary of naming specific people whose views I know personally but haven't expressed them publicly, so I'll just give some orgs who mostly work in those two segments, if you don't mind:

  • 'Long-term + EA':  the APPG for Future Generations does a lot of work here, and I'd add Tyler John's work (here & here), plausibly Beckstead's thesis.
  • 'Xrisk + EA':  my impression is that some of the more normy groups OpenPhil have funded are here, working with the EA community on xrisk topics, but not necessarily buying longtermism.
Comment by HaydnBelfield on Effective altruism is no longer the right name for the movement · 2022-08-31T14:57:04.031Z · EA · GW

I agree that Effective Altruism and the existential risk prevention movement are not the same thing. Let me use this as an opportunity to trot out my Venn diagrams again. The point is that these  communities and ideas overlap but don't necessarily imply each other - you don't have to agree to all of them because you agree with one of them, and there are good people doing good work in all the segments.


Comment by HaydnBelfield on How might we align transformative AI if it’s developed very soon? · 2022-08-30T11:14:19.092Z · EA · GW

I found this presentation of a deployment problem really concrete and useful, thanks.

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-08-28T09:11:53.233Z · EA · GW

Hi Stefan,

Thanks for this response. 

You're quite right that if this post were arguing that there is an overall pattern, it would quite clearly be inadequate. It doesn't define the universe of cases or make clear how representative these cases are of that universe, the two main studies could be criticised for selecting on the dependent variable, and its based primarily on quotes from two books.

However, I didn't set out to answer something like the research question "which is more common in 20th century history, mistakenly sprinting or mistakenly failing to sprint?" - though I think that's a very interesting question, and would like someone to look into it!

My intention for this blog post was for it to be fairly clear and memorable, aimed at a general audience - especially perhaps a machine learning researcher who doesn't know much about history. The main takeaway I wanted wasn't for people to think "this is the most common/likely outcome" but rather to add a historic example to their repertoire that they can refer to - "this was an outcome". It was supposed to be a cautionary tale, a prompt to people to think not "all sprints are wrong" but rather "wait am I in an Ellsberg situation?" - and if so to have some general, sensible recommendations and questions to ask.

My aim was to express a worry ("be careful about mistaken sprints") and illustrate that with two clear, memorable stories. There's a reasonable scenario in the next few decades that we're in a situation where we feel we need to back a sprint, prompted by concern about another group/country's sprint. If we do, and I'm not around to say "hey lets be careful about this and check we're actually in a race" then I hope these two case studies may stick in someone's mind and lead them to say "OK but lets just check, don't want to make the same mistake as Szilard and Ellsberg..."

Comment by HaydnBelfield on Supplement to "The Brussels Effect and AI: How EU AI regulation will impact the global AI market" · 2022-08-17T10:21:50.649Z · EA · GW

Congratulations! So glad this is out.

Comment by HaydnBelfield on Historical EA funding data · 2022-08-15T16:30:43.714Z · EA · GW

This is hugely useful, thanks for putting it together!

Comment by HaydnBelfield on Some concerns about policy work funding and the Long Term Future Fund · 2022-08-14T11:16:47.390Z · EA · GW

If these experts regularly have a large impact on these decisions, that's an argument for transparency about them. This is a factor that could of course be outweighed by other considerations (ability to give frank advice, confidentiality, etc). Perhaps might be worth asking them how they'd feel about being named (with no pressure attached, obviously).

Also, can one volunteer as an expert? I would - and I imagine others (just on this post, perhaps Ian and Sam?) would too.

Comment by HaydnBelfield on Some concerns about policy work funding and the Long Term Future Fund · 2022-08-14T11:13:14.662Z · EA · GW

Yep great news!

Comment by HaydnBelfield on Islands, nuclear winter, and trade disruption as a human existential risk factor · 2022-08-07T06:50:47.446Z · EA · GW

Just a quick one: this is great and groundbreaking work, thanks for doing it!

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-07-08T22:49:56.489Z · EA · GW

Thank you, and I agree on both counts.

Comment by HaydnBelfield on Linkpost: The Scientists, the Statesmen, and the Bomb · 2022-07-08T11:19:20.523Z · EA · GW

The authors' takeaway is:

"The implication of these historical outcomes is that in order to reliably affect decision-making, you must yourself be the decision-maker. Prestige, access to decision-makers, relevant expertise, and cogent reasoning are not sufficient; even with all these you are liable to be ignored. By understanding the complex workings of decision-making at the highest levels, you can improve your chances of influencing outcomes in the way you desire, but even if you understand how the game is played, you are ultimately subject to the judgment of those who wield power, and this judgment can be frustratingly capricious. Without even such an understanding, you stand little or no chance whatsoever. "

I'm sympathetic to this view, and think they're right about this case study (eg see my Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg).

Nevertheless, I think this claim is overconfident and unfounded. We can't just generalise from one case to the entire universe of cases! A more accurate assessment needs to reckon with the success of the nuclear and biological weapons arms control epistemic community in the early 1970s (such as Kissinger and Meselson) - as well as the many other examples of scientific advisers being influential.

Comment by HaydnBelfield on [Book] On Assessing the Risk of Nuclear War · 2022-07-07T21:35:43.939Z · EA · GW

This looks like absolutely fascinating, much-needed work. I particularly appreciate the variety of methodological approaches. Looking forward to reading!

Comment by HaydnBelfield on The established nuke risk field deserves more engagement · 2022-07-05T12:40:35.135Z · EA · GW


Comment by HaydnBelfield on The established nuke risk field deserves more engagement · 2022-07-04T20:43:46.451Z · EA · GW

Definitely agree! We should definitely engage more with the field. I would note there's good stuff, eg here, here, here, here.

Who critiques EA, and at what timestamp in the podcast?

Comment by HaydnBelfield on Why AGI Timeline Research/Discourse Might Be Overrated · 2022-07-04T09:12:18.875Z · EA · GW

I assume Carl is thinking of something along the lines of "try and buy most new high-end chips". See eg Sam interviewed by Rob.

Comment by HaydnBelfield on Strategic Perspectives on Long-term AI Governance: Introduction · 2022-07-04T08:31:42.562Z · EA · GW

Probably "environment-shaping", but I imagine future posts will discuss each perspective in more detail.

Comment by HaydnBelfield on A Critique of The Precipice: Chapter 6 - The Risk Landscape [Red Team Challenge] · 2022-06-27T15:53:25.588Z · EA · GW

It's really important that there is public, good-faith, well-reasoned critique of this important chapter in a central book in the field. You raise some excellent points that I'd love to see Ord (and/or others) respond to. Congratulations on your work, and thank you!

Comment by HaydnBelfield on On Deference and Yudkowsky's AI Risk Estimates · 2022-06-20T20:20:56.713Z · EA · GW

More than Philip Tetlock (author of Superforecasting)?

Does that particular quote from Yudkowsky not strike you as slightly arrogant?

Comment by HaydnBelfield on What are EA's biggest legible achievements in x-risk? · 2022-06-14T22:15:57.927Z · EA · GW

There's a whole AI ethics and safety field that would have been much smaller and less influential.

From my paper Activism by the AI Community: Analysing Recent Achievements and Future Prospects.

"2.2 Ethics and safety 

There has been sustained activism from the AI community to emphasise that AI should be developed and deployed in a safe and beneficial manner. This has involved Open Letters, AI principles, the establishment of new centres, and influencing governments. 

The Puerto Rico Conference in January 2015 was a landmark event to promote the beneficial and safe development of AI. It led to an Open Letter signed by over 8,000 people calling for the safe and beneficial development of AI, and a research agenda to that end [21]. The Asilomar Conference in January 2017 led to the Asilomar AI Principles, signed by several thousand AI researchers [23]. Over a dozen sets of principles from a range of groups followed [61]. 

The AI community has established several research groups to understand and shape the societal impact of AI. AI conferences have also expanded their work to consider the impact of AI. New groups include: 

  • OpenAI (December 2015)
  • Centre for Human-Compatible AI (August 2016) 
  • Leverhulme Centre for the Future of Intelligence (October 2016)3 
  •  DeepMind Ethics and Society (October 2017)
  • UK Government’s Centre for Data Ethics and Innovation (November 2017)"
Comment by HaydnBelfield on Things usually end slowly · 2022-06-07T22:42:00.004Z · EA · GW

Great post! Mass extinctions and historical societal collapses are important data sources - I would also suggest ecological regime shifts. My main takeaway is actually about multicausality: several ‘external’ shocks typically occur in a similar period. ‘Internal’ factors matter too - very similar shocks can affect societies very differently depending on their internal structure and leadership. When complex adaptive systems shift equilibria, several causes are normally at play.

Myself, Luke Kemp and Anders Sandberg (and many others!) have three seperate chapters touching on these topics in a forthcoming book on 'Historical Systemic Collapse' edited by Princeton's Miguel Centeno et al . Hopefully coming out this year.

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-30T12:46:26.796Z · EA · GW

Thanks for this. I'm more counselling "be careful about secrecy" rather than "don't be secret". Especially be careful about secret sprints, being told you're in  a race but can't see the secret information why, and careful about "you have to take part in this secret project".

On the capability side, the shift in AI/ML publication and release norms towards staged release (not releasing full model immediately but carefully checking for misuse potential first), structured access (through APIs) and so on has been positive, I think. 

On the risks/analysis side, MIRI have their own “nondisclosed-by-default” policy on publication. CSER and other academic research groups tend towards more of a "disclosed-by-default” policy.

Comment by HaydnBelfield on How Could AI Governance Go Wrong? · 2022-05-28T21:37:00.715Z · EA · GW

Hi both,

Yes behavioural science isn't a topic I'm super familiar with, but it seems very important!

I think most of the focus so far has been on shifting norms/behaviour at top AI labs, for example nudging Publication and Release Norms for Responsible AI.

Recommender systems are a great example of a broader concern. Another is lethal autonomous weapons, where a big focus is "meaningful human control". Automation bias is an issue even up to the nuclear level - the concern is that people will more blindly trust ML systems, and won't disbelieve them as people did in several Cold War close calls (eg Petrov not believing his computer warning of an attack). See Autonomy and machine learning at the interface of nuclear weapons, computers and people.

Jess Whittlestone's PhD was in Behavioural Science, now she's Head of AI Policy at the Centre for Long-Term Resilience.

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-28T21:28:53.178Z · EA · GW


This was very much Ellsberg's view on eg the 80,000 Hours podcast:

"And it was just a lot better for Boeing and Lockheed and Northrop Grumman and General Dynamics to go that way than not to have them, then they wouldn’t be selling the weapons. And by the way what I’ve learned just recently by books like … A guys named Kofsky wrote a book called Harry Truman And The War Scare of 1947.

Reveals that at the end of the war, Ford and GM who had made most of our bombers went back to making cars very profitably. But Boeing and Lockheed didn’t make products for the commercial market, only for commercial air except there wasn’t a big enough market to keep them from bankruptcy. They had suddenly lost their vast orders for military planes in mid 1945. The only way they could avoid bankruptcy was to sell a lot of planes to the government, military planes. But against who? Not Germany we were occupying Germany, not Japan we were occupying Japan. Who was our enemy that you needed a lot of planes against. Well Russia had been our ally during the war, but Russia had enough targets to justify, so they had to be an enemy and they had to be the enemy, and we went off from there.

I would say that having read that book and a few others I could say, I now see since my book was written nine months ago, that the Cold War was a marketing campaign for selling war planes to the government and to our allies. It was a marketing campaign for annual subsidies to the aerospace industry, and the electronics industry. And also the basis for a protection racket for Europe, that kept us as a major European power. Strictly speaking we’re not a European power. But we are in effect because we provide their protection against Russia the super enemy with nuclear weapons, and for that purpose it’s better for the Russians to have ICBM, and missiles, and H-bombs, as an enemy we can prepare against. It’s the preparations that are profitable. All wars have been very profitable for the arms manufacturers, nuclear war will not be, but preparation for it is very profitable, and therefore we have to be prepared."

Comment by HaydnBelfield on How Could AI Governance Go Wrong? · 2022-05-27T14:51:47.151Z · EA · GW

Hi, yes good question, and one that has been much discussed - here's three papers on the topic. I'm personally of the view that there shouldn't really be much conflict/contradictions - we're all pushing for the safe, beneficial and responsible development and deployment of AI, and there's lots of common ground.

Bridging near- and long-term concerns about AI 

Bridging the Gap: the case for an Incompletely Theorized Agreement on AI policy 

Reconciliation between Factions Focused on Near-Term and Long-Term Artificial Intelligence 

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-25T23:24:48.587Z · EA · GW

Apologies! LAWS = Lethal Autonomous Weapons. Have edited the text.

Comment by HaydnBelfield on Some unfun lessons I learned as a junior grantmaker · 2022-05-24T09:44:33.564Z · EA · GW

This is how I've responded to positive funding news before, seems right.

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-20T10:30:38.663Z · EA · GW

Thanks! And thanks for this link. Very moving on their sense of powerlessness.

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-20T10:28:43.748Z · EA · GW

Thanks Rohin. Yes I should perhaps have spelled this out more. I was thinking about two things - focussed on those two stages of advocacy and participation.

1. Don't just get swept up in race rhetoric and join the advocacy: "oh there's nothing we can do to prevent this, we may as well just join and be loud advocates so we have some chance to shape it". Well no, whether a sprint occurs is not just in the hands of politicians and the military, but also to a large extent in the hands of scientists. Scientists have proven crucial to advocacy for, and participation in, sprints. Don't give up your power too easily.

2. You don't have to stay if it turns out you're not actually in a race and you don't have any influence on the sprint program. There were several times in 1945 when it seems to me that scientists gave up their power too easily - over when and how the bomb was used, and what information was given to the US public. Its striking that Rotblat was the only one to resign - and he was leant on to keep his real reasons secret.

One can also see this later in 1949 and the decision to go for the thermonuclear bomb. Oppenheimer, Conant, Fermi and Bethe all strongly opposed that second 'sprint' ("It is neccessarily an evil thing considerd in any light."). They were overruled, and yet continued to actively participate in the program. The only person to leave the program (Ellsberg thinks, p.291-296) was Ellsberg's own father, a factory designer - who also kept it secret.

Exit or the threat of exit can be a powerful way to shape outcomes - I discuss this further in Activism by the AI Community. Don't give up your power too easily.

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-20T10:11:33.142Z · EA · GW

Thanks Pablo for those thoughts and the link - very interesting to read in his own words.

I completely agree that stopping a 'sprint' project is very hard - probably harder than not beginning one. The US didn't slow down on ICBMs in 1960-2 either. 

We can see some of the mechanisms by which this occurs around biological weapons programs. Nixon unilaterally ended the US one; Brezhnev increased the size of the secret Soviet one. So in the USSR there was a big political/military/industrial complex with a stake in the growth of the program and substantial lobbying power, and it shaped Soviet perceptions of 'sunk costs', precedent, doctrine, strategic need for a weapons technology, identities and norms; while in the US the oppossite occured. 

Comment by HaydnBelfield on Risks from Autonomous Weapon Systems and Military AI · 2022-05-19T14:25:13.963Z · EA · GW

I don't think its a hole at all, I think its quite reasonable to focus on major states. The private sector approach is a different one with a whole different set of actors/interventions/literature - completely makes sense that its outside the scope of this report. I was just doing classic whatabouterism, wondering about your take on a related but seperate approach.

Btw I completely agree with you about cluster munitions. 

Comment by HaydnBelfield on Risks from Autonomous Weapon Systems and Military AI · 2022-05-19T13:46:14.328Z · EA · GW

Great report! Looking forward to digging into it more. 

It definitely makes sense to focus on (major) states. However a different intervention I don't think I saw in the piece is about targeting the private sector - those actually developing the tech. E.g. Reprogramming war by Pax for Peace, a Dutch NGO. They describe the project as follows:

"This is part of the PAX project aimed at dissuading the private sector from contributing to the development of lethal autonomous weapons. These weapons pose a serious threat to international peace and security, and would violate fundamental legal and ethical principles. PAX aims to engage with the private sector to help prevent lethal autonomous weapons from becoming a reality. In a series of four reports we look into which actors could potentially be involved in the development of these weapons. Each report looks at a different group of actors, namely states, the tech sector, universities & research institutes, and arms producers. This project is aimed at creating awareness in the private sector about the concerns related to lethal autonomous weapons, and at working with private sector actors to develop guidelines and regulations to ensure their work does not contribute to the development of these weapons."

It follows fairly successful investor campaigns on e.g. cluster munitions. This project could form the basis for shareholder activism or divestment by investors, and/or wider activism by the AI community  by students, researchers, employees, etc - building on eg FLI's "we won't work on LAWS" pledge

I'd be interested in your views on that kind of approach.

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-19T13:04:19.808Z · EA · GW

Thanks for these questions! I tried to answer your first in my reply to Christian.

On your second, "delaying development" makes it sound like the natural outcome/null hypothesis is a sprint - but its remarkable how the more 'natural' outcome was to not sprint, and how much effort it took to make the US sprint.

To get initial interest at the beginning of the war required lots of advocacy from top scientists, like Einstein. Even then, the USA  didn't really do anything from 1939 until 1941, when an Australian scientist went to the USA, persuaded US scientists and promised that Britain would share all its research and resources. Britain was later cut out by the Americans, and didn't have a serious independent program for the rest of the war. Germany considered it in the early war, but decided against in 1942. During the war, neither the USSR nor Japan had serious programs (and France was collaborating with Germany). All four major states (UK, Germany, USSR, Japan) realised it would cost a huge amount in terms of money, people and scarce resources like iron, and probably not come in time to affect the course of the war.

The counterfactual is just "The US acts like the other major powers of the time and decides not to launch a sprint program that costs 0.4% of GDP during a total war, and that probably won't affect who wins the war".

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-19T12:44:03.913Z · EA · GW

Thanks for the kind words Christian - I'm looking forward to reading that report, it sounds fascinating.

I agree with your first point - I say "They were arguably right, ex ante, to advocate for and participate in a project to deter the Nazi use of nuclear weapons." Actions in 1939-42 or around 1957-1959 are defensible. However, I think this highlights 1) accurate information in 1942-3 (and 1957) would have been useful and 2) when they found out the accurate information (in 1944 and 1961) , its very interesting that it didn't stop the arms buildup.

The question of whether over, under or calibrated confidence is more common is an interesting one that I'd like someone to research. It perhaps could be usefully narrowed to WWII & postwar USA. I offered some short examples, but this could easily be a paper. There are some theoretical reasons to expect overconfidence, I'd think: such as paranoia and risk-aversion, or  political economy incentives for the military-industrial complex to overemphasise risk (to get funding). But yes, an interesting open empirical question.

Comment by HaydnBelfield on Are you really in a race? The Cautionary Tales of Szilárd and Ellsberg · 2022-05-19T09:20:18.203Z · EA · GW

woops thanks for catching - have cut

Comment by HaydnBelfield on Climate change - Problem profile · 2022-05-18T20:56:08.848Z · EA · GW

Pretty sure jackva is responding to the linked article, not just this post, as e.g. they quote footnote 25 in full.

On first point, I think that that kind of argument could be found in Jonathan B. Wiener's work on "'risk-superior moves'—better options that reduce multiple risks in concert." See e.g.

On the second point, what about climate change in India-Pakistan? e.g. an event worse than the current terrible heatwave - heat stress and agriculture/economic shock leads to migration, instability, rise in tension and accidental use of nuclear weapons. The recent modelling papers indicate that would lead to 'nuclear autumn' and probably be a global catastrophe.

Comment by HaydnBelfield on Climate change - Problem profile · 2022-05-18T20:41:45.534Z · EA · GW

Note that "humanity is doomed" is not the same as 'direct extinction', as there are many other ways for us to waste our potential.

I think its an interesting argument, but I'm unsure that we can get to a rigorous, defensible distinction between 'direct' and 'indirect' risks. I'm also unsure how this framework fits with the "risk/risk factor" framework, or the 'hazard/vulnerability/exposure' framework that's common across disaster risk reduction, business + govt planning, etc. I'd be interested in hearing more in favour of this view, and in favour of the 2 claims I picked out above.

We've talked about this before, but in general I've got such uncertainty about the state of our knowledge and the future of the world that I incline towards grouping together nuclear, bio and climate as being in roughly the same scale/importance 'tier' and then spending most of our focus seeing if any particular research strand or intervention is neglected and solvable (e.g. your work flagging something underexplored like cement).

On your food production point, as I understand it the issue is more shocks than averages. Food system shocks that can lead to "economic shocks, socio-political instability as well as starvation, migration and conflict" (from the 'causal loop diagram' paper). However, I'm not a food systems expert, I'd suggest the best people to discuss this with more are our Catherine Richards and Asaf Tzachor, authors of e.g. Future Foods For Risk-Resilient Diets.

Comment by HaydnBelfield on Climate change - Problem profile · 2022-05-18T18:37:20.808Z · EA · GW

For other readers that might be similarly confused to me - there's more in the profile on 'indirect extinction risks' and on other longrun effects on humanity's potential.

Seems a bit odd to me to just post the 'direct extinction' bit, as essentially no serious researcher argues that there is a significant chance that climate change could 'directly' (and we can debate what that means) cause extinction. However, maybe this view is more widespread amongst the general public (and therefore worth responding to)?

On 'indirect risk', I'd be interested in hearing more on these two claims:

  • "it's less important to reduce upstream issues that could be making them worse vs trying to fix them directly" (footnote 25); and
  • "our guess is that [climate change's 'indirect'] contribution to other existential risks is at most an order of magnitude higher — so something like 1 in 1,000" - which "still seems more than 10 times less likely to cause extinction than nuclear war or pandemics."

If people are interested in reading more about climate change as a contributor to GCR, here are two CSER papers from last year (and we have a big one coming out soon)

Comment by HaydnBelfield on Information security considerations for AI and the long term future · 2022-05-04T19:59:38.258Z · EA · GW

Thanks for this Jeffrey and Lennart! Very interesting, and I broadly agree. Good area for people to gain skills/expertise, and private companies should beef up their infosec to make it harder for them to be hacked and stop some adversaries.

However,  I think its worth being humble/realistic. IMO a small/medium tech company (even Big Tech themselves) are not going to be able to stop a motivated state-linked actor from the P5. Would you broadly agree?

Comment by HaydnBelfield on What to include in a guest lecture on existential risks from AI? · 2022-04-13T17:46:50.745Z · EA · GW

AGI Safety Fundamentals has the best resources and reading guides. Best short intros are the very short (500 words) intro and a slightly longer one, both from Kelsey Piper.

You might find a lecture of mine useful: