AI risk hub in Singapore?post by kokotajlod · 2020-10-29T11:51:49.741Z · EA · GW · 3 comments
I tentatively guess that if Singapore were to become a thriving hub for AI risk reduction, this would reduce AI risk by 16%. Moreover I think making this happen is fairly tractable and extremely neglected. In this post I sketch my reasons. I'm interested to hear what the community thinks.
My experience (and what I've been told) is that everyone generally agrees that it would be good for AI risk awareness to be raised in Asia, but conventional wisdom is that it's the job of people like Brian Tse to do that and most other people would only make things worse by trying to help. I think this is mostly right; my only disagreement is that I think the rest of us should look harder for ways to help, and be willing to sacrifice more if need be. For example, I suggested to MIRI that they move to Singapore, not because they could or should try to influence the government or anything like that, but because their presence in Singapore would make it a more attractive place for AI risk reducers (e.g. Singaporean EAs), thereby helping to create an AI risk hub there (instead of the current situation, which is brain drain from Singapore to the Bay and London).
I put my calculation of expected value at the end; for now, here are some basic facts about Singapore and the major pathways by which I expect good things from an AI risk hub there.
Thanks to Jia Yuan Loke, Vaidehi Agarwalla, and others for conversations that led to this post.
Some basic background facts about Singapore:
1. Smart, educated, english-speaking population; a tech, trade, and financial hub for Asia.
2. Cost of living lower than London but higher than Toronto. Haven’t looked into this much, just googled and found this.
3. Is already an EA hub compared to most of Asia, but has very little EA presence compared to many places in the West.
4. Singaporean government is unusually rational, in both epistemic and instrumental senses. It is a one-party state run by very smart son of Lee Kwan Yew, the man who said: “I am not following any prescription given to me by any theoretician on democracy or whatever. I work from first principles: what will get me there?”
5. Government of Singapore is widely respected throughout the world, including by China and the USA. Thousands of Chinese officials visit Singapore to learn from how they do things. Singapore has served as a bridge between East and West on multiple occasions, laying the groundwork for the ping-pong diplomacy between China and USA and hosting the talks between North Korea and USA.
6. Drugs and male-male sex are illegal in Singapore, though restrictions on the latter are poorly enforced and may be loosening.
Path to impact #1: Picks low-hanging fruit for raising awareness and recruitment
For all we know, there are loads of people in Asia who would contribute to AI risk reduction if only they found out about AI risk or had an easier way to contribute.
There may be lots of people who MIRI or some other AI risk org would want to hire, who they have trouble hiring because of immigration restrictions in the US and/or UK. (And remember, these restrictions could increase in the future!) Having a hub in Singapore--or even just one or two organizations--would provide these people with a place to work.
Path to impact #2: Singapore govt takes AI risk seriously sooner:
Importance: Singapore govt is unusually rational. Were they to become convinced that AI risk is real, they probably would do something useful to reduce it. Here are some things Singapore could do to reduce AI risk:
1. Merely publicly announcing that AI risk is real would make it much easier to convince other governments around the world to take it seriously, since Singapore’s govt is so universally respected.
2. Relatedly, there’s a history of other countries copying Singapore’s policies. It will be a lot easier for AI policy folks to convince the US govt (or the CCP) to implement policy X if something similar has already been implemented by Singapore.
3. Singapore could spearhead and organize international treaties and collaborations to reduce AI risk, given their general competence and position as respected bridge between East and West
4. Singapore could throw lots of money and talent at the problem.
5. Singapore could build AGI themselves; they spend billions on AI annually already. They could assimilate OpenAI in the process, potentially making for a better strategic situation (Singapore govt better than Microsoft+US Govt, conditional on Singapore govt convinced of AI risk)
6. Singapore could probably think of things to do that I haven’t yet thought of.
Tractability and neglectedness: From talking to Jia and Vaidehi I tentatively conclude that this is both extremely tractable and extremely neglected. There are Singaporeans concerned about AI risk, but they are mostly outside Singapore at the moment. The government has already demonstrated openness to input from some of them. If Singapore became an AI risk hub, these Singaporeans would return home, become influential, and likely (IMO) make the Singaporean government take AI risk seriously several months (and maybe several years) earlier than it otherwise would.
The calculation: (My estimate of expected value)
Here are three possible futures:
AsianTAI: Transformative AI is built first / primarily in Asia. I think there's a 20% chance of this, and I have short timelines. If I had longer timelines the probability would go up.
AsianAwarenessNeeded: AsianTAI is false, but nevertheless Asian awareness of AI risk turns out to be necessary for making the future go well, perhaps because there is a distributed slow takeoff and there needs to be worldwide coordination on AI safety standards. I say there's a 40% chance of this.
None: None of the above; TAI is created probably in the USA and what Asia thinks isn't directly relevant. I say there's a 40% chance of this.
Note that both paths to impact are beneficial in all three scenarios. However, they are obviously more beneficial in AsianAwarenessNeeded and most beneficial in AsianTAI.
My shaky unconfident guess is that having a big AI risk hub in Asia would reduce AI risk by 30% conditional on AsianTAI, 20% conditional on AsianAwarenessNeeded, and 5% conditional on None. This works out to an unconditional 16%.
(Did I mention I'm extremely unconfident in these guesses? I am. Were I to think more about this, I'd model it over time, with the intervention being "hub sooner" vs. "hub later or never" rather than hub vs. no hub. I'd also do it on a relative rather than absolute scale, e.g. "How much diminishing returns are there to reducing AI risk in the West, where loads of people are already doing it, vs. Asia?")
OK, so where in Asia should there be a hub? Not sure, but Singapore seems like a good option. It's english-speaking, which makes it easier for people from the West to contribute, and that may be important for getting the hub started. This is a major source of uncertainty for me; maybe there should be a hub in Asia but maybe Singapore isn't the best place for it. Maybe e.g. Hong Kong would be better.
Tractability: Making a hub is hard insofar as you have to compete with other hubs, and easy otherwise. (Network effects!) Currently the Bay is the biggest hub but London/Oxford/Cambridge is reasonably big too. This makes it hard. However, there are no other hubs, and in particular no hubs in Asia. This makes it easier; presumably there are quite a few people who would go to Singapore but wouldn't go to London or SF.
Neglectedness: Currently Singaporean EAs and AI risk people tend to leave Singapore and go to the west. The fact that I mentioned Brian Tse above instead of, say, an entire Asian think tank dedicated to reducing AI risk (there is none, as far as I know?) also says something about the neglectedness of this cause...
Comments sorted by top scores.