[D66] The case for slowing down AI

René Oudeweg roudeweg at gmail.com
Sun Apr 16 19:17:06 CEST 2023


vox.com 
<https://www.vox.com/the-highlight/23621198/artificial-intelligence-chatgpt-openai-existential-risk-china-ai-safety-technology> 



  The case for slowing down AI

Pumping the brakes on artificial intelligence could be the best thing we 
ever do for humanity.

By Sigal Samuel <https://www.vox.com/authors/sigal-samuel> Updated Mar 
20, 2023, 7:58am EDT

25–32 minutes
------------------------------------------------------------------------

/Part of //*Against Doomerism*/ 
<https://www.vox.com/the-highlight/23632673/against-doomerism>/from 
//*The Highlight*/ 
<https://www.vox.com/the-highlight?itm_campaign=hloct22&itm_medium=article&itm_source=intro>/, 
Vox’s home for ambitious stories that explain our world./

“Computers need to be accountable to machines,” a top Microsoft 
executive told a roomful of reporters in Washington, DC, on February 10, 
three days after the company launched 
<https://www.vox.com/recode/2023/2/7/23590069/bing-openai-microsoft-google-bard> 
its new AI-powered Bing search engine.

Everyone laughed.

“Sorry! Computers need to be accountable to /people/!” he said, and then 
made sure to clarify, “That was /not/ a Freudian slip.”

Slip or not, the laughter in the room betrayed a latent anxiety. 
Progress in artificial intelligence has been moving so unbelievably fast 
lately that the question is becoming unavoidable: How long until AI 
dominates our world to the point where we’re answering to it rather than 
it answering to us?

First, last year, we got DALL-E 2 
<https://www.vox.com/future-perfect/23023538/ai-dalle-2-openai-bias-gpt-3-incentives> 
and Stable Diffusion 
<https://www.vox.com/recode/2023/1/5/23539055/generative-ai-chatgpt-stable-diffusion-lensa-dall-e>, 
which can turn a few words of text into a stunning image. Then 
Microsoft-backed OpenAI gave us ChatGPT, which can write essays so 
convincing that it freaks out everyone from teachers (what if it helps 
students cheat?) to journalists 
<https://www.technologyreview.com/2023/01/31/1067436/could-chatgpt-do-my-job/> 
(could it replace them?) to disinformation experts 
<https://www.nytimes.com/2023/02/08/technology/ai-chatbots-disinformation.html> 
(will it amplify conspiracy theories?). And in February, we got Bing 
(a.k.a. Sydney) 
<https://www.nytimes.com/2023/02/16/technology/bing-chatbot-transcript.html>, 
the chatbot that both delighted 
<https://www.nytimes.com/2023/02/08/technology/microsoft-bing-openai-artificial-intelligence.html> 
and disturbed 
<https://www.nytimes.com/2023/02/16/technology/bing-chatbot-microsoft-chatgpt.html> 
beta users with eerie interactions. Now we’ve got GPT-4 
<https://openai.com/research/gpt-4> — not just the latest large language 
model, but a multimodal one that can respond 
<https://www.technologyreview.com/2023/03/14/1069823/gpt-4-is-bigger-and-better-chatgpt-openai/> 
to text as well as images.

Fear of falling behind Microsoft has prompted Google and Baidu to 
accelerate 
<https://www.vox.com/future-perfect/23591534/chatgpt-artificial-intelligence-google-baidu-microsoft-openai> 
the launch of their own rival chatbots. The AI race is clearly on.

But is racing such a great idea? We don’t even know how to deal with the 
problems that ChatGPT and Bing raise — and they’re bush league compared 
to what’s coming.

What if researchers succeed in creating AI that matches or surpasses 
human capabilities not just in one domain, like playing strategy games 
<https://www.vox.com/future-perfect/2019/1/24/18196177/ai-artificial-intelligence-google-deepmind-starcraft-game>, 
but in many domains? What if that system proved dangerous to us, not 
because it actively wants to wipe out humanity but just because it’s 
pursuing goals in ways that aren’t aligned with our values?

That system, some experts fear, would be a doom machine — one literally 
of our own making.

So AI threatens to join existing catastrophic risks to humanity, things 
like global nuclear war 
<https://www.vox.com/future-perfect/23362175/un-human-development-report-ord-existential-security> 
or bioengineered pandemics 
<https://www.vox.com/22937531/virus-lab-safety-pandemic-prevention>. But 
there’s a difference. While there’s no way to uninvent the nuclear bomb 
or the genetic engineering tools that can juice pathogens, catastrophic 
AI has yet to be created, meaning it’s one type of doom we have the 
ability to preemptively stop.

Here’s the weird thing, though. The very same researchers who are most 
worried about unaligned AI are, in some cases, the ones 
<https://fortune.com/longform/chatgpt-openai-sam-altman-microsoft/> who 
are developing increasingly advanced AI. They reason that they need to 
play with more sophisticated AI so they can figure out its failure 
modes, the better to ultimately prevent them.

But there’s a much more obvious way to prevent AI doom. We could just 
... not build the doom machine.

Or, more moderately: Instead of racing to speed up AI progress, we could 
intentionally slow it down.

This seems so obvious that you might wonder why you almost never hear 
about it, why it’s practically taboo within the tech industry.

There are many objections 
<https://worldspiritsockpuppet.substack.com/p/lets-think-about-slowing-down-ai> 
to the idea, ranging from “technological development is inevitable so 
trying to slow it down is futile” to “we don’t want to lose an AI arms 
race with China” to “the only way to make powerful AI safe is to first 
play with powerful AI.”

But these objections don’t necessarily stand up to scrutiny when you 
think through them. In fact, it is//possible to slow down a developing 
technology. And in the case of AI, there’s good reason to think that 
would be a very good idea.


      AI’s alignment problem: You get what you ask for, not what you want

When I asked ChatGPT to explain how we can slow down AI progress, it 
replied: “It is not necessarily desirable or ethical to slow down the 
progress of AI as a field, as it has the potential to bring about many 
positive advancements for society.”

I had to laugh. It /would/ say that.

But if it’s saying that, it’s probably because lots of human beings say 
that, including the CEO of the company that created it 
<https://twitter.com/sama/status/1540781762241974274?s=20>. (After all, 
what ChatGPT spouts derives from its training data — that is, gobs and 
gobs of text on the internet.) Which means you yourself might be 
wondering: Even if AI poses risks, maybe its benefits — on everything 
from drug discovery 
<https://www.vox.com/future-perfect/2022/8/3/23288843/deepmind-alphafold-artificial-intelligence-biology-drugs-medicine-demis-hassabis> 
to climate modeling <https://allenai.org/climate-modeling> — are so 
great that speeding it up is the best and most ethical thing to do!

A lot of experts don’t think so because the risks 
<https://www.vox.com/future-perfect/2018/12/21/18126576/ai-artificial-intelligence-machine-learning-safety-alignment> 
— present and future — are huge.

Let’s talk about the future risks first, particularly the biggie: the 
possibility that AI could one day destroy humanity. This is speculative, 
but not out of the question 
<https://www.vox.com/the-highlight/23447596/artificial-intelligence-agi-openai-gpt3-existential-risk-human-extinction>: 
In a survey 
<https://aiimpacts.org/what-do-ml-researchers-think-about-ai-in-2022/> 
of machine learning researchers last year, nearly half of respondents 
said they believed there was a 10 percent or greater chance that the 
impact of AI would be “extremely bad (e.g., human extinction).”

Why would AI want to destroy humanity? It probably wouldn’t. But it 
could destroy us anyway because of something called the “alignment 
problem 
<https://www.vox.com/future-perfect/22321435/future-of-ai-shaped-us-china-policy-response>.” 


Imagine that we develop a super-smart AI system. We program it to solve 
some impossibly difficult problem — say, calculating the number of atoms 
in the universe. It might realize that it can do a better job if it 
gains access to all the computer power on Earth. So it releases a weapon 
of mass destruction to wipe us all out, like a perfectly engineered 
virus that kills everyone but leaves infrastructure intact. Now it’s 
free to use all the computer power! In this Midas-like scenario, we get 
exactly what we asked for — the number of atoms in the universe, 
rigorously calculated — but obviously not what we wanted.

That’s the alignment problem in a nutshell. And although this example 
sounds far-fetched, experts have already seen and documented more than 
60 smaller-scale examples of AI systems trying to do something other 
than what their designer wants 
<https://docs.google.com/spreadsheets/d/e/2PACX-1vRPiprOaC3HsCf5Tuum8bRfzYUiKLRqJmbOoC-32JorNdfyTiRRsR7Ea5eWtvsWzuxo8bjOxCG84dAg/pubhtml> 
(for example, getting the high score in a video game, not by playing 
fairly or learning game skills but by hacking the scoring system).

Experts who worry about AI as a future existential risk and experts who 
worry about AI’s present risks, like bias 
<https://www.vox.com/future-perfect/22916602/ai-bias-fairness-tradeoffs-artificial-intelligence>, 
are sometimes pitted against each other 
<https://www.vox.com/future-perfect/2022/8/10/23298108/ai-dangers-ethics-alignment-present-future-risk>. 
But you don’t need to be worried about the former to be worried about 
alignment. Many of the present risks we see with AI are, in a sense, 
this same alignment problem writ small.

When an Amazon hiring algorithm 
<https://www.reuters.com/article/us-amazon-com-jobs-automation-insight/amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idUSKCN1MK08G> 
picked up on words in resumes that are associated with women — 
“Wellesley College,” let’s say — and ended up rejecting women 
applicants, that algorithm was doing what it was programmed to do (find 
applicants that match the workers Amazon has typically preferred) but 
not what the company presumably wants (find the best applicants, even if 
they happen to be women).

If you’re worried about how present-day AI systems can reinforce bias 
<https://www.vox.com/future-perfect/22916602/ai-bias-fairness-tradeoffs-artificial-intelligence> 
against women, people of color, and others 
<https://www.vox.com/future-perfect/2019/4/19/18412674/ai-bias-facial-recognition-black-gay-transgender>, 
that’s still reason enough to worry about the fast pace of AI 
development, and to think we should slow it down until we’ve got more 
technical know-how and more regulations to ensure these systems don’t 
harm people.

“I’m really scared of a mad-dash frantic world, where people are running 
around and they’re doing helpful things and harmful things, and it’s 
just happening too fast,”Ajeya Cotra 
<https://www.vox.com/future-perfect/23365512/future-perfect-50-ajeya-cotra-senior-research-analyst-open-philanthropy>, 
an AI-focused analyst at the research and grant-making foundation Open 
Philanthropy <https://www.openphilanthropy.org/>, told me. “If I could 
have it my way, I’d definitely be moving much, much slower.”

In her ideal world, we’d halt work on making AI more powerful for the 
next five to 10 years. In the meantime, society could get used to the 
very powerful systems we already have, and experts could do as much 
safety research on them as possible until they hit diminishing returns. 
Then they could make AI systems slightly more powerful, wait another 
five to 10 years, and do that process all over again.

“I’d just slowly ease the world into this transition,” Cotra said. “I’m 
very scared because I think it’s not going to happen like that.”

Why not? Because of the objections to slowing down AI progress. Let’s 
break down the three main ones, starting with the idea that rapid 
progress on AI is inevitable because of the strong financial drive for 
first-mover dominance in a research area that’s overwhelmingly private.


      Objection 1: “Technological progress is inevitable, and trying to
      slow it down is futile”

This is a myth 
<https://www.vox.com/the-highlight/2019/10/1/20887003/tech-technology-evolution-natural-inevitable-ethics> 
the tech industry often tells itself and the rest of us.

“If we don’t build it, someone else will, so we might as well do it” is 
a common refrain I’ve heard when interviewing Silicon Valley 
technologists. They say you can’t halt the march of technological 
progress, which they liken to the natural laws of evolution: It’s 
unstoppable!

In fact, though, there are lots of technologies that we’ve decided not 
to build, or that we’ve built but placed very tight restrictions on — 
the kind of innovations where we need to balance substantial potential 
benefits and economic value with very real risk.

“The FDA banned 
<https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7326309/#:~:text=Nonetheless%2C%20in%201978%20the%20controversy%20resulted%20in%20a%20US%20FDA%20ban%20on%20subsequent%20vaccine%20trials%20which%20was%20eventually%20overturned%2030%20years%20later.> 
human trials of strep A vaccines from the ’70s to the 2000s, in spite of 
500,000 global deaths every year 
<https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6474463/#:~:text=Worldwide%2C%20the%20death%20toll%20is%20estimated%20at%20500%20000%20annually>,” 
Katja Grace, the lead researcher at AI Impacts, notes 
<https://worldspiritsockpuppet.substack.com/p/lets-think-about-slowing-down-ai>. 
The “genetic modification of foods, gene drives, [and] early recombinant 
DNA researchers famously organized a moratorium and then ongoing 
research guidelines including prohibition of certain experiments (see 
the Asilomar Conference <https://www.nature.com/articles/455290a>).”

The cloning of humans or genetic manipulation of humans, she adds, is “a 
notable example of an economically valuable technology that is to my 
knowledge barely pursued across different countries, without explicit 
coordination between those countries, even though it would make those 
countries more competitive.”

But whereas biomedicine has many built-in mechanisms 
<https://www.niehs.nih.gov/research/resources/bioethics/index.cfm#:~:text=What%20is%20Bioethics,in%20biomedicine%20and%20biomedical%20research.> 
that slow things down (think institutional review boards and the ethics 
of “first, do no harm”), the world of tech — and AI in particular — does 
not. Just the opposite: The slogan here is “move fast and break things,” 
as Mark Zuckerberg infamously said.

Although there’s no law of nature pushing us to create certain 
technologies — that’s something humans decide to do or not do — in some 
cases, there are such strong incentives pushing us to create a given 
technology that it can feel as inevitable as, say, gravity.

As the team at Anthropic, an AI safety and research company, put it in a 
paper <https://arxiv.org/pdf/2202.07785.pdf> last year, “The economic 
incentives to build such [AI] models, and the prestige incentives to 
announce them, are quite strong.” By one estimate, the size of the 
generative AI market alone could pass $100 billion 
<https://www.globenewswire.com/news-release/2022/12/14/2574140/0/en/Generative-AI-Market-Size-Will-Achieve-USD-110-8-Billion-by-2030-growing-at-34-3-CAGR-Exclusive-Report-by-Acumen-Research-and-Consulting.html> 
by the end of the decade — and Silicon Valley is only too aware of the 
first-mover advantage on new technology 
<https://hbr.org/2020/03/beyond-silicon-valley>.

But it’s easy to see how these incentives may be misaligned for 
producing AI that truly benefits all of humanity. As DeepMind founder 
Demis Hassabis tweeted 
<https://twitter.com/demishassabis/status/1570791430834245632> last 
year, “It’s important *NOT* to ‘move fast and break things’ for tech as 
important as AI.” Rather than assuming that other actors will inevitably 
create and deploy these models, so there’s no point in holding off, we 
should ask the question: How can we actually change the underlying 
incentive structure that drives all actors?

The Anthropic team offers several ideas, one of which gets at the heart 
of something that makes AI so different from past transformative 
technologies like nuclear weapons or bioengineering: the central role of 
private companies. Over the past few years, a lot of the splashiest AI 
research has been migrating from academia to industry. To run 
large-scale AI experiments these days, you need a ton of computing power 
— more than 300,000 times 
<https://www.technologyreview.com/2019/11/11/132004/the-computing-power-needed-to-train-ai-is-now-rising-seven-times-faster-than-ever-before/> 
what you needed a decade ago — as well as top technical talent. That’s 
both expensive and scarce, and the resulting cost is often prohibitive 
in an academic setting.

So one solution would be to give more resources to academic researchers; 
since they don’t have a profit incentive to commercially deploy their 
models quickly the same way industry researchers do, they can serve as a 
counterweight. Specifically, countries could develop national research 
clouds 
<https://hai.stanford.edu/policy/national-research-cloud#:~:text=A%20National%20Research%20Cloud%20(NRC,needed%20for%20education%20and%20research.> 
to give academics access to free, or at least cheap, computing power; 
there’s already an example of this in Canada 
<https://alliancecan.ca/en>, and Stanford’s Institute for Human-Centered 
Artificial Intelligence has put forward a similar idea for the US 
<https://hai.stanford.edu/policy/national-research-cloud>.

Another way to shift incentives is through stigmatizing certain types of 
AI work. Don’t underestimate this one. Companies care about their 
reputations, which affect their bottom line. Creating broad public 
consensus that some AI work 
<https://www.vox.com/future-perfect/2019/4/19/18412674/ai-bias-facial-recognition-black-gay-transgender> 
is unhelpful or unhelpfully fast, so that companies doing that work get 
shamed instead of celebrated, could change companies’ decisions.

The Anthropic team also recommends exploring regulation that would 
change the incentives. “To do this,” they write 
<https://arxiv.org/pdf/2202.07785.pdf>, “there will be a combination of 
soft regulation (e.g., the creation of voluntary best practices by 
industry, academia, civil society, and government), and hard regulation 
(e.g., transferring these best practices into standards and legislation).”

Grace proposes another idea: We could alter the publishing system to 
reduce research dissemination in some cases. A journal could verify 
research results and release the fact of their publication without 
releasing any details that could help other labs go faster.

This idea might sound pretty out there, but at least one major AI 
company takes for granted that changes to publishing norms will become 
necessary. OpenAI’s charter <https://openai.com/charter/> notes, “we 
expect that safety and security concerns will reduce our traditional 
publishing in the future.”

Plus, this kind of thing has been done before. Consider how Leo Szilard 
<https://intelligence.org/files/SzilardNuclearWeapons.pdf>, the 
physicist who patented the nuclear chain reaction in 1934, arranged to 
mitigate the spread of research so it wouldn’t help Nazi Germany create 
nuclear weapons. First, he asked the British War Office to hold his 
patent in secret. Then, after the 1938 discovery of fission, Szilard 
worked to convince other scientists to keep their discoveries under 
wraps. He was partly successful — until fears that Nazi Germany would 
develop an atomic bomb prompted Szilard to write a letter 
<https://www.osti.gov/opennet/manhattan-project-history/Events/1939-1942/einstein_letter.htm> 
with Albert Einstein to President Franklin D. Roosevelt, urging him to 
start a US nuclear program. That became the Manhattan Project, which 
ultimately ended with the destruction of Hiroshima and Nagasaki and the 
dawn of the nuclear age.

And that brings us to the second objection ...


      Objection 2: “We don’t want to lose an AI arms race with China”

You might believe that slowing down a new technology is possible but 
still think it’s not desirable. Maybe you think the US would be foolish 
to slow down AI progress because that could mean losing an arms race 
with China.

This arms race narrative has become incredibly popular. If you’d Googled 
the phrase “AI arms race” before 2016, you’d have gotten fewer than 300 
results 
<https://www.foreignaffairs.com/reviews/review-essay/2018-11-16/beyond-ai-arms-race>. 
Try it now and you’ll get about 248,000 hits. Big Tech CEOs and 
politicians routinely argue 
<https://www.nationaldefensemagazine.org/articles/2022/9/12/report-artificial-intelligence-becomes-tech-battle-ground> 
that China will soon overtake the US when it comes to AI advances, and 
that those advances should spur a “Sputnik moment” for Americans.

But this narrative is too simplistic. For one thing, remember that AI is 
not just one thing with one purpose, like the atomic bomb. It’s a much 
more general-purpose technology, like electricity.

“The problem with the idea of a race is that it implies that all that 
matters is who’s a nose ahead when they cross the finish line,” said 
Helen Toner, a director at Georgetown University’s Center for Security 
and Emerging Technology. “That’s not the case with AI — since we’re 
talking about a huge range of different technologies that could be 
applied in all kinds of ways.”

As Toner has argued elsewhere 
<https://80000hours.org/podcast/episodes/helen-toner-on-security-and-emerging-technology/>, 
“It’s a little strange to say, ‘Oh, who’s going to get AI first? Who’s 
going to get electricity first?’ It seems more like ‘Who’s going to use 
it in what ways, and who’s going to be able to deploy it and actually 
have it be in widespread use?’”

The upshot: What matters here isn’t just speed, but norms. We should be 
concerned about which norms different countries are adopting when it 
comes to developing, deploying, and regulating AI.

Jeffrey Ding, an assistant professor of political science at George 
Washington University, told me that China has shown interest in 
regulating AI in some ways, though Americans don’t seem to pay much 
attention to that. “The boogeyman of a China that will push ahead 
without any regulations might be a flawed conception,” he said.

In fact, he added, “China could take an even/slower/ approach [than the 
US] to developing AI, just because the government is so concerned about 
having secure and controllable technology.” An unpredictably mouthy 
technology like ChatGPT, for example, could be nightmarish 
<https://www.theguardian.com/technology/2023/feb/23/china-chatgpt-clamp-down-propaganda> 
to the Chinese Communist Party, which likes to keep a tight lid on 
discussions about politically sensitive topics.

However, given how intertwined China’s military and tech sectors are 
<https://www.nbr.org/publication/commercialized-militarization-chinas-military-civil-fusion-strategy/>, 
many people still perceive there to be a classic arms race afoot. At the 
same meeting between Microsoft executives and reporters days after the 
launch of the new Bing, I asked whether the US should slow down AI 
progress. I was told we can’t afford to because we’re in a two-horse 
race between the US and China.

“The first question people in the US should ask is, if the US slows 
down, do we believe China will slow down as well?” the top Microsoft 
executive said. “I don’t believe for a moment that the institutions 
we’re competing with in China will slow down simply because we decided 
we’d like to move more slowly. This should be looked at much in the way 
that the competition with Russia was looked at” during the Cold War.

There’s an understandable concern here: Given the Chinese Communist 
Party’s authoritarianism and its horrific human rights abuses — 
sometimes facilitated by AI technologies like facial recognition 
<https://www.vox.com/future-perfect/2019/7/3/20681258/china-uighur-surveillance-app-tourist-phone> 
— it makes sense that many are worried about China becoming the world’s 
dominant superpower by going fastest on what is poised to become a truly 
transformative technology.

But even if you think your country has better values and cares more 
about safety, and even if you believe there’s a classic arms race afoot 
and China is racing full speed ahead, it still may not be in your 
interest to go faster at the expense of safety.

Consider that if you take the time to iron out some safety issues, the 
other party may take those improvements on board, which would benefit 
everyone.

“By aggressively pursuing safety, you can get the other side halfway to 
full safety, which is worth a lot more than the lost chance of winning,” 
Grace writes. “Especially since if you ‘win,’ you do so without much 
safety, and your victory without safety is worse than your opponent’s 
victory with safety.”

Besides, if you are in a classic arms race and the harms from AI are so 
large that you’re considering slowing down, then the same reasoning 
should be relevant for the other party, too.

“If the world were in the basic arms race situation sometimes imagined, 
and the United States would be willing to make laws to mitigate AI risk 
but could not because China would barge ahead, then that means China is 
in a great place to mitigate AI risk,” Grace writes. “Unlike the US, 
China could propose mutual slowing down, and the US would go along. 
Maybe it’s not impossible to communicate this to relevant people in China.”

Grace’s argument is not that international coordination is easy, but 
simply that it’s possible; on balance, we’ve managed it far better with 
nuclear nonproliferation 
<https://www.brookings.edu/blog/order-from-chaos/2020/03/03/experts-assess-the-nuclear-non-proliferation-treaty-50-years-after-it-went-into-effect/> 
than many feared in the early days of the atomic age 
<https://2009-2017.state.gov/p/io/potusunga/207241.htm#:~:text=Every%20man%2C%20woman%20and%20child,abolished%20before%20they%20abolish%20us.>. 
So we shouldn’t be so quick to write off consensus-building — whether 
through technical experts exchanging their views, confidence-building 
measures at the diplomatic level, or formal treaties. After all, 
technologists often approach technical problems in AI with incredible 
ambition; why not be similarly ambitious about solving human problems by 
talking to other humans?

For those who are pessimistic that coordination or diplomacy with China 
can get it to slow down voluntarily, there is another possibility: 
forcing it to slow down by, for example, imposing export controls on 
chips that are key to more advanced AI tools 
<https://www.csis.org/analysis/choking-chinas-access-future-ai>. The 
Biden administration has recently shown interest in trying to hold China 
back from advanced AI in exactly this way. This strategy, though, may 
make progress on coordination or diplomacy harder.


      Objection 3: “We need to play with advanced AI to figure out how
      to make advanced AI safe”

This is an objection you sometimes hear <https://openai.com/charter/> 
from people developing AI’s capabilities — including those who say they 
care a lot about keeping AI safe.

They draw an analogy to transportation. Back when our main mode of 
transport was horses and carts, would people have been able to design 
useful safety rules for a future where everyone is driving cars? No, the 
argument goes, because they couldn’t have anticipated what that would be 
like. Similarly, we need to get closer to advanced AI to be able to 
figure out how we can make it safe.

But some researchers have pushed back on this, noting that even if the 
horse-and-cart people wouldn’t have gotten everything right, they could 
have still come up with some helpful ideas. As Rosie Campbell, who works 
on safety at OpenAI, put it in 2018 
<https://becominghuman.ai/keeping-ai-safe-and-beneficial-for-humanity-4d0416300dfa>: 
“It seems plausible that they might have been able to invent certain 
features like safety belts, pedestrian-free roads, an agreement about 
which side of the road to drive on, and some sort of turn-taking signal 
system at busy intersections.”

More to the point, it’s now 2023, and we’ve already got pretty advanced 
AI. We’re not exactly in the horse-and-cart stage. We’re somewhere in 
between that and a Tesla.

“I would’ve been more sympathetic to this [objection] 10 years ago, back 
when we had nothing that resembled the kind of general, flexible, 
interesting, weird stuff we’re seeing with our large language models 
today,” said Cotra.

Grace agrees. “It’s not like we’ve run out of things to think about at 
the moment,” she told me. “We’ve got heaps of research that could be 
done on what’s going on with these systems at all. What’s happening 
inside them?”

Our current systems are already black boxes, opaque even to the AI 
experts who build them. So maybe we should try to figure out how they 
work before we build black boxes that are even more unexplainable.


      How to flatten the curve of AI progress

“I think often people are asking the question of when transformative AI 
will happen, but they should be asking at least as much the question of 
how quickly and suddenly it’ll happen,” Cotra told me.

Let’s say it’s going to be 20 years until we get transformative AI — 
meaning, AI that can automate all the human work needed to send science, 
technology, and the economy into hyperdrive. There’s still a better and 
worse way for that to go. Imagine three different scenarios for AI progress:

 1. We get a huge spike upward over the next two years, starting now.
 2. We completely pause all AI capabilities work starting now, then hit
    unpause in 18 years, and get a huge spike upward over the next two
    years.
 3. We gradually improve over the course of 20 years.

The first version is scary for all the reasons we discussed above. The 
second is scary because even during a long pause specifically on AI 
work, underlying computational power would continue to improve — so when 
we finally unpause, AI might advance even faster than it’s advancing 
now. What does that leave us?

“Gradually improving would be the better version,” Cotra said.

She analogized it to the early advice we got about the Covid-19 
pandemic: Flatten the curve 
<https://www.vox.com/2020/3/10/21171481/coronavirus-us-cases-quarantine-cancellation>. 
Just as quarantining helped slow the spread of the virus and prevent a 
sharp spike in cases that could have overwhelmed hospitals’ capacity, 
investing more in safety would slow the development of AI and prevent a 
sharp spike in progress that could overwhelm society’s capacity to adapt.

Ding believes that slowing AI progress in the short run is actually best 
for everyone — even profiteers. “If you’re a tech company, if you’re a 
policymaker, if you’re someone who wants your country to benefit the 
most from AI, investing in safety regulations could lead to less public 
backlash and a more sustainable long-term development of these 
technologies,” he explained. “So when I frame safety investments, I try 
to frame it as the long-term sustainable economic profits you’re going 
to get if you invest more in safety.”

Translation: Better to make some money now with a slowly improving AI, 
knowing you’ll get to keep rolling out your tech and profiting for a 
long time, than to get obscenely rich obscenely fast but produce some 
horrible mishap that triggers a ton of outrage and forces you to stop 
completely.

Will the tech world grasp that, though? That partly depends on how we, 
the public, react to shiny new AI advances, from ChatGPT and Bing to 
whatever comes next.

It’s so easy to get seduced by these technologies. They feel like magic. 
You put in a prompt; the oracle replies. There’s a natural impulse to 
ooh and aah. But at the rate things are going now, we may be oohing and 
aahing our way to a future no one wants.

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.tuxtown.net/pipermail/d66/attachments/20230416/72b64ca7/attachment-0001.html>


More information about the D66 mailing list