The Nonlinear Library: EA Forum
Education
EA - What is the current most representative EA AI x-risk argument? by Matthew Barnett
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: What is the current most representative EA AI x-risk argument?, published by Matthew Barnett on December 16, 2023 on The Effective Altruism Forum.I tend to disagree with most EAs about existential risk from AI. Unfortunately, my disagreements are all over the place. It's not that I disagree with one or two key points: there are many elements of the standard argument that I diverge from, and depending on the audience, I don't know which points of disagreement people think are most important.I want to write a post highlighting all the important areas where I disagree, and offering my own counterarguments as an alternative. This post would benefit from responding to an existing piece, along the same lines as Quintin Pope's article "My Objections to "We're All Gonna Die with Eliezer Yudkowsky"". By contrast, it would be intended to address the EA community as a whole, since I'm aware many EAs already disagree with Yudkowsky even if they buy the basic arguments for AI x-risks.My question is: what is the current best single article (or set of articles) that provide a well-reasoned and comprehensive case for believing that there is a substantial (>10%) probability of an AI catastrophe this century?I was considering replying to Joseph Carlsmith's article, "Is Power-Seeking AI an Existential Risk?", since it seemed reasonably comprehensive and representative of the concerns EAs have about AI x-risk. However, I'm a bit worried that the article is not very representative of EAs who have substantial probabilities of doom, since he originally estimated a total risk of catastrophe at only 5% before 2070. In May 2022, Carlsmith changed his mind and reported a higher probability, but I am not sure whether this is because he has been exposed to new arguments, or because he simply thinks the stated arguments are stronger than he originally thought.I suspect I have both significant moral disagreements and significant empirical disagreements with EAs, and I want to include both in such an article, while mainly focusing on the empirical points. For example, I have the feeling that I disagree with most EAs about:How bad human disempowerment would likely be from a utilitarian perspective, and what "human disempowerment" even means in the first placeWhether there will be a treacherous turn event, during which AIs violently take over the world after previously having been behaviorally aligned with humansHow likely AIs are to coordinate near-perfectly with each other as a unified front, leaving humans out of their coalitionWhether we should expect AI values to be "alien" (like paperclip maximizers) in the absence of extraordinary efforts to align them with humansWhether the AIs themselves will be significant moral patients, on par with humansWhether there will be a qualitative moment when "the AGI" is created, rather than systems incrementally getting more advanced, with no clear finish lineWhether we get only "one critical try" to align AGIWhether "AI lab leaks" are an important source of AI riskHow likely AIs are to kill every single human if they are unaligned with humansWhether there will be a "value lock-in" event soon after we create powerful AI that causes values to cease their evolution over the coming billions of yearsHow bad problems related to "specification gaming" will be in the futureHow society is likely to respond to AI risks, and whether they'll sleepwalk into a catastropheHowever, I also disagree with points made by many other EAs who have argued against the standard AI risk case. For example, I think that,AIs will eventually become vastly more powerful and smarter than humans. So, I think AIs will eventually be able to "defeat all of us combined"I think a benign "AI takeover" event is very likely even if we align AIs successfullyAIs will likely be goal-...
Create your
podcast in
minutes
It is Free