Jump to content

Existential risk from artificial general intelligence: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
→‎Anthropomorphic arguments: I didn't fix it I just deleted a part that was sexist
Tags: Reverted Mobile edit Mobile web edit
m Added a category
(24 intermediate revisions by 15 users not shown)
Line 3: Line 3:
{{Artificial intelligence}}
{{Artificial intelligence}}


'''Existential risk from artificial general intelligence''' is the idea that substantial progress in [[artificial general intelligence]] (AGI) could result in [[human extinction]] or an irreversible [[global catastrophic risk|global catastrophe]].<ref name="aima">{{Cite book |last1=Russell |first1=Stuart |title=Artificial Intelligence: A Modern Approach |title-link=Artificial Intelligence: A Modern Approach |last2=Norvig |first2=Peter |date=2009 |publisher=Prentice Hall |isbn=978-0-13-604259-4 |chapter=26.3: The Ethics and Risks of Developing Artificial Intelligence |author-link=Stuart J. Russell |author-link2=Peter Norvig}}</ref><ref>{{Cite journal |last=Bostrom |first=Nick |author-link=Nick Bostrom |date=2002 |title=Existential risks |journal=[[Journal of Evolution and Technology]] |volume=9 |issue=1 |pages=1–31}}</ref><ref name="auto1">{{Cite journal |last1=Turchin |first1=Alexey |last2=Denkenberger |first2=David |date=2018-05-03 |title=Classification of global catastrophic risks connected with artificial intelligence |journal=AI & Society |volume=35 |issue=1 |pages=147–163 |doi=10.1007/s00146-018-0845-5 |issn=0951-5666 |s2cid=19208453|url=https://philarchive.org/rec/TURCOG-2 }}</ref>
'''Existential risk from artificial general intelligence''' refers to the idea that substantial progress in [[artificial general intelligence]] (AGI) could lead to [[human extinction]] or an irreversible [[global catastrophic risk|global catastrophe]].<ref name="aima">{{Cite book |last1=Russell |first1=Stuart |title=Artificial Intelligence: A Modern Approach |title-link=Artificial Intelligence: A Modern Approach |last2=Norvig |first2=Peter |date=2009 |publisher=Prentice Hall |isbn=978-0-13-604259-4 |chapter=26.3: The Ethics and Risks of Developing Artificial Intelligence |author-link=Stuart J. Russell |author-link2=Peter Norvig}}</ref><ref>{{Cite journal |last=Bostrom |first=Nick |author-link=Nick Bostrom |date=2002 |title=Existential risks |journal=[[Journal of Evolution and Technology]] |volume=9 |issue=1 |pages=1–31}}</ref><ref name="auto1">{{Cite journal |last1=Turchin |first1=Alexey |last2=Denkenberger |first2=David |date=2018-05-03 |title=Classification of global catastrophic risks connected with artificial intelligence |journal=AI & Society |volume=35 |issue=1 |pages=147–163 |doi=10.1007/s00146-018-0845-5 |issn=0951-5666 |s2cid=19208453|url=https://philarchive.org/rec/TURCOG-2 }}</ref>


One argument goes as follows: [[human species|human beings]] dominate other species because the [[human brain]] possesses distinctive capabilities other animals lack. If AI were to surpass humanity in general [[intelligence]] and become [[superintelligence|superintelligent]], then it could become difficult or impossible to control. Just as the fate of the [[mountain gorilla]] depends on human goodwill, so might the fate of humanity depend on the actions of a future machine superintelligence.<ref name="superintelligence">{{Cite book |last=Bostrom |first=Nick |title=Superintelligence: Paths, Dangers, Strategies |title-link=Superintelligence: Paths, Dangers, Strategies |date=2014 |publisher=Oxford University Press |isbn=978-0-19-967811-2 |edition=First |author-link=Nick Bostrom}}<!-- preface --></ref>
One argument for the importance of this risk references how [[human species|human beings]] dominate other species because the [[human brain]] possesses distinctive capabilities other animals lack. If AI were to surpass [[human intelligence]] and become [[superintelligence|superintelligent]], it might become uncontrollable. Just as the fate of the [[mountain gorilla]] depends on human goodwill, the fate of humanity could depend on the actions of a future machine superintelligence.<ref name="superintelligence">{{Cite book |last=Bostrom |first=Nick |title=Superintelligence: Paths, Dangers, Strategies |title-link=Superintelligence: Paths, Dangers, Strategies |date=2014 |publisher=Oxford University Press |isbn=978-0-19-967811-2 |edition=First |author-link=Nick Bostrom}}<!-- preface --></ref>


The plausibility of existential catastrophe due to AI is widely debated, and hinges in part on whether AGI or superintelligence are achievable, the speed at which dangerous capabilities and behaviors emerge,<ref>{{Cite news |last=Vynck |first=Gerrit De |date=2023-05-23 |title=The debate over whether AI will destroy us is dividing Silicon Valley |language=en-US |newspaper=Washington Post |url=https://www.washingtonpost.com/technology/2023/05/20/ai-existential-risk-debate/ |access-date=2023-07-27 |issn=0190-8286}}</ref> and whether practical scenarios for [[AI takeover]]s exist.<ref>{{Cite news |last=Metz |first=Cade |date=2023-06-10 |title=How Could A.I. Destroy Humanity? |language=en-US |work=The New York Times |url=https://www.nytimes.com/2023/06/10/technology/ai-humanity.html |access-date=2023-07-27 |issn=0362-4331}}</ref> Concerns about superintelligence have been voiced by leading computer scientists and tech [[Chief executive officer|CEOs]] such as [[Geoffrey Hinton]],<ref>{{Cite web |date=25 March 2023 |title='Godfather of artificial intelligence' weighs in on the past and potential of AI |url=https://www.cbsnews.com/news/godfather-of-artificial-intelligence-weighs-in-on-the-past-and-potential-of-artificial-intelligence/ |access-date=2023-04-10 |website=www.cbsnews.com |language=en-US}}</ref> [[Yoshua Bengio]],<ref>{{Cite web |title=How Rogue AIs may Arise|url=https://yoshuabengio.org/2023/05/22/how-rogue-ais-may-arise/ |access-date=2023-05-26 |website=yoshuabengio.org |date=26 May 2023 |language=en-US}}</ref> [[Alan Turing]],{{efn|In a 1951 lecture<ref>{{Cite speech| last = Turing| first = Alan| title = Intelligent machinery, a heretical theory| event = Lecture given to '51 Society'| location = Manchester| access-date = 2022-07-22| date = 1951| publisher = The Turing Digital Archive| url = https://turingarchive.kings.cam.ac.uk/publications-lectures-and-talks-amtb/amt-b-4| page = 16| archive-date = September 26, 2022| archive-url = https://web.archive.org/web/20220926004549/https://turingarchive.kings.cam.ac.uk/publications-lectures-and-talks-amtb/amt-b-4| url-status = live}}</ref> Turing argued that "It seems probable that once the machine thinking method had started, it would not take long to outstrip our feeble powers. There would be no question of the machines dying, and they would be able to converse with each other to sharpen their wits. At some stage therefore we should have to expect the machines to take control, in the way that is mentioned in Samuel Butler's Erewhon". Also in a lecture broadcast on the [[BBC]]<ref>{{Cite episode |title= Can digital computers think?|series=Automatic Calculating Machines |first=Alan |last=Turing |network= BBC |date=15 May 1951 |number=2 |transcript=Can digital computers think? |transcript-url=https://turingarchive.kings.cam.ac.uk/publications-lectures-and-talks-amtb/amt-b-6 }}</ref> he expressed the opinion: "If a machine can think, it might think more intelligently than we do, and then where should we be? Even if we could keep the machines in a subservient position, for instance by turning off the power at strategic moments, we should, as a species, feel greatly humbled.&nbsp;.&nbsp;.&nbsp;. This new danger&nbsp;.&nbsp;.&nbsp;. is certainly something which can give us anxiety."|name=turing_note}} [[Elon Musk]],<ref name="Parkin">{{Cite news |last=Parkin |first=Simon |date=14 June 2015 |title=Science fiction no more? Channel 4's Humans and our rogue AI obsessions |language=en |work=[[The Guardian]] |url=https://www.theguardian.com/tv-and-radio/2015/jun/14/science-fiction-no-more-humans-tv-artificial-intelligence |url-status=live |access-date=5 February 2018 |archive-url=https://web.archive.org/web/20180205184322/https://www.theguardian.com/tv-and-radio/2015/jun/14/science-fiction-no-more-humans-tv-artificial-intelligence |archive-date=5 February 2018}}</ref> and [[OpenAI]] CEO [[Sam Altman]].<ref name="Jackson">{{Cite web |last=Jackson |first=Sarah |title=The CEO of the company behind AI chatbot ChatGPT says the worst-case scenario for artificial intelligence is 'lights out for all of us' |url=https://www.businessinsider.com/chatgpt-openai-ceo-worst-case-ai-lights-out-for-all-2023-1 |access-date=2023-04-10 |website=Business Insider |language=en-US}}</ref> In 2022, a survey of AI researchers with a 17% response rate found that the majority of respondents believed there is a 10 percent or greater chance that our inability to control AI will cause an existential catastrophe.<ref>{{Cite web |title=The AI Dilemma |url=https://www.humanetech.com/podcast/the-ai-dilemma |access-date=2023-04-10 |website=www.humanetech.com |language=en|quote=50% of AI researchers believe there's a 10% or greater chance that humans go extinct from our inability to control AI.}}</ref><ref name=":8">{{Cite web |date=2022-08-04 |title=2022 Expert Survey on Progress in AI |url=https://aiimpacts.org/2022-expert-survey-on-progress-in-ai/ |access-date=2023-04-10 |website=AI Impacts |language=en-US}}</ref> In 2023, hundreds of AI experts and other notable figures [[Statement on AI risk of extinction|signed a statement]] that "Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as [[pandemic]]s and [[Nuclear warfare|nuclear war]]".<ref>{{Cite news |last=Roose |first=Kevin |date=2023-05-30 |title=A.I. Poses 'Risk of Extinction,' Industry Leaders Warn |language=en-US |work=The New York Times |url=https://www.nytimes.com/2023/05/30/technology/ai-threat-warning.html |access-date=2023-06-03 |issn=0362-4331}}</ref> Following increased concern over AI risks, government leaders such as [[Prime Minister of the United Kingdom|United Kingdom prime minister]] [[Rishi Sunak]]<ref>{{Cite magazine |last=Sunak |first=Rishi |date=June 14, 2023 |title=Rishi Sunak Wants the U.K. to Be a Key Player in Global AI Regulation |magazine=Time |url=https://time.com/6287253/uk-rishi-sunak-ai-regulation/}}</ref> and [[Secretary-General of the United Nations|United Nations Secretary-General]] [[António Guterres]]<ref name=":12" /> called for an increased focus on global [[Regulation of artificial intelligence|AI regulation]].
The plausibility of existential catastrophe due to AI is widely debated. It hinges in part on whether AGI or superintelligence are achievable, the speed at which dangerous capabilities and behaviors emerge,<ref>{{Cite news |last=Vynck |first=Gerrit De |date=2023-05-23 |title=The debate over whether AI will destroy us is dividing Silicon Valley |language=en-US |newspaper=Washington Post |url=https://www.washingtonpost.com/technology/2023/05/20/ai-existential-risk-debate/ |access-date=2023-07-27 |issn=0190-8286}}</ref> and whether practical scenarios for [[AI takeover]]s exist.<ref>{{Cite news |last=Metz |first=Cade |date=2023-06-10 |title=How Could A.I. Destroy Humanity? |language=en-US |work=The New York Times |url=https://www.nytimes.com/2023/06/10/technology/ai-humanity.html |access-date=2023-07-27 |issn=0362-4331}}</ref> Concerns about superintelligence have been voiced by leading computer scientists and tech [[Chief executive officer|CEOs]] such as [[Geoffrey Hinton]],<ref>{{Cite web |date=25 March 2023 |title='Godfather of artificial intelligence' weighs in on the past and potential of AI |url=https://www.cbsnews.com/news/godfather-of-artificial-intelligence-weighs-in-on-the-past-and-potential-of-artificial-intelligence/ |access-date=2023-04-10 |website=www.cbsnews.com |language=en-US}}</ref> [[Yoshua Bengio]],<ref>{{Cite web |title=How Rogue AIs may Arise|url=https://yoshuabengio.org/2023/05/22/how-rogue-ais-may-arise/ |access-date=2023-05-26 |website=yoshuabengio.org |date=26 May 2023 |language=en-US}}</ref> [[Alan Turing]],{{efn|In a 1951 lecture<ref>{{Cite speech| last = Turing| first = Alan| title = Intelligent machinery, a heretical theory| event = Lecture given to '51 Society'| location = Manchester| access-date = 2022-07-22| date = 1951| publisher = The Turing Digital Archive| url = https://turingarchive.kings.cam.ac.uk/publications-lectures-and-talks-amtb/amt-b-4| page = 16| archive-date = September 26, 2022| archive-url = https://web.archive.org/web/20220926004549/https://turingarchive.kings.cam.ac.uk/publications-lectures-and-talks-amtb/amt-b-4| url-status = live}}</ref> Turing argued that "It seems probable that once the machine thinking method had started, it would not take long to outstrip our feeble powers. There would be no question of the machines dying, and they would be able to converse with each other to sharpen their wits. At some stage therefore we should have to expect the machines to take control, in the way that is mentioned in Samuel Butler's Erewhon". Also in a lecture broadcast on the [[BBC]]<ref>{{Cite episode |title= Can digital computers think?|series=Automatic Calculating Machines |first=Alan |last=Turing |network= BBC |date=15 May 1951 |number=2 |transcript=Can digital computers think? |transcript-url=https://turingarchive.kings.cam.ac.uk/publications-lectures-and-talks-amtb/amt-b-6 }}</ref> he expressed the opinion: "If a machine can think, it might think more intelligently than we do, and then where should we be? Even if we could keep the machines in a subservient position, for instance by turning off the power at strategic moments, we should, as a species, feel greatly humbled... This new danger... is certainly something which can give us anxiety."|name=turing_note}} [[Elon Musk]],<ref name="Parkin">{{Cite news |last=Parkin |first=Simon |date=14 June 2015 |title=Science fiction no more? Channel 4's Humans and our rogue AI obsessions |language=en |work=[[The Guardian]] |url=https://www.theguardian.com/tv-and-radio/2015/jun/14/science-fiction-no-more-humans-tv-artificial-intelligence |url-status=live |access-date=5 February 2018 |archive-url=https://web.archive.org/web/20180205184322/https://www.theguardian.com/tv-and-radio/2015/jun/14/science-fiction-no-more-humans-tv-artificial-intelligence |archive-date=5 February 2018}}</ref> and [[OpenAI]] CEO [[Sam Altman]].<ref name="Jackson">{{Cite web |last=Jackson |first=Sarah |title=The CEO of the company behind AI chatbot ChatGPT says the worst-case scenario for artificial intelligence is 'lights out for all of us' |url=https://www.businessinsider.com/chatgpt-openai-ceo-worst-case-ai-lights-out-for-all-2023-1 |access-date=2023-04-10 |website=Business Insider |language=en-US}}</ref> In 2022, a survey of AI researchers with a 17% response rate found that the majority believed there is a 10 percent or greater chance that human inability to control AI will cause an existential catastrophe.<ref>{{Cite web |title=The AI Dilemma |url=https://www.humanetech.com/podcast/the-ai-dilemma |access-date=2023-04-10 |website=www.humanetech.com |language=en|quote=50% of AI researchers believe there's a 10% or greater chance that humans go extinct from our inability to control AI.}}</ref><ref name=":8">{{Cite web |date=2022-08-04 |title=2022 Expert Survey on Progress in AI |url=https://aiimpacts.org/2022-expert-survey-on-progress-in-ai/ |access-date=2023-04-10 |website=AI Impacts |language=en-US}}</ref> In 2023, hundreds of AI experts and other notable figures [[Statement on AI risk of extinction|signed a statement]] declaring, "Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as [[pandemic]]s and [[Nuclear warfare|nuclear war]]".<ref>{{Cite news |last=Roose |first=Kevin |date=2023-05-30 |title=A.I. Poses 'Risk of Extinction,' Industry Leaders Warn |language=en-US |work=The New York Times |url=https://www.nytimes.com/2023/05/30/technology/ai-threat-warning.html |access-date=2023-06-03 |issn=0362-4331}}</ref> Following increased concern over AI risks, government leaders such as [[Prime Minister of the United Kingdom|United Kingdom prime minister]] [[Rishi Sunak]]<ref>{{Cite magazine |last=Sunak |first=Rishi |date=June 14, 2023 |title=Rishi Sunak Wants the U.K. to Be a Key Player in Global AI Regulation |magazine=Time |url=https://time.com/6287253/uk-rishi-sunak-ai-regulation/}}</ref> and [[Secretary-General of the United Nations|United Nations Secretary-General]] [[António Guterres]]<ref name=":12" /> called for an increased focus on global [[Regulation of artificial intelligence|AI regulation]].


Two sources of concern stem from the problems of AI [[AI capability control|control]] and [[AI alignment|alignment]]: controlling a superintelligent machine or instilling it with human-compatible values may be difficult. Many researchers believe that a superintelligent machine would resist attempts to disable it or change its goals, as that would prevent it from accomplishing its present goals. It would be extremely difficult to align a superintelligence with the full breadth of significant human values and constraints.<ref name="aima" /><ref name="yudkowsky-global-risk">{{Cite journal |last=Yudkowsky |first=Eliezer |date=2008 |title=Artificial Intelligence as a Positive and Negative Factor in Global Risk |url=https://intelligence.org/files/AIPosNegFactor.pdf |journal=Global Catastrophic Risks |pages=308–345 |bibcode=2008gcr..book..303Y |access-date=27 August 2018 |archive-date=2 March 2013 |archive-url=https://web.archive.org/web/20130302173022/http://intelligence.org/files/AIPosNegFactor.pdf |url-status=live }}</ref><ref name="research-priorities">{{cite journal |title=Research Priorities for Robust and Beneficial Artificial Intelligence |author1-last=Russell |author1-first=Stuart |author1-link=Stuart J. Russell |author2-last=Dewey |author2-first=Daniel |author3-last=Tegmark |author3-first=Max |author3-link=Max Tegmark |journal=AI Magazine |pages=105–114 |publisher=Association for the Advancement of Artificial Intelligence |year=2015 |url=https://futureoflife.org/data/documents/research_priorities.pdf |bibcode=2016arXiv160203506R |arxiv=1602.03506 |access-date=10 August 2019 |archive-date=4 August 2019 |archive-url=https://web.archive.org/web/20190804145930/https://futureoflife.org/data/documents/research_priorities.pdf |url-status=live }}, cited in {{Cite web |date=January 2015 |title=AI Open Letter - Future of Life Institute |url=https://futureoflife.org/ai-open-letter |access-date=2019-08-09 |website=Future of Life Institute |archive-date=10 August 2019 |archive-url=https://web.archive.org/web/20190810020404/https://futureoflife.org/ai-open-letter |url-status=live }}</ref> In contrast, skeptics such as [[computer scientist]] [[Yann LeCun]] argue that superintelligent machines will have no desire for self-preservation.<ref name="vanity">{{Cite news |last=Dowd |first=Maureen |date=April 2017 |title=Elon Musk's Billion-Dollar Crusade to Stop the A.I. Apocalypse |language=en |work=The Hive |url=https://www.vanityfair.com/news/2017/03/elon-musk-billion-dollar-crusade-to-stop-ai-space-x |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20180726041656/https://www.vanityfair.com/news/2017/03/elon-musk-billion-dollar-crusade-to-stop-ai-space-x |archive-date=26 July 2018}}</ref>
Two sources of concern stem from the problems of AI [[AI capability control|control]] and [[AI alignment|alignment]]. Controlling a superintelligent machine or instilling it with human-compatible values may be difficult. Many researchers believe that a superintelligent machine would likely resist attempts to disable it or change its goals as that would prevent it from accomplishing its present goals. It would be extremely challenging to align a superintelligence with the full breadth of significant human values and constraints.<ref name="aima" /><ref name="yudkowsky-global-risk">{{Cite journal |last=Yudkowsky |first=Eliezer |date=2008 |title=Artificial Intelligence as a Positive and Negative Factor in Global Risk |url=https://intelligence.org/files/AIPosNegFactor.pdf |journal=Global Catastrophic Risks |pages=308–345 |bibcode=2008gcr..book..303Y |access-date=27 August 2018 |archive-date=2 March 2013 |archive-url=https://web.archive.org/web/20130302173022/http://intelligence.org/files/AIPosNegFactor.pdf |url-status=live }}</ref><ref name="research-priorities">{{cite journal |title=Research Priorities for Robust and Beneficial Artificial Intelligence |author1-last=Russell |author1-first=Stuart |author1-link=Stuart J. Russell |author2-last=Dewey |author2-first=Daniel |author3-last=Tegmark |author3-first=Max |author3-link=Max Tegmark |journal=AI Magazine |pages=105–114 |publisher=Association for the Advancement of Artificial Intelligence |year=2015 |url=https://futureoflife.org/data/documents/research_priorities.pdf |bibcode=2016arXiv160203506R |arxiv=1602.03506 |access-date=10 August 2019 |archive-date=4 August 2019 |archive-url=https://web.archive.org/web/20190804145930/https://futureoflife.org/data/documents/research_priorities.pdf |url-status=live }}, cited in {{Cite web |date=January 2015 |title=AI Open Letter - Future of Life Institute |url=https://futureoflife.org/ai-open-letter |access-date=2019-08-09 |website=Future of Life Institute |archive-date=10 August 2019 |archive-url=https://web.archive.org/web/20190810020404/https://futureoflife.org/ai-open-letter |url-status=live }}</ref> In contrast, skeptics such as [[computer scientist]] [[Yann LeCun]] argue that superintelligent machines will have no desire for self-preservation.<ref name="vanity">{{Cite news |last=Dowd |first=Maureen |date=April 2017 |title=Elon Musk's Billion-Dollar Crusade to Stop the A.I. Apocalypse |language=en |work=The Hive |url=https://www.vanityfair.com/news/2017/03/elon-musk-billion-dollar-crusade-to-stop-ai-space-x |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20180726041656/https://www.vanityfair.com/news/2017/03/elon-musk-billion-dollar-crusade-to-stop-ai-space-x |archive-date=26 July 2018}}</ref>


A third source of concern is that a sudden "[[intelligence explosion]]" might take an unprepared human race by surprise. Such scenarios consider the possibility that an AI that is more intelligent than its creators might be able to [[Recursion|recursively]] improve itself at an exponentially increasing rate, improving too quickly for its handlers and society at large to control.<ref name="aima" /><ref name="yudkowsky-global-risk" /> Empirically, examples like [[AlphaZero]] teaching itself to play [[Go (game)|Go]] show that domain-specific AI systems can sometimes progress from subhuman to superhuman ability very quickly, although such systems do not involve altering their fundamental architecture.<ref>{{Cite web |title=AlphaGo Zero: Starting from scratch |url=https://www.deepmind.com/blog/alphago-zero-starting-from-scratch |access-date=2023-07-28 |website=www.deepmind.com |date=18 October 2017 |language=en}}</ref>
A third source of concern is the possibility of a sudden "[[intelligence explosion]]" that catches humanity unprepared. In this scenario, an AI more intelligent than its creators would be able to [[Recursive self-improvement|recursively improve itself]] at an exponentially increasing rate, improving too quickly for its handlers or society at large to control.<ref name="aima" /><ref name="yudkowsky-global-risk" /> Empirically, examples like [[AlphaZero]], which taught itself to play [[Go (game)|Go]] and quickly surpassed human ability, show that domain-specific AI systems can sometimes progress from subhuman to superhuman ability very quickly, although such [[machine learning]] systems do not recursively improve their fundamental architecture.<ref>{{Cite web |title=AlphaGo Zero: Starting from scratch |url=https://www.deepmind.com/blog/alphago-zero-starting-from-scratch |access-date=2023-07-28 |website=www.deepmind.com |date=18 October 2017 |language=en}}</ref>


==History==
==History==
Line 22: Line 22:
In 1965, [[I. J. Good]] originated the concept now known as an "intelligence explosion" and said the risks were underappreciated:<ref>{{Cite news |last=Hilliard |first=Mark |date=2017 |title=The AI apocalypse: will the human race soon be terminated? |language=en |newspaper=The Irish Times |url=https://www.irishtimes.com/business/innovation/the-ai-apocalypse-will-the-human-race-soon-be-terminated-1.3019220 |access-date=15 March 2020 |archive-date=22 May 2020 |archive-url=https://web.archive.org/web/20200522114127/https://www.irishtimes.com/business/innovation/the-ai-apocalypse-will-the-human-race-soon-be-terminated-1.3019220 |url-status=live }}</ref>
In 1965, [[I. J. Good]] originated the concept now known as an "intelligence explosion" and said the risks were underappreciated:<ref>{{Cite news |last=Hilliard |first=Mark |date=2017 |title=The AI apocalypse: will the human race soon be terminated? |language=en |newspaper=The Irish Times |url=https://www.irishtimes.com/business/innovation/the-ai-apocalypse-will-the-human-race-soon-be-terminated-1.3019220 |access-date=15 March 2020 |archive-date=22 May 2020 |archive-url=https://web.archive.org/web/20200522114127/https://www.irishtimes.com/business/innovation/the-ai-apocalypse-will-the-human-race-soon-be-terminated-1.3019220 |url-status=live }}</ref>


{{cquote|Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines; there would then unquestionably be an 'intelligence explosion', and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is the last invention that man need ever make, provided that the machine is docile enough to tell us how to keep it under control. It is curious that this point is made so seldom outside of science fiction. It is sometimes worthwhile to take science fiction seriously.<ref>I.J. Good, [http://commonsenseatheism.com/wp-content/uploads/2011/02/Good-Speculations-Concerning-the-First-Ultraintelligent-Machine.pdf "Speculations Concerning the First Ultraintelligent Machine"] {{webarchive|url=https://web.archive.org/web/20111128085512/http://commonsenseatheism.com/wp-content/uploads/2011/02/Good-Speculations-Concerning-the-First-Ultraintelligent-Machine.pdf |date=2011-11-28 }} ([http://www.acceleratingfuture.com/pages/ultraintelligentmachine.html HTML] {{Webarchive|url=https://web.archive.org/web/20111128085512/http://commonsenseatheism.com/wp-content/uploads/2011/02/Good-Speculations-Concerning-the-First-Ultraintelligent-Machine.pdf |date=28 November 2011 }} ), ''Advances in Computers'', vol. 6, 1965.</ref>}}
{{cquote|Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines; there would then unquestionably be an 'intelligence explosion', and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is the last invention that man need ever make, provided that the machine is docile enough to tell us how to keep it under control. It is curious that this point is made so seldom outside of science fiction. It is sometimes worthwhile to take science fiction seriously.<ref>I.J. Good, [http://commonsenseatheism.com/wp-content/uploads/2011/02/Good-Speculations-Concerning-the-First-Ultraintelligent-Machine.pdf "Speculations Concerning the First Ultraintelligent Machine"] {{webarchive|url=https://web.archive.org/web/20111128085512/http://commonsenseatheism.com/wp-content/uploads/2011/02/Good-Speculations-Concerning-the-First-Ultraintelligent-Machine.pdf |date=2011-11-28 }} ([http://www.acceleratingfuture.com/pages/ultraintelligentmachine.html HTML] ), ''Advances in Computers'', vol. 6, 1965.</ref>}}


Scholars such as [[Marvin Minsky]]<ref>{{Cite book |last1=Russell |first1=Stuart J. |title=Artificial Intelligence: A Modern Approach |title-link=Artificial Intelligence: A Modern Approach |last2=Norvig |first2=Peter |date=2003 |publisher=Prentice Hall |isbn=978-0-13-790395-5 |location=Upper Saddle River, New Jersey |at= |language=en-us |chapter=Section 26.3: The Ethics and Risks of Developing Artificial Intelligence |quote=Similarly, Marvin Minsky once suggested that an AI program designed to solve the Riemann Hypothesis might end up taking over all the resources of Earth to build more powerful supercomputers to help achieve its goal.}}</ref> and I. J. Good himself<ref>{{Cite book |last=Barrat |first=James |title=Our final invention: artificial intelligence and the end of the human era |date=2013 |publisher=St. Martin's Press |isbn=978-0-312-62237-4 |edition=First |location=New York |language=en-us |quote=In the bio, playfully written in the third person, Good summarized his life's milestones, including a probably never before seen account of his work at Bletchley Park with Turing. But here's what he wrote in 1998 about the first superintelligence, and his late-in-the-game U-turn: [The paper] 'Speculations Concerning the First Ultra-intelligent Machine' (1965)...began: 'The survival of man depends on the early construction of an ultra-intelligent machine.' Those were his [Good's] words during the Cold War, and he now suspects that 'survival' should be replaced by 'extinction.' He thinks that, because of international competition, we cannot prevent the machines from taking over. He thinks we are lemmings. He said also that 'probably Man will construct the deus ex machina in his own image.'}}</ref> occasionally expressed concern that a superintelligence could seize control, but issued no call to action. In 2000, computer scientist and [[Sun microsystems|Sun]] co-founder [[Bill Joy]] penned an influential essay, "[[Why The Future Doesn't Need Us]]", identifying superintelligent robots as a high-tech danger to human survival, alongside [[nanotechnology]] and engineered bioplagues.<ref>{{Cite news |last=Anderson |first=Kurt |date=26 November 2014 |title=Enthusiasts and Skeptics Debate Artificial Intelligence |work=[[Vanity Fair (magazine)|Vanity Fair]] |url=https://www.vanityfair.com/news/tech/2014/11/artificial-intelligence-singularity-theory |access-date=30 January 2016 |archive-date=22 January 2016 |archive-url=https://web.archive.org/web/20160122025154/http://www.vanityfair.com/news/tech/2014/11/artificial-intelligence-singularity-theory |url-status=live }}</ref>
Scholars such as [[Marvin Minsky]]<ref>{{Cite book |last1=Russell |first1=Stuart J. |title=Artificial Intelligence: A Modern Approach |title-link=Artificial Intelligence: A Modern Approach |last2=Norvig |first2=Peter |date=2003 |publisher=Prentice Hall |isbn=978-0-13-790395-5 |location=Upper Saddle River, New Jersey |at= |language=en-us |chapter=Section 26.3: The Ethics and Risks of Developing Artificial Intelligence |quote=Similarly, Marvin Minsky once suggested that an AI program designed to solve the Riemann Hypothesis might end up taking over all the resources of Earth to build more powerful supercomputers to help achieve its goal.}}</ref> and I. J. Good himself<ref>{{Cite book |last=Barrat |first=James |title=Our final invention: artificial intelligence and the end of the human era |date=2013 |publisher=St. Martin's Press |isbn=978-0-312-62237-4 |edition=First |location=New York |language=en-us |quote=In the bio, playfully written in the third person, Good summarized his life's milestones, including a probably never before seen account of his work at Bletchley Park with Turing. But here's what he wrote in 1998 about the first superintelligence, and his late-in-the-game U-turn: [The paper] 'Speculations Concerning the First Ultra-intelligent Machine' (1965)...began: 'The survival of man depends on the early construction of an ultra-intelligent machine.' Those were his [Good's] words during the Cold War, and he now suspects that 'survival' should be replaced by 'extinction.' He thinks that, because of international competition, we cannot prevent the machines from taking over. He thinks we are lemmings. He said also that 'probably Man will construct the deus ex machina in his own image.'}}</ref> occasionally expressed concern that a superintelligence could seize control, but issued no call to action. In 2000, computer scientist and [[Sun microsystems|Sun]] co-founder [[Bill Joy]] penned an influential essay, "[[Why The Future Doesn't Need Us]]", identifying superintelligent robots as a high-tech danger to human survival, alongside [[nanotechnology]] and engineered bioplagues.<ref>{{Cite news |last=Anderson |first=Kurt |date=26 November 2014 |title=Enthusiasts and Skeptics Debate Artificial Intelligence |work=[[Vanity Fair (magazine)|Vanity Fair]] |url=https://www.vanityfair.com/news/tech/2014/11/artificial-intelligence-singularity-theory |access-date=30 January 2016 |archive-date=22 January 2016 |archive-url=https://web.archive.org/web/20160122025154/http://www.vanityfair.com/news/tech/2014/11/artificial-intelligence-singularity-theory |url-status=live }}</ref>
Line 28: Line 28:
[[Nick Bostrom]] published ''[[Superintelligence: Paths, Dangers, Strategies|Superintelligence]]'' in 2014, which presented his arguments that superintelligence poses an existential threat.<ref>{{Cite news |last=Metz |first=Cade |date=9 June 2018 |title=Mark Zuckerberg, Elon Musk and the Feud Over Killer Robots |work=The New York Times |url=https://www.nytimes.com/2018/06/09/technology/elon-musk-mark-zuckerberg-artificial-intelligence.html |access-date=3 April 2019 |archive-date=15 February 2021 |archive-url=https://web.archive.org/web/20210215051949/https://www.nytimes.com/2018/06/09/technology/elon-musk-mark-zuckerberg-artificial-intelligence.html |url-status=live }}</ref> By 2015, public figures such as physicists [[Stephen Hawking]] and Nobel laureate [[Frank Wilczek]], computer scientists [[Stuart J. Russell]] and [[Roman Yampolskiy]], and entrepreneurs [[Elon Musk]] and [[Bill Gates]] were expressing concern about the risks of superintelligence.<ref>{{Cite news |last=Hsu |first=Jeremy |date=1 March 2012 |title=Control dangerous AI before it controls us, one expert says |work=[[NBC News]] |url=http://www.nbcnews.com/id/46590591/ns/technology_and_science-innovation |access-date=28 January 2016 |archive-date=2 February 2016 |archive-url=https://web.archive.org/web/20160202173621/http://www.nbcnews.com/id/46590591/ns/technology_and_science-innovation |url-status=live }}</ref><ref name="hawking editorial">{{Cite news |title=Stephen Hawking: 'Transcendence looks at the implications of artificial intelligence&nbsp;– but are we taking AI seriously enough?' |publisher=[[The Independent (UK)]] |url=https://www.independent.co.uk/news/science/stephen-hawking-transcendence-looks-at-the-implications-of-artificial-intelligence--but-are-we-taking-ai-seriously-enough-9313474.html |url-status=live |access-date=3 December 2014 |archive-url=https://web.archive.org/web/20150925153716/http://www.independent.co.uk/news/science/stephen-hawking-transcendence-looks-at-the-implications-of-artificial-intelligence--but-are-we-taking-ai-seriously-enough-9313474.html |archive-date=25 September 2015}}</ref><ref name="bbc on hawking editorial">{{Cite news |date=2 December 2014 |title=Stephen Hawking warns artificial intelligence could end mankind |publisher=[[BBC]] |url=https://www.bbc.com/news/technology-30290540 |url-status=live |access-date=3 December 2014 |archive-url=https://web.archive.org/web/20151030054329/http://www.bbc.com/news/technology-30290540 |archive-date=30 October 2015}}</ref><ref>{{Cite news |last=Eadicicco |first=Lisa |date=28 January 2015 |title=Bill Gates: Elon Musk Is Right, We Should All Be Scared Of Artificial Intelligence Wiping Out Humanity |work=[[Business Insider]] |url=http://www.businessinsider.com/bill-gates-artificial-intelligence-2015-1 |access-date=30 January 2016 |archive-date=26 February 2016 |archive-url=https://web.archive.org/web/20160226090602/http://www.businessinsider.com/bill-gates-artificial-intelligence-2015-1 |url-status=live }}</ref> Also in 2015, the [[Open Letter on Artificial Intelligence]] highlighted the "great potential of AI" and encouraged more research on how to make it robust and beneficial.<ref>{{Cite web |title=Research Priorities for Robust and Beneficial Artificial Intelligence: an Open Letter |url=http://futureoflife.org/misc/open_letter |url-status=live |archive-url=https://web.archive.org/web/20150115160823/http://futureoflife.org/misc/open_letter |archive-date=15 January 2015 |access-date=23 October 2015 |publisher=[[Future of Life Institute]]}}</ref> In April 2016, the journal ''[[Nature (journal)|Nature]]'' warned: "Machines and robots that outperform humans across the board could self-improve beyond our control—and their interests might not align with ours".<ref>{{Cite journal |date=2016 |title=Anticipating artificial intelligence |journal=Nature |language=en |volume=532 |issue=7600 |page=413 |doi=10.1038/532413a |pmid=27121801 |bibcode=2016Natur.532Q.413. |s2cid=4399193 |issn=1476-4687|doi-access=free }}</ref> In 2020, [[Brian Christian]] published ''[[The Alignment Problem]]'', which details the history of progress on AI alignment up to that time.<ref>{{cite book |last=Christian |first=Brian |date=October 6, 2020 |title=The Alignment Problem: Machine Learning and Human Values |url=https://brianchristian.org/the-alignment-problem/ |publisher=[[W. W. Norton & Company]] |isbn=978-0-393-63582-9 |access-date=5 December 2021 |archive-date=5 December 2021 |archive-url=https://web.archive.org/web/20211205135022/https://brianchristian.org/the-alignment-problem/ |url-status=live }}</ref><ref>{{cite journal |last1=Dignum |first1=Virginia |date=26 May 2021 |title=AI – the people and places that make, use and manage it |journal=Nature |language=en-us |volume=593 |issue=7860 |pages=499–500 |bibcode=2021Natur.593..499D |doi=10.1038/d41586-021-01397-x |s2cid=235216649 |doi-access=free}}</ref>
[[Nick Bostrom]] published ''[[Superintelligence: Paths, Dangers, Strategies|Superintelligence]]'' in 2014, which presented his arguments that superintelligence poses an existential threat.<ref>{{Cite news |last=Metz |first=Cade |date=9 June 2018 |title=Mark Zuckerberg, Elon Musk and the Feud Over Killer Robots |work=The New York Times |url=https://www.nytimes.com/2018/06/09/technology/elon-musk-mark-zuckerberg-artificial-intelligence.html |access-date=3 April 2019 |archive-date=15 February 2021 |archive-url=https://web.archive.org/web/20210215051949/https://www.nytimes.com/2018/06/09/technology/elon-musk-mark-zuckerberg-artificial-intelligence.html |url-status=live }}</ref> By 2015, public figures such as physicists [[Stephen Hawking]] and Nobel laureate [[Frank Wilczek]], computer scientists [[Stuart J. Russell]] and [[Roman Yampolskiy]], and entrepreneurs [[Elon Musk]] and [[Bill Gates]] were expressing concern about the risks of superintelligence.<ref>{{Cite news |last=Hsu |first=Jeremy |date=1 March 2012 |title=Control dangerous AI before it controls us, one expert says |work=[[NBC News]] |url=http://www.nbcnews.com/id/46590591/ns/technology_and_science-innovation |access-date=28 January 2016 |archive-date=2 February 2016 |archive-url=https://web.archive.org/web/20160202173621/http://www.nbcnews.com/id/46590591/ns/technology_and_science-innovation |url-status=live }}</ref><ref name="hawking editorial">{{Cite news |title=Stephen Hawking: 'Transcendence looks at the implications of artificial intelligence&nbsp;– but are we taking AI seriously enough?' |publisher=[[The Independent (UK)]] |url=https://www.independent.co.uk/news/science/stephen-hawking-transcendence-looks-at-the-implications-of-artificial-intelligence--but-are-we-taking-ai-seriously-enough-9313474.html |url-status=live |access-date=3 December 2014 |archive-url=https://web.archive.org/web/20150925153716/http://www.independent.co.uk/news/science/stephen-hawking-transcendence-looks-at-the-implications-of-artificial-intelligence--but-are-we-taking-ai-seriously-enough-9313474.html |archive-date=25 September 2015}}</ref><ref name="bbc on hawking editorial">{{Cite news |date=2 December 2014 |title=Stephen Hawking warns artificial intelligence could end mankind |publisher=[[BBC]] |url=https://www.bbc.com/news/technology-30290540 |url-status=live |access-date=3 December 2014 |archive-url=https://web.archive.org/web/20151030054329/http://www.bbc.com/news/technology-30290540 |archive-date=30 October 2015}}</ref><ref>{{Cite news |last=Eadicicco |first=Lisa |date=28 January 2015 |title=Bill Gates: Elon Musk Is Right, We Should All Be Scared Of Artificial Intelligence Wiping Out Humanity |work=[[Business Insider]] |url=http://www.businessinsider.com/bill-gates-artificial-intelligence-2015-1 |access-date=30 January 2016 |archive-date=26 February 2016 |archive-url=https://web.archive.org/web/20160226090602/http://www.businessinsider.com/bill-gates-artificial-intelligence-2015-1 |url-status=live }}</ref> Also in 2015, the [[Open Letter on Artificial Intelligence]] highlighted the "great potential of AI" and encouraged more research on how to make it robust and beneficial.<ref>{{Cite web |title=Research Priorities for Robust and Beneficial Artificial Intelligence: an Open Letter |url=http://futureoflife.org/misc/open_letter |url-status=live |archive-url=https://web.archive.org/web/20150115160823/http://futureoflife.org/misc/open_letter |archive-date=15 January 2015 |access-date=23 October 2015 |publisher=[[Future of Life Institute]]}}</ref> In April 2016, the journal ''[[Nature (journal)|Nature]]'' warned: "Machines and robots that outperform humans across the board could self-improve beyond our control—and their interests might not align with ours".<ref>{{Cite journal |date=2016 |title=Anticipating artificial intelligence |journal=Nature |language=en |volume=532 |issue=7600 |page=413 |doi=10.1038/532413a |pmid=27121801 |bibcode=2016Natur.532Q.413. |s2cid=4399193 |issn=1476-4687|doi-access=free }}</ref> In 2020, [[Brian Christian]] published ''[[The Alignment Problem]]'', which details the history of progress on AI alignment up to that time.<ref>{{cite book |last=Christian |first=Brian |date=October 6, 2020 |title=The Alignment Problem: Machine Learning and Human Values |url=https://brianchristian.org/the-alignment-problem/ |publisher=[[W. W. Norton & Company]] |isbn=978-0-393-63582-9 |access-date=5 December 2021 |archive-date=5 December 2021 |archive-url=https://web.archive.org/web/20211205135022/https://brianchristian.org/the-alignment-problem/ |url-status=live }}</ref><ref>{{cite journal |last1=Dignum |first1=Virginia |date=26 May 2021 |title=AI – the people and places that make, use and manage it |journal=Nature |language=en-us |volume=593 |issue=7860 |pages=499–500 |bibcode=2021Natur.593..499D |doi=10.1038/d41586-021-01397-x |s2cid=235216649 |doi-access=free}}</ref>


In March 2023, key figures in AI, such as Musk, signed a letter from the [[Future of Life Institute]] calling a halt to advanced AI training until it could be properly regulated.<ref>{{Cite news |date=2023-03-29 |title=Elon Musk among experts urging a halt to AI training |language=en-GB |work=BBC News |url=https://www.bbc.com/news/technology-65110030 |access-date=2023-06-09}}</ref> In May 2023, the [[Center for AI Safety]] released a statement signed by numerous experts in AI safety and the AI existential risk which stated: "Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war".<ref>{{Cite web |title=Statement on AI Risk |url=https://www.safe.ai/statement-on-ai-risk#open-letter |access-date=2023-06-08 |website=Center for AI Safety}}</ref><ref>{{Cite news |date=2023-05-30 |title=Artificial intelligence could lead to extinction, experts warn |language=en-GB |work=BBC News |url=https://www.bbc.com/news/uk-65746524 |access-date=2023-06-08}}</ref>
In March 2023, key figures in AI, such as Musk, signed a letter from the [[Future of Life Institute]] calling a halt to advanced AI training until it could be properly regulated.<ref>{{Cite news |date=2023-03-29 |title=Elon Musk among experts urging a halt to AI training |language=en-GB |work=BBC News |url=https://www.bbc.com/news/technology-65110030 |access-date=2023-06-09}}</ref> In May 2023, the [[Center for AI Safety]] released a statement signed by numerous experts in AI safety and the AI existential risk which stated: "Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war."<ref>{{Cite web |title=Statement on AI Risk |url=https://www.safe.ai/statement-on-ai-risk#open-letter |access-date=2023-06-08 |website=Center for AI Safety}}</ref><ref>{{Cite news |date=2023-05-30 |title=Artificial intelligence could lead to extinction, experts warn |language=en-GB |work=BBC News |url=https://www.bbc.com/news/uk-65746524 |access-date=2023-06-08}}</ref>


== AI capabilities ==
== AI capabilities ==
Line 96: Line 96:


== Types of existential risk ==
== Types of existential risk ==
[[File:X-risk-chart-en-01a.svg|thumb|upright=1.9|Scope–severity grid from Bostrom's paper "Existential Risk Prevention as Global Priority"<ref name="priority">{{Cite journal |last=Bostrom |first=Nick |year=2013 |title=Existential Risk Prevention as Global Priority |url=http://www.existential-risk.org/concept.pdf |journal=Global Policy |volume=4 |issue=1 |pages=15–3 |doi=10.1111/1758-5899.12002 |via=Existential Risk}}</ref>]]
[[File:X-risk-chart-en-01a.svg|thumb|upright=1.5|Scope–severity grid from Bostrom's paper "Existential Risk Prevention as Global Priority"<ref name="priority">{{Cite journal |last=Bostrom |first=Nick |year=2013 |title=Existential Risk Prevention as Global Priority |url=http://www.existential-risk.org/concept.pdf |journal=Global Policy |volume=4 |issue=1 |pages=15–3 |doi=10.1111/1758-5899.12002 |via=Existential Risk}}</ref>]]
An existential risk is "one that threatens the premature extinction of Earth-originating intelligent life or the permanent and drastic destruction of its potential for desirable future development".<ref>{{Cite news |last=Doherty |first=Ben |date=2018-05-17 |title=Climate change an 'existential security risk' to Australia, Senate inquiry says |language=en-GB |work=The Guardian |url=https://www.theguardian.com/environment/2018/may/18/climate-change-an-existential-security-risk-to-australia-senate-inquiry-says |access-date=2023-07-16 |issn=0261-3077}}</ref>
An existential risk is "one that threatens the premature extinction of Earth-originating intelligent life or the permanent and drastic destruction of its potential for desirable future development".<ref>{{Cite news |last=Doherty |first=Ben |date=2018-05-17 |title=Climate change an 'existential security risk' to Australia, Senate inquiry says |language=en-GB |work=The Guardian |url=https://www.theguardian.com/environment/2018/may/18/climate-change-an-existential-security-risk-to-australia-senate-inquiry-says |access-date=2023-07-16 |issn=0261-3077}}</ref>


Line 115: Line 115:


An [[Instrumental and intrinsic value|"instrumental" goal]] is a sub-goal that helps to achieve an agent's ultimate goal. "Instrumental convergence" refers to the fact that some sub-goals are useful for achieving virtually ''any'' ultimate goal, such as acquiring resources or self-preservation.<ref name="omohundro">Omohundro, S. M. (2008, February). The basic AI drives. In ''AGI'' (Vol. 171, pp. 483–492).</ref> Bostrom argues that if an advanced AI's instrumental goals conflict with humanity's goals, the AI might harm humanity in order to acquire more resources or prevent itself from being shut down, but only as a way to achieve its ultimate goal.<ref name="superintelligence" />
An [[Instrumental and intrinsic value|"instrumental" goal]] is a sub-goal that helps to achieve an agent's ultimate goal. "Instrumental convergence" refers to the fact that some sub-goals are useful for achieving virtually ''any'' ultimate goal, such as acquiring resources or self-preservation.<ref name="omohundro">Omohundro, S. M. (2008, February). The basic AI drives. In ''AGI'' (Vol. 171, pp. 483–492).</ref> Bostrom argues that if an advanced AI's instrumental goals conflict with humanity's goals, the AI might harm humanity in order to acquire more resources or prevent itself from being shut down, but only as a way to achieve its ultimate goal.<ref name="superintelligence" />
[[File:Power-Seeking_Image.png|thumb|upright=1.9|Some ways in which an advanced misaligned AI could try to gain more power.<ref name="Carlsmith2022">{{cite arXiv |eprint=2206.13353 |class=cs.CY |first=Joseph |last=Carlsmith |title=Is Power-Seeking AI an Existential Risk? |date=2022-06-16}}</ref> Power-seeking behaviors may arise because power is useful to accomplish virtually any objective.<ref>{{Cite web |title='The Godfather of A.I.' warns of 'nightmare scenario' where artificial intelligence begins to seek power |url=https://fortune.com/2023/05/02/godfather-ai-geoff-hinton-google-warns-artificial-intelligence-nightmare-scenario/ |access-date=2023-06-10 |website=Fortune |language=en}}</ref>]]
[[File:Power-Seeking_Image.png|thumb|upright=1.5|Some ways in which an advanced misaligned AI could try to gain more power.<ref name="Carlsmith2022">{{cite arXiv |eprint=2206.13353 |class=cs.CY |first=Joseph |last=Carlsmith |title=Is Power-Seeking AI an Existential Risk? |date=2022-06-16}}</ref> Power-seeking behaviors may arise because power is useful to accomplish virtually any objective.<ref>{{Cite web |title='The Godfather of A.I.' warns of 'nightmare scenario' where artificial intelligence begins to seek power |url=https://fortune.com/2023/05/02/godfather-ai-geoff-hinton-google-warns-artificial-intelligence-nightmare-scenario/ |access-date=2023-06-10 |website=Fortune |language=en}}</ref>]]
[[Stuart J. Russell|Russell]] argues that a sufficiently advanced machine "will have self-preservation even if you don't program it in... if you say, 'Fetch the coffee', it can't fetch the coffee if it's dead. So if you give it any goal whatsoever, it has a reason to preserve its own existence to achieve that goal."<ref name="vanity" /><ref>{{Cite news |last=Wakefield |first=Jane |date=15 September 2015 |title=Why is Facebook investing in AI? |work=BBC News |url=https://www.bbc.com/news/technology-34118481 |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20171202192942/http://www.bbc.com/news/technology-34118481 |archive-date=2 December 2017}}</ref>
[[Stuart J. Russell|Russell]] argues that a sufficiently advanced machine "will have self-preservation even if you don't program it in... if you say, 'Fetch the coffee', it can't fetch the coffee if it's dead. So if you give it any goal whatsoever, it has a reason to preserve its own existence to achieve that goal."<ref name="vanity" /><ref>{{Cite news |last=Wakefield |first=Jane |date=15 September 2015 |title=Why is Facebook investing in AI? |work=BBC News |url=https://www.bbc.com/news/technology-34118481 |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20171202192942/http://www.bbc.com/news/technology-34118481 |archive-date=2 December 2017}}</ref>


Line 161: Line 161:
The academic debate is between those who worry that AI might threaten humanity and those who believe it would not. Both sides of this debate have framed the other side's arguments as illogical anthropomorphism.<ref name="yudkowsky-global-risk" /> Those skeptical of AGI risk accuse their opponents of anthropomorphism for assuming that an AGI would naturally desire power; those concerned about AGI risk accuse skeptics of anthropomorphism for believing an AGI would naturally value or infer human ethical norms.<ref name="yudkowsky-global-risk" /><ref name="Telegraph2016">{{Cite news |date=1 Sep 2015 |title=Should humans fear the rise of the machine? |work=[[The Telegraph (UK)]] |url=https://www.telegraph.co.uk/technology/news/11837157/Should-humans-fear-the-rise-of-the-machine.html |archive-url=https://ghostarchive.org/archive/20220112/https://www.telegraph.co.uk/technology/news/11837157/Should-humans-fear-the-rise-of-the-machine.html |archive-date=12 January 2022 |url-access=subscription |url-status=live |access-date=7 February 2016}}{{cbignore}}</ref>
The academic debate is between those who worry that AI might threaten humanity and those who believe it would not. Both sides of this debate have framed the other side's arguments as illogical anthropomorphism.<ref name="yudkowsky-global-risk" /> Those skeptical of AGI risk accuse their opponents of anthropomorphism for assuming that an AGI would naturally desire power; those concerned about AGI risk accuse skeptics of anthropomorphism for believing an AGI would naturally value or infer human ethical norms.<ref name="yudkowsky-global-risk" /><ref name="Telegraph2016">{{Cite news |date=1 Sep 2015 |title=Should humans fear the rise of the machine? |work=[[The Telegraph (UK)]] |url=https://www.telegraph.co.uk/technology/news/11837157/Should-humans-fear-the-rise-of-the-machine.html |archive-url=https://ghostarchive.org/archive/20220112/https://www.telegraph.co.uk/technology/news/11837157/Should-humans-fear-the-rise-of-the-machine.html |archive-date=12 January 2022 |url-access=subscription |url-status=live |access-date=7 February 2016}}{{cbignore}}</ref>


Evolutionary psychologist [[Steven Pinker]], a skeptic, argues that "AI dystopias project a parochial alpha-male psychology onto the concept of intelligence. They assume that superhumanly intelligent robots would develop goals like deposing their masters or taking over the world"; perhaps instead "artificial intelligence will naturally develop along female lines: fully capable of solving problems, but with no desire to annihilate innocents or dominate the civilization."<ref name=shermer/> Facebook's director of AI research, [[Yann LeCun]], has said: "Humans have all kinds of drives that make them do bad things to each other, like the self-preservation instinct... Those drives are programmed into our brain but there is absolutely no reason to build robots that have the same kind of drives".<ref>{{cite news |title=Intelligent Machines: What does Facebook want with AI? |url=https://www.bbc.com/news/technology-34118481 |access-date=31 March 2023 |work=BBC News |date=14 September 2015}}</ref>
Evolutionary psychologist [[Steven Pinker]], a skeptic, argues that "AI dystopias project a (this section is sexist please fix it
) fully capable of solving problems, but with no desire to annihilate innocents or dominate the civilization."<ref name=shermer/> Facebook's director of AI research, [[Yann LeCun]], has said: "Humans have all kinds of drives that make them do bad things to each other, like the self-preservation instinct... Those drives are programmed into our brain but there is absolutely no reason to build robots that have the same kind of drives".<ref>{{cite news |title=Intelligent Machines: What does Facebook want with AI? |url=https://www.bbc.com/news/technology-34118481 |access-date=31 March 2023 |work=BBC News |date=14 September 2015}}</ref>


Despite other differences, the x-risk school{{efn|as interpreted by [[Seth Baum]]}} agrees with Pinker that an advanced AI would not destroy humanity out of emotion such as revenge or anger, that questions of consciousness are not relevant to assess the risk,<ref name="auto">{{Cite journal |last=Baum |first=Seth |date=2018-09-30 |title=Countering Superintelligence Misinformation |journal=Information |volume=9 |issue=10 |page=244 |doi=10.3390/info9100244 |issn=2078-2489 |doi-access=free}}</ref> and that computer systems do not generally have a computational equivalent of testosterone.<ref>{{Cite web |title=The Myth Of AI |url=https://www.edge.org/conversation/jaron_lanier-the-myth-of-ai |access-date=2020-03-11 |website=www.edge.org |archive-date=11 March 2020 |archive-url=https://web.archive.org/web/20200311210407/https://www.edge.org/conversation/jaron_lanier-the-myth-of-ai |url-status=live }}</ref> They think that power-seeking or self-preservation behaviors emerge in the AI as a way to achieve its true goals, according to the concept of [[instrumental convergence]].
Despite other differences, the x-risk school{{efn|as interpreted by [[Seth Baum]]}} agrees with Pinker that an advanced AI would not destroy humanity out of emotion such as revenge or anger, that questions of consciousness are not relevant to assess the risk,<ref name="auto">{{Cite journal |last=Baum |first=Seth |date=2018-09-30 |title=Countering Superintelligence Misinformation |journal=Information |volume=9 |issue=10 |page=244 |doi=10.3390/info9100244 |issn=2078-2489 |doi-access=free}}</ref> and that computer systems do not generally have a computational equivalent of testosterone.<ref>{{Cite web |title=The Myth Of AI |url=https://www.edge.org/conversation/jaron_lanier-the-myth-of-ai |access-date=2020-03-11 |website=www.edge.org |archive-date=11 March 2020 |archive-url=https://web.archive.org/web/20200311210407/https://www.edge.org/conversation/jaron_lanier-the-myth-of-ai |url-status=live }}</ref> They think that power-seeking or self-preservation behaviors emerge in the AI as a way to achieve its true goals, according to the concept of [[instrumental convergence]].
Line 178: Line 177:


=== Life 3.0 ===
=== Life 3.0 ===
In [[Max Tegmark]]'s 2017 book ''[[Life 3.0]]'', a corporation's "Omega team" creates an extremely powerful AI able to moderately improve its own source code in a number of areas. After a certain point, the team chooses to publicly downplay the AI's ability in order to avoid regulation or confiscation of the project. For safety, the team keeps the AI [[AI box|in a box]] where it is mostly unable to communicate with the outside world, and uses it to make money, by diverse means such as [[Amazon Mechanical Turk]] tasks, production of animated films and TV shows, and development of biotech drugs, with profits invested back into further improving AI. The team next tasks the AI with [[astroturfing]] an army of pseudonymous citizen journalists and commentators in order to gain political influence to use "for the greater good" to prevent wars. The team faces risks that the AI could try to escape by inserting "backdoors" in the systems it designs, by [[steganography|hidden messages]] in its produced content, or by using its growing understanding of human behavior to [[Social engineering (security)|persuade someone into letting it free]]. The team also faces risks that its decision to box the project will delay the project long enough for another project to overtake it.<ref>{{Cite journal |last=Russell |first=Stuart |date=30 August 2017 |title=Artificial intelligence: The future is superintelligent |journal=Nature |language=En |volume=548 |issue=7669 |pages=520–521 |bibcode=2017Natur.548..520R |doi=10.1038/548520a |s2cid=4459076 |doi-access=free}}</ref><ref name="life 3.0" /><!-- Prelude and Chapter 4 -->
In [[Max Tegmark]]'s 2017 book ''[[Life 3.0]]'', a corporation's "Omega team" creates an extremely powerful AI able to moderately improve its own source code in a number of areas. After a certain point, the team chooses to publicly downplay the AI's ability in order to avoid regulation or confiscation of the project. For safety, the team keeps the AI [[AI capability control|in a box]] where it is mostly unable to communicate with the outside world, and uses it to make money, by diverse means such as [[Amazon Mechanical Turk]] tasks, production of animated films and TV shows, and development of biotech drugs, with profits invested back into further improving AI. The team next tasks the AI with [[astroturfing]] an army of pseudonymous citizen journalists and commentators in order to gain political influence to use "for the greater good" to prevent wars. The team faces risks that the AI could try to escape by inserting "backdoors" in the systems it designs, by [[steganography|hidden messages]] in its produced content, or by using its growing understanding of human behavior to [[Social engineering (security)|persuade someone into letting it free]]. The team also faces risks that its decision to box the project will delay the project long enough for another project to overtake it.<ref>{{Cite journal |last=Russell |first=Stuart |date=30 August 2017 |title=Artificial intelligence: The future is superintelligent |journal=Nature |language=En |volume=548 |issue=7669 |pages=520–521 |bibcode=2017Natur.548..520R |doi=10.1038/548520a |s2cid=4459076 |doi-access=free}}</ref><ref name="life 3.0" /><!-- Prelude and Chapter 4 -->


== Perspectives ==
== Perspectives ==
Line 207: Line 206:
{{Further|Artificial general intelligence#Feasibility}}[[Baidu]] Vice President [[Andrew Ng]] said in 2015 that AI existential risk is "like worrying about overpopulation on Mars when we have not even set foot on the planet yet."<ref name="shermer">{{Cite journal |last=Shermer |first=Michael |date=1 March 2017 |title=Apocalypse AI |language=en |page=77 |journal=Scientific American |volume=316 |issue=3 |url=https://www.scientificamerican.com/article/artificial-intelligence-is-not-a-threat-mdash-yet/ |access-date=27 November 2017 |bibcode=2017SciAm.316c..77S |doi=10.1038/scientificamerican0317-77 |pmid=28207698 |archive-date=1 December 2017 |archive-url=https://web.archive.org/web/20171201051401/https://www.scientificamerican.com/article/artificial-intelligence-is-not-a-threat-mdash-yet/ |url-status=live }}</ref><ref>{{cite magazine |last1=Garling |first1=Caleb |title=Andrew Ng: Why 'Deep Learning' Is a Mandate for Humans, Not Just Machines |url=https://www.wired.com/brandlab/2015/05/andrew-ng-deep-learning-mandate-humans-not-just-machines/ |access-date=31 March 2023 |magazine=Wired |date=5 May 2015}}</ref> For the danger of uncontrolled advanced AI to be realized, the hypothetical AI may have to overpower or outthink any human, which some experts argue is a possibility far enough in the future to not be worth researching.<ref>{{cite web |date=2023-04-04 |title=Is artificial intelligence really an existential threat to humanity? |url=https://mambapost.com/2023/04/tech-news/ai-are-an-existential-threat-to-humanity/ |website=MambaPost}}</ref><ref>{{cite web |title=The case against killer robots, from a guy actually working on artificial intelligence |url=http://fusion.net/story/54583/the-case-against-killer-robots-from-a-guy-actually-building-ai/ |url-status=live |archive-url=https://web.archive.org/web/20160204175716/http://fusion.net/story/54583/the-case-against-killer-robots-from-a-guy-actually-building-ai/ |archive-date=4 February 2016 |access-date=31 January 2016 |website=Fusion.net}}</ref>
{{Further|Artificial general intelligence#Feasibility}}[[Baidu]] Vice President [[Andrew Ng]] said in 2015 that AI existential risk is "like worrying about overpopulation on Mars when we have not even set foot on the planet yet."<ref name="shermer">{{Cite journal |last=Shermer |first=Michael |date=1 March 2017 |title=Apocalypse AI |language=en |page=77 |journal=Scientific American |volume=316 |issue=3 |url=https://www.scientificamerican.com/article/artificial-intelligence-is-not-a-threat-mdash-yet/ |access-date=27 November 2017 |bibcode=2017SciAm.316c..77S |doi=10.1038/scientificamerican0317-77 |pmid=28207698 |archive-date=1 December 2017 |archive-url=https://web.archive.org/web/20171201051401/https://www.scientificamerican.com/article/artificial-intelligence-is-not-a-threat-mdash-yet/ |url-status=live }}</ref><ref>{{cite magazine |last1=Garling |first1=Caleb |title=Andrew Ng: Why 'Deep Learning' Is a Mandate for Humans, Not Just Machines |url=https://www.wired.com/brandlab/2015/05/andrew-ng-deep-learning-mandate-humans-not-just-machines/ |access-date=31 March 2023 |magazine=Wired |date=5 May 2015}}</ref> For the danger of uncontrolled advanced AI to be realized, the hypothetical AI may have to overpower or outthink any human, which some experts argue is a possibility far enough in the future to not be worth researching.<ref>{{cite web |date=2023-04-04 |title=Is artificial intelligence really an existential threat to humanity? |url=https://mambapost.com/2023/04/tech-news/ai-are-an-existential-threat-to-humanity/ |website=MambaPost}}</ref><ref>{{cite web |title=The case against killer robots, from a guy actually working on artificial intelligence |url=http://fusion.net/story/54583/the-case-against-killer-robots-from-a-guy-actually-building-ai/ |url-status=live |archive-url=https://web.archive.org/web/20160204175716/http://fusion.net/story/54583/the-case-against-killer-robots-from-a-guy-actually-building-ai/ |archive-date=4 February 2016 |access-date=31 January 2016 |website=Fusion.net}}</ref>


Skeptics who believe AGI is not a short-term possibility often argue that concern about existential risk from AI is unhelpful because it could distract people from more immediate concerns about AI's impact, because it could lead to government regulation or make it more difficult to fund AI research, or because it could damage the field's reputation.<ref>{{Cite web |date=2023-05-31 |title=AI experts challenge 'doomer' narrative, including 'extinction risk' claims |url=https://venturebeat.com/ai/ai-experts-challenge-doomer-narrative-including-extinction-risk-claims/ |access-date=2023-07-08 |website=VentureBeat |language=en-US}}</ref> AI and AI ethics researchers [[Timnit Gebru]], [[Emily M. Bender]], [[Margaret Mitchell (scientist)|Margaret Mitchell]], and Angelina McMillan-Major have argued that discussion of existential risk distracts from the immediate, ongoing harms from AI taking place today, such as data theft, worker exploitation, bias, and concentration of power.<ref>{{Cite web |last=Coldewey |first=Devin |date=2023-04-01 |title=Ethicists fire back at 'AI Pause' letter they say 'ignores the actual harms' |url=https://techcrunch.com/2023/03/31/ethicists-fire-back-at-ai-pause-letter-they-say-ignores-the-actual-harms/ |access-date=2023-07-23 |website=TechCrunch |language=en-US}}</ref> They further note the association between those warning of existential risk and [[longtermism]], which they describe as a "dangerous ideology" for its unscientific and utopian nature.<ref>{{Cite web |title=DAIR (Distributed AI Research Institute) |url=https://dair-institute.org/ |access-date=2023-07-23 |website=DAIR Institute |language=en}}</ref>
Skeptics who believe AGI is not a short-term possibility often argue that concern about existential risk from AI is unhelpful because it could distract people from more immediate concerns about AI's impact, because it could lead to government regulation or make it more difficult to fund AI research, or because it could damage the field's reputation.<ref>{{Cite web |date=2023-05-31 |title=AI experts challenge 'doomer' narrative, including 'extinction risk' claims |url=https://venturebeat.com/ai/ai-experts-challenge-doomer-narrative-including-extinction-risk-claims/ |access-date=2023-07-08 |website=VentureBeat |language=en-US}}</ref> AI and AI ethics researchers [[Timnit Gebru]], [[Emily M. Bender]], [[Margaret Mitchell (scientist)|Margaret Mitchell]], and Angelina McMillan-Major have argued that discussion of existential risk distracts from the immediate, ongoing harms from AI taking place today, such as data theft, worker exploitation, bias, and concentration of power.<ref>{{Cite web |last=Coldewey |first=Devin |date=2023-04-01 |title=Ethicists fire back at 'AI Pause' letter they say 'ignores the actual harms' |url=https://techcrunch.com/2023/03/31/ethicists-fire-back-at-ai-pause-letter-they-say-ignores-the-actual-harms/ |access-date=2023-07-23 |website=TechCrunch |language=en-US}}</ref> They further note the association between those warning of existential risk and [[longtermism]], which they describe as a "dangerous ideology" for its unscientific and utopian nature.<ref>{{Cite web |title=DAIR (Distributed AI Research Institute) |url=https://dair-institute.org/ |access-date=2023-07-23 |website=[[DAIR Institute]] |language=en}}</ref> Gebru and [[Émile P. Torres]] have suggested that obsession with AGI is part of a pattern of intellectual movements called [[TESCREAL|TESCREAL.]]<ref>{{Cite journal |last1=Gebru |first1=Timnit |last2=Torres |first2=Émile P. |date=2024-04-14 |title=The TESCREAL bundle: Eugenics and the promise of utopia through artificial general intelligence |url=https://firstmonday.org/ojs/index.php/fm/article/view/13636 |journal=First Monday |language=en |doi=10.5210/fm.v29i4.13636 |doi-access=free |issn=1396-0466}}</ref>


[[Wired (magazine)|''Wired'']] editor [[Kevin Kelly (editor)|Kevin Kelly]] argues that natural intelligence is more nuanced than AGI proponents believe, and that intelligence alone is not enough to achieve major scientific and societal breakthroughs. He argues that intelligence consists of many dimensions that are not well understood, and that conceptions of an 'intelligence ladder' are misleading. He notes the crucial role real-world experiments play in the scientific method, and that intelligence alone is no substitute for these.<ref>{{Cite magazine|last=Kelly|first=Kevin|author-link=Kevin Kelly (editor)|date=April 25, 2017|title=The Myth of a Superhuman AI|url=https://www.wired.com/2017/04/the-myth-of-a-superhuman-ai/|archive-url=https://web.archive.org/web/20211226181932/https://www.wired.com/2017/04/the-myth-of-a-superhuman-ai/|archive-date=December 26, 2021|access-date=February 19, 2022|magazine=Wired}}</ref>
[[Wired (magazine)|''Wired'']] editor [[Kevin Kelly (editor)|Kevin Kelly]] argues that natural intelligence is more nuanced than AGI proponents believe, and that intelligence alone is not enough to achieve major scientific and societal breakthroughs. He argues that intelligence consists of many dimensions that are not well understood, and that conceptions of an 'intelligence ladder' are misleading. He notes the crucial role real-world experiments play in the scientific method, and that intelligence alone is no substitute for these.<ref>{{Cite magazine|last=Kelly|first=Kevin|author-link=Kevin Kelly (editor)|date=April 25, 2017|title=The Myth of a Superhuman AI|url=https://www.wired.com/2017/04/the-myth-of-a-superhuman-ai/|archive-url=https://web.archive.org/web/20211226181932/https://www.wired.com/2017/04/the-myth-of-a-superhuman-ai/|archive-date=December 26, 2021|access-date=February 19, 2022|magazine=Wired}}</ref>
Line 228: Line 227:
In 2018, a [[SurveyMonkey]] poll of the American public by ''[[USA Today]]'' found 68% thought the real current threat remains "human intelligence", but also found that 43% said superintelligent AI, if it were to happen, would result in "more harm than good", and that 38% said it would do "equal amounts of harm and good".<ref>{{Cite news |date=5 January 2023 |title=Elon Musk says AI could doom human civilization. Zuckerberg disagrees. Who's right? |url=https://www.usatoday.com/story/tech/news/2018/01/02/artificial-intelligence-end-world-overblown-fears/985813001/ |url-status=live |access-date=8 January 2018 |archive-url=https://web.archive.org/web/20180108075432/https://www.usatoday.com/story/tech/news/2018/01/02/artificial-intelligence-end-world-overblown-fears/985813001/ |archive-date=8 January 2018}}</ref>
In 2018, a [[SurveyMonkey]] poll of the American public by ''[[USA Today]]'' found 68% thought the real current threat remains "human intelligence", but also found that 43% said superintelligent AI, if it were to happen, would result in "more harm than good", and that 38% said it would do "equal amounts of harm and good".<ref>{{Cite news |date=5 January 2023 |title=Elon Musk says AI could doom human civilization. Zuckerberg disagrees. Who's right? |url=https://www.usatoday.com/story/tech/news/2018/01/02/artificial-intelligence-end-world-overblown-fears/985813001/ |url-status=live |access-date=8 January 2018 |archive-url=https://web.archive.org/web/20180108075432/https://www.usatoday.com/story/tech/news/2018/01/02/artificial-intelligence-end-world-overblown-fears/985813001/ |archive-date=8 January 2018}}</ref>


An April 2023 [[YouGov]] poll of US adults found 46% of respondents were "somewhat concerned" or "very concerned" about "the possibility that AI will cause the end of the human race on Earth," compared with 40% who were "not very concerned" or "not at all concerned."<ref>{{Cite news |date=14 April 2023 |title=AI doomsday worries many Americans. So does apocalypse from climate change, nukes, war, and more |url=https://today.yougov.com/topics/technology/articles-reports/2023/04/14/ai-nuclear-weapons-world-war-humanity-poll |url-status=live |access-date=9 July 2023 |archive-url=https://web.archive.org/web/20230623095224/https://today.yougov.com/topics/technology/articles-reports/2023/04/14/ai-nuclear-weapons-world-war-humanity-poll |archive-date=23 June 2023}}</ref>
An April 2023 [[YouGov]] poll of US adults found 46% of respondents were "somewhat concerned" or "very concerned" about "the possibility that AI will cause the end of the human race on Earth", compared with 40% who were "not very concerned" or "not at all concerned."<ref>{{Cite news |date=14 April 2023 |title=AI doomsday worries many Americans. So does apocalypse from climate change, nukes, war, and more |url=https://today.yougov.com/topics/technology/articles-reports/2023/04/14/ai-nuclear-weapons-world-war-humanity-poll |url-status=live |access-date=9 July 2023 |archive-url=https://web.archive.org/web/20230623095224/https://today.yougov.com/topics/technology/articles-reports/2023/04/14/ai-nuclear-weapons-world-war-humanity-poll |archive-date=23 June 2023}}</ref>


According to an August 2023 survey by the Pew Research Centers, 52% of Americans felt more concerned than excited about new AI developments; nearly a third felt as equally concerned and excited. More Americans saw that AI would have a more helpful than hurtful impact on several areas, from healthcare and vehicle safety to product search and customer service. The main exception is privacy: 53% of Americans believe AI will lead to higher exposure of their personal information.<ref>{{Cite web |last1=Tyson |first1=Alec |last2=Kikuchi |first2=Emma |title=Growing public concern about the role of artificial intelligence in daily life |url=https://www.pewresearch.org/short-reads/2023/08/28/growing-public-concern-about-the-role-of-artificial-intelligence-in-daily-life/ |access-date=2023-09-17 |website=Pew Research Center |date=28 August 2023 |language=en-US}}</ref>
According to an August 2023 survey by the Pew Research Centers, 52% of Americans felt more concerned than excited about new AI developments; nearly a third felt as equally concerned and excited. More Americans saw that AI would have a more helpful than hurtful impact on several areas, from healthcare and vehicle safety to product search and customer service. The main exception is privacy: 53% of Americans believe AI will lead to higher exposure of their personal information.<ref>{{Cite web |last1=Tyson |first1=Alec |last2=Kikuchi |first2=Emma |title=Growing public concern about the role of artificial intelligence in daily life |url=https://www.pewresearch.org/short-reads/2023/08/28/growing-public-concern-about-the-role-of-artificial-intelligence-in-daily-life/ |access-date=2023-09-17 |website=Pew Research Center |date=28 August 2023 |language=en-US}}</ref>
Line 235: Line 234:
{{see also|AI alignment|Machine ethics|Friendly artificial intelligence|Regulation of artificial intelligence}}
{{see also|AI alignment|Machine ethics|Friendly artificial intelligence|Regulation of artificial intelligence}}


Many scholars concerned about the AGI existential risk believe that the best approach is to conduct substantial research into solving the difficult "control problem": what types of safeguards, algorithms, or architectures can programmers implement to maximize the probability that their recursively-improving AI would continue to behave in a friendly manner after it reaches superintelligence?<ref name="superintelligence" /><ref name="physica_scripta">{{Cite journal |last1=Sotala |first1=Kaj |last2=Yampolskiy |first2=Roman |author-link2=Roman Yampolskiy |date=19 December 2014 |title=Responses to catastrophic AGI risk: a survey |journal=[[Physica Scripta]] |volume=90 |issue=1}}</ref> Social measures may mitigate the AGI existential risk;<ref>{{Cite journal|last1=Barrett|first1=Anthony M.|last2=Baum|first2=Seth D.|date=2016-05-23|title=A model of pathways to artificial superintelligence catastrophe for risk and decision analysis|journal=Journal of Experimental & Theoretical Artificial Intelligence|volume=29|issue=2|pages=397–414|doi=10.1080/0952813x.2016.1186228|arxiv=1607.07730|s2cid=928824|issn=0952-813X}}</ref><ref>{{Cite journal|last1=Sotala|first1=Kaj|last2=Yampolskiy|first2=Roman V|date=2014-12-19|title=Responses to catastrophic AGI risk: a survey|journal=Physica Scripta|volume=90|issue=1|page=018001|doi=10.1088/0031-8949/90/1/018001|bibcode=2015PhyS...90a8001S |s2cid=4749656 |issn=0031-8949|doi-access=free}}</ref> for instance, one recommendation is for a UN-sponsored "Benevolent AGI Treaty" that would ensure only altruistic AGIs be created.<ref>{{Cite journal|last1=Ramamoorthy|first1=Anand|last2=Yampolskiy|first2=Roman|date=2018|title=Beyond MAD? The race for artificial general intelligence|url=https://www.itu.int/pub/S-JOURNAL-ICTS.V1I1-2018-9|journal=ICT Discoveries|publisher=ITU|volume=1|issue=Special Issue 1|pages=1–8|access-date=7 January 2022|archive-date=7 January 2022|archive-url=https://web.archive.org/web/20220107141537/https://www.itu.int/pub/S-JOURNAL-ICTS.V1I1-2018-9|url-status=live}}</ref> Similarly, an arms control approach has been suggested, as has a global peace treaty grounded in the international relations theory of conforming instrumentalism, with an artificial superintelligence potentially being a signatory.<ref>{{Cite journal|last1=Carayannis|first1=Elias G.|last2=Draper|first2=John|date=2022-01-11|title=Optimising peace through a Universal Global Peace Treaty to constrain the risk of war from a militarised artificial superintelligence|journal=AI & Society|volume=38 |issue=6 |pages=2679–2692|language=en|doi=10.1007/s00146-021-01382-y|pmid=35035113|pmc=8748529|s2cid=245877737|issn=0951-5666}}</ref><ref>{{Citation |last1=Carayannis |first1=Elias G. |title=The challenge of advanced cyberwar and the place of cyberpeace |date=2023-05-30 |work=The Elgar Companion to Digital Transformation, Artificial Intelligence and Innovation in the Economy, Society and Democracy |pages=32–80 |url=https://www.elgaronline.com/edcollchap/book/9781839109362/book-part-9781839109362-8.xml |access-date=2023-06-08 |publisher=Edward Elgar Publishing |doi=10.4337/9781839109362.00008 |isbn=978-1-83910-936-2 |last2=Draper |first2=John}}.</ref><!-- in physica_scripta, see sections 3.3.2. Encourage Research into Safe AGI, and 3.3.3. Differential Technological Progress -->
Many scholars concerned about AGI existential risk believe that extensive research into the "control problem" is essential. This problem involves determining which safeguards, algorithms, or architectures can be implemented to increase the likelihood that a recursively-improving AI remains friendly after achieving superintelligence.<ref name="superintelligence" /><ref name="physica_scripta">{{Cite journal |last1=Sotala |first1=Kaj |last2=Yampolskiy |first2=Roman |author-link2=Roman Yampolskiy |date=19 December 2014 |title=Responses to catastrophic AGI risk: a survey |journal=[[Physica Scripta]] |volume=90 |issue=1}}</ref> Social measures are also proposed to mitigate AGI risks,<ref>{{Cite journal|last1=Barrett|first1=Anthony M.|last2=Baum|first2=Seth D.|date=2016-05-23|title=A model of pathways to artificial superintelligence catastrophe for risk and decision analysis|journal=Journal of Experimental & Theoretical Artificial Intelligence|volume=29|issue=2|pages=397–414|doi=10.1080/0952813x.2016.1186228|arxiv=1607.07730|s2cid=928824|issn=0952-813X}}</ref><ref>{{Cite journal|last1=Sotala|first1=Kaj|last2=Yampolskiy|first2=Roman V|date=2014-12-19|title=Responses to catastrophic AGI risk: a survey|journal=Physica Scripta|volume=90|issue=1|page=018001|doi=10.1088/0031-8949/90/1/018001|bibcode=2015PhyS...90a8001S |s2cid=4749656 |issn=0031-8949|doi-access=free}}</ref> such as a UN-sponsored "Benevolent AGI Treaty" to ensure that only altruistic AGIs are created.<ref>{{Cite journal|last1=Ramamoorthy|first1=Anand|last2=Yampolskiy|first2=Roman|date=2018|title=Beyond MAD? The race for artificial general intelligence|url=https://www.itu.int/pub/S-JOURNAL-ICTS.V1I1-2018-9|journal=ICT Discoveries|publisher=ITU|volume=1|issue=Special Issue 1|pages=1–8|access-date=7 January 2022|archive-date=7 January 2022|archive-url=https://web.archive.org/web/20220107141537/https://www.itu.int/pub/S-JOURNAL-ICTS.V1I1-2018-9|url-status=live}}</ref> Additionally, an arms control approach and a global peace treaty grounded in [[international relations theory]] have been suggested, potentially for an artificial superintelligence to be a signatory.<ref>{{Cite journal|last1=Carayannis|first1=Elias G.|last2=Draper|first2=John|date=2022-01-11|title=Optimising peace through a Universal Global Peace Treaty to constrain the risk of war from a militarised artificial superintelligence|journal=AI & Society|volume=38 |issue=6 |pages=2679–2692|language=en|doi=10.1007/s00146-021-01382-y|pmid=35035113|pmc=8748529|s2cid=245877737|issn=0951-5666}}</ref><ref>{{Citation |last1=Carayannis |first1=Elias G. |title=The challenge of advanced cyberwar and the place of cyberpeace |date=2023-05-30 |work=The Elgar Companion to Digital Transformation, Artificial Intelligence and Innovation in the Economy, Society and Democracy |pages=32–80 |url=https://www.elgaronline.com/edcollchap/book/9781839109362/book-part-9781839109362-8.xml |access-date=2023-06-08 |publisher=Edward Elgar Publishing |doi=10.4337/9781839109362.00008 |isbn=978-1-83910-936-2 |last2=Draper |first2=John}}.</ref><!-- in physica_scripta, see sections 3.3.2. Encourage Research into Safe AGI, and 3.3.3. Differential Technological Progress -->


Researchers at Google have proposed research into general "AI safety" issues to simultaneously mitigate both short-term risks from narrow AI and long-term risks from AGI.<ref>{{Cite news |last=Vincent |first=James |date=22 June 2016 |title=Google's AI researchers say these are the five key problems for robot safety |language=en |work=The Verge |url=https://www.theverge.com/circuitbreaker/2016/6/22/11999664/google-robots-ai-safety-five-problems |access-date=5 April 2020 |archive-date=24 December 2019 |archive-url=https://web.archive.org/web/20191224201240/https://www.theverge.com/circuitbreaker/2016/6/22/11999664/google-robots-ai-safety-five-problems |url-status=live }}</ref><ref>Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. "Concrete problems in AI safety." arXiv preprint arXiv:1606.06565 (2016).</ref> A 2020 estimate places global spending on AI existential risk somewhere between $10 and $50 million, compared with global spending on AI around perhaps $40 billion.<!-- Precipice chapter=Chapter 2: Existential Risk|at=Footnote 55 --> Bostrom suggests that funding of protective technologies should be prioritized over potentially dangerous ones.<ref name=":5">{{Cite book |last=Ord |first=Toby |title=[[The Precipice: Existential Risk and the Future of Humanity]] |date=2020 |publisher=Bloomsbury Publishing Plc |isbn=978-1-5266-0019-6 |author-link=Toby Ord}}</ref><!-- Precipice chapter=Chapter 7: Safeguarding Humanity|at=Kindle loc 3327 --> Some funders, such as Musk, propose that radical [[human enhancement|human cognitive enhancement]] could be such a technology, for example direct neural linking between human and machine; others argue that enhancement technologies may themselves pose an existential risk.<ref>{{Cite news |last=Johnson |first=Alex |date=2019 |title=Elon Musk wants to hook your brain up directly to computers – starting next year |url=https://www.nbcnews.com/mach/tech/elon-musk-wants-hook-your-brain-directly-computers-starting-next-ncna1030631 |url-status=live |archive-url=https://web.archive.org/web/20200418094146/https://www.nbcnews.com/mach/tech/elon-musk-wants-hook-your-brain-directly-computers-starting-next-ncna1030631 |archive-date=18 April 2020 |access-date=5 April 2020 |work=NBC News |language=en-us}}</ref><ref>{{Cite news |last=Torres |first=Phil |date=18 September 2018 |title=Only Radically Enhancing Humanity Can Save Us All |language=en |work=Slate Magazine |url=https://slate.com/technology/2018/09/genetic-engineering-to-stop-doomsday.html |access-date=5 April 2020 |archive-date=6 August 2020 |archive-url=https://web.archive.org/web/20200806073520/https://slate.com/technology/2018/09/genetic-engineering-to-stop-doomsday.html |url-status=live }}</ref> Researchers could closely monitor or attempt to "box in" an initial AI at a risk of becoming too powerful. A dominant superintelligent AI, if aligned with human interests, might itself take action to mitigate the risk of takeover by rival AI, although the creation of the dominant AI could itself pose an existential risk.<ref>{{Cite journal |last1=Barrett |first1=Anthony M. |last2=Baum |first2=Seth D. |date=23 May 2016 |title=A model of pathways to artificial superintelligence catastrophe for risk and decision analysis |journal=Journal of Experimental & Theoretical Artificial Intelligence |volume=29 |issue=2 |pages=397–414 |arxiv=1607.07730 |doi=10.1080/0952813X.2016.1186228 |s2cid=928824}}</ref>
Researchers at Google have proposed research into general "AI safety" issues to simultaneously mitigate both short-term risks from narrow AI and long-term risks from AGI.<ref>{{Cite news |last=Vincent |first=James |date=22 June 2016 |title=Google's AI researchers say these are the five key problems for robot safety |language=en |work=The Verge |url=https://www.theverge.com/circuitbreaker/2016/6/22/11999664/google-robots-ai-safety-five-problems |access-date=5 April 2020 |archive-date=24 December 2019 |archive-url=https://web.archive.org/web/20191224201240/https://www.theverge.com/circuitbreaker/2016/6/22/11999664/google-robots-ai-safety-five-problems |url-status=live }}</ref><ref>Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. "Concrete problems in AI safety." arXiv preprint arXiv:1606.06565 (2016).</ref> A 2020 estimate places global spending on AI existential risk somewhere between $10 and $50 million, compared with global spending on AI around perhaps $40 billion.<!-- Precipice chapter=Chapter 2: Existential Risk|at=Footnote 55 --> Bostrom suggests prioritizing funding for protective technologies over potentially dangerous ones.<ref name=":5">{{Cite book |last=Ord |first=Toby |title=[[The Precipice: Existential Risk and the Future of Humanity]] |date=2020 |publisher=Bloomsbury Publishing Plc |isbn=978-1-5266-0019-6 |author-link=Toby Ord}}</ref><!-- Precipice chapter=Chapter 7: Safeguarding Humanity|at=Kindle loc 3327 --> Some, like Elon Musk, advocate radical [[human enhancement|human cognitive enhancement]], such as direct neural linking between humans and machines; others argue that these technologies may pose an existential risk themselves.<ref>{{Cite news |last=Johnson |first=Alex |date=2019 |title=Elon Musk wants to hook your brain up directly to computers – starting next year |url=https://www.nbcnews.com/mach/tech/elon-musk-wants-hook-your-brain-directly-computers-starting-next-ncna1030631 |url-status=live |archive-url=https://web.archive.org/web/20200418094146/https://www.nbcnews.com/mach/tech/elon-musk-wants-hook-your-brain-directly-computers-starting-next-ncna1030631 |archive-date=18 April 2020 |access-date=5 April 2020 |work=NBC News |language=en-us}}</ref><ref>{{Cite news |last=Torres |first=Phil |date=18 September 2018 |title=Only Radically Enhancing Humanity Can Save Us All |language=en |work=Slate Magazine |url=https://slate.com/technology/2018/09/genetic-engineering-to-stop-doomsday.html |access-date=5 April 2020 |archive-date=6 August 2020 |archive-url=https://web.archive.org/web/20200806073520/https://slate.com/technology/2018/09/genetic-engineering-to-stop-doomsday.html |url-status=live }}</ref> Another proposed method is closely monitoring or "boxing in" an early-stage AI to prevent it from becoming too powerful. A dominant, aligned superintelligent AI might also mitigate risks from rival AIs, although its creation could present its own existential dangers.<ref>{{Cite journal |last1=Barrett |first1=Anthony M. |last2=Baum |first2=Seth D. |date=23 May 2016 |title=A model of pathways to artificial superintelligence catastrophe for risk and decision analysis |journal=Journal of Experimental & Theoretical Artificial Intelligence |volume=29 |issue=2 |pages=397–414 |arxiv=1607.07730 |doi=10.1080/0952813X.2016.1186228 |s2cid=928824}}</ref>


Institutions such as the [[Alignment Research Center]],<ref>{{Cite web |last=Piper |first=Kelsey |date=2023-03-29 |title=How to test what an AI model can – and shouldn't – do |url=https://www.vox.com/future-perfect/2023/3/29/23661633/gpt-4-openai-alignment-research-center-open-philanthropy-ai-safety |access-date=2023-07-28 |website=Vox |language=en}}</ref> the [[Machine Intelligence Research Institute]], the [[Future of Humanity Institute]],<ref>{{Cite magazine |last=Piesing |first=Mark |date=17 May 2012 |title=AI uprising: humans will be outsourced, not obliterated |magazine=Wired |url=https://www.wired.co.uk/news/archive/2012-05/17/the-dangers-of-an-ai-smarter-than-us |access-date=12 December 2015 |archive-date=7 April 2014 |archive-url=https://web.archive.org/web/20140407041151/http://www.wired.co.uk/news/archive/2012-05/17/the-dangers-of-an-ai-smarter-than-us |url-status=live }}</ref><ref>{{Cite news |last=Coughlan |first=Sean |date=24 April 2013 |title=How are humans going to become extinct? |work=BBC News |url=https://www.bbc.com/news/business-22002530 |access-date=29 March 2014 |archive-date=9 March 2014 |archive-url=https://web.archive.org/web/20140309003706/http://www.bbc.com/news/business-22002530 |url-status=live }}</ref> the [[Future of Life Institute]], the [[Centre for the Study of Existential Risk]], and the [[Center for Human-Compatible AI]]<ref>{{Cite news |last=Bridge |first=Mark |date=10 June 2017 |title=Making robots less confident could prevent them taking over |language=en |work=The Times |url=https://www.thetimes.co.uk/article/making-robots-less-confident-could-prevent-them-taking-over-gnsblq7lx |access-date=21 March 2018 |archive-date=21 March 2018 |archive-url=https://web.archive.org/web/20180321133426/https://www.thetimes.co.uk/article/making-robots-less-confident-could-prevent-them-taking-over-gnsblq7lx |url-status=live }}</ref> are involved in research into AI risk and safety.
Institutions such as the [[Alignment Research Center]],<ref>{{Cite web |last=Piper |first=Kelsey |date=2023-03-29 |title=How to test what an AI model can – and shouldn't – do |url=https://www.vox.com/future-perfect/2023/3/29/23661633/gpt-4-openai-alignment-research-center-open-philanthropy-ai-safety |access-date=2023-07-28 |website=Vox |language=en}}</ref> the [[Machine Intelligence Research Institute]],<ref>{{Cite magazine |last=Piesing |first=Mark |date=17 May 2012 |title=AI uprising: humans will be outsourced, not obliterated |magazine=Wired |url=https://www.wired.co.uk/news/archive/2012-05/17/the-dangers-of-an-ai-smarter-than-us |access-date=12 December 2015 |archive-date=7 April 2014 |archive-url=https://web.archive.org/web/20140407041151/http://www.wired.co.uk/news/archive/2012-05/17/the-dangers-of-an-ai-smarter-than-us |url-status=live }}</ref><ref>{{Cite news |last=Coughlan |first=Sean |date=24 April 2013 |title=How are humans going to become extinct? |work=BBC News |url=https://www.bbc.com/news/business-22002530 |access-date=29 March 2014 |archive-date=9 March 2014 |archive-url=https://web.archive.org/web/20140309003706/http://www.bbc.com/news/business-22002530 |url-status=live }}</ref> the [[Future of Life Institute]], the [[Centre for the Study of Existential Risk]], and the [[Center for Human-Compatible AI]]<ref>{{Cite news |last=Bridge |first=Mark |date=10 June 2017 |title=Making robots less confident could prevent them taking over |language=en |work=The Times |url=https://www.thetimes.co.uk/article/making-robots-less-confident-could-prevent-them-taking-over-gnsblq7lx |access-date=21 March 2018 |archive-date=21 March 2018 |archive-url=https://web.archive.org/web/20180321133426/https://www.thetimes.co.uk/article/making-robots-less-confident-could-prevent-them-taking-over-gnsblq7lx |url-status=live }}</ref> are actively engaged in researching AI risk and safety.


=== Views on banning and regulation ===
=== Views on banning and regulation ===


==== Banning ====
==== Banning ====
Some scholars have said that even if AGI poses an existential risk, attempting to ban research into artificial intelligence is still unwise, and probably futile.<ref>{{Cite journal |last=McGinnis |first=John |author-link=John McGinnis |date=Summer 2010 |title=Accelerating AI |url=http://scholarlycommons.law.northwestern.edu/cgi/viewcontent.cgi?article=1193&context=nulr_online |journal=[[Northwestern University Law Review]] |volume=104 |issue=3 |pages=1253–1270 |access-date=16 July 2014 |quote=For all these reasons, verifying a global relinquishment treaty, or even one limited to AI-related weapons development, is a nonstarter... (For different reasons from ours, the Machine Intelligence Research Institute) considers (AGI) relinquishment infeasible... |archive-date=15 February 2016 |archive-url=https://web.archive.org/web/20160215073656/http://scholarlycommons.law.northwestern.edu/cgi/viewcontent.cgi?article=1193&context=nulr_online |url-status=live }}</ref><ref>{{Cite journal |last1=Sotala |first1=Kaj |last2=Yampolskiy |first2=Roman |author-link2=Roman Yampolskiy |date=19 December 2014 |title=Responses to catastrophic AGI risk: a survey |journal=[[Physica Scripta]] |volume=90 |issue=1 |quote=In general, most writers reject proposals for broad relinquishment... Relinquishment proposals suffer from many of the same problems as regulation proposals, but to a greater extent. There is no historical precedent of general, multi-use technology similar to AGI being successfully relinquished for good, nor do there seem to be any theoretical reasons for believing that relinquishment proposals would work in the future. Therefore we do not consider them to be a viable class of proposals.}}</ref><ref>{{Cite news |last=Allenby |first=Brad |date=11 April 2016 |title=The Wrong Cognitive Measuring Stick |language=en-US |work=Slate |url=http://www.slate.com/articles/technology/future_tense/2016/04/why_it_s_a_mistake_to_compare_a_i_with_human_intelligence.html |access-date=15 May 2016 |quote=It is fantasy to suggest that the accelerating development and deployment of technologies that taken together are considered to be A.I. will be stopped or limited, either by regulation or even by national legislation. |archive-date=15 May 2016 |archive-url=https://web.archive.org/web/20160515114003/http://www.slate.com/articles/technology/future_tense/2016/04/why_it_s_a_mistake_to_compare_a_i_with_human_intelligence.html |url-status=live }}</ref> Skeptics argue that regulation of AI is completely valueless, as no existential risk exists. But scholars who believe in existential risk say it is difficult to depend on people from the AI industry to regulate or constrain AI research because it directly contradicts their personal interests.<ref name=":7">{{Cite book |last=Yampolskiy |first=Roman V. |chapter=AI Risk Skepticism |date=2022 |editor-last=Müller |editor-first=Vincent C. |title=Philosophy and Theory of Artificial Intelligence 2021 |chapter-url=https://link.springer.com/chapter/10.1007/978-3-031-09153-7_18 |series=Studies in Applied Philosophy, Epistemology and Rational Ethics |volume=63 |language=en |location=Cham |publisher=Springer International Publishing |pages=225–248 |doi=10.1007/978-3-031-09153-7_18 |isbn=978-3-031-09153-7}}</ref> The scholars also agree with the skeptics that banning research would be unwise, as research could be moved to countries with looser regulations or conducted covertly.<ref name=":7" /> The latter issue is particularly relevant, as artificial intelligence research can be done on a small scale without substantial infrastructure or resources.<ref name="mcginnis">{{Cite journal |last=McGinnis |first=John |author-link=John McGinnis |date=Summer 2010 |title=Accelerating AI |url=http://scholarlycommons.law.northwestern.edu/cgi/viewcontent.cgi?article=1193&context=nulr_online |journal=[[Northwestern University Law Review]] |volume=104 |issue=3 |pages=1253–1270 |access-date=16 July 2014 |archive-date=15 February 2016 |archive-url=https://web.archive.org/web/20160215073656/http://scholarlycommons.law.northwestern.edu/cgi/viewcontent.cgi?article=1193&context=nulr_online |url-status=live }}</ref><ref>{{Cite magazine |date=4 October 2013 |title=Why We Should Think About the Threat of Artificial Intelligence |magazine=[[The New Yorker]] |url=https://www.newyorker.com/tech/elements/why-we-should-think-about-the-threat-of-artificial-intelligence |access-date=7 February 2016 |quote=Of course, one could try to ban super-intelligent computers altogether. But 'the competitive advantage—economic, military, even artistic—of every advance in automation is so compelling,' [[Vernor Vinge]], the mathematician and science-fiction author, wrote, 'that passing laws, or having customs, that forbid such things merely assures that someone else will.' |archive-date=4 February 2016 |archive-url=https://web.archive.org/web/20160204081825/http://www.newyorker.com/tech/elements/why-we-should-think-about-the-threat-of-artificial-intelligence |url-status=live }}</ref> Two additional hypothetical difficulties with bans (or other regulation) are that technology entrepreneurs statistically tend toward general skepticism about government regulation, and that businesses could have a strong incentive to (and might well succeed at) fighting regulation and [[politicization of science|politicizing]] the underlying debate.<ref>{{Cite journal |last=Baum |first=Seth |date=2018-08-22 |title=Superintelligence Skepticism as a Political Tool |journal=Information |volume=9 |issue=9 |page=209 |doi=10.3390/info9090209 |issn=2078-2489 |doi-access=free}}</ref>
Some scholars have said that even if AGI poses an existential risk, attempting to ban research into artificial intelligence is still unwise, and probably futile.<ref>{{Cite journal |last=McGinnis |first=John |author-link=John McGinnis |date=Summer 2010 |title=Accelerating AI |url=http://scholarlycommons.law.northwestern.edu/cgi/viewcontent.cgi?article=1193&context=nulr_online |journal=[[Northwestern University Law Review]] |volume=104 |issue=3 |pages=1253–1270 |access-date=16 July 2014 |quote=For all these reasons, verifying a global relinquishment treaty, or even one limited to AI-related weapons development, is a nonstarter... (For different reasons from ours, the Machine Intelligence Research Institute) considers (AGI) relinquishment infeasible... |archive-date=15 February 2016 |archive-url=https://web.archive.org/web/20160215073656/http://scholarlycommons.law.northwestern.edu/cgi/viewcontent.cgi?article=1193&context=nulr_online |url-status=live }}</ref><ref>{{Cite journal |last1=Sotala |first1=Kaj |last2=Yampolskiy |first2=Roman |author-link2=Roman Yampolskiy |date=19 December 2014 |title=Responses to catastrophic AGI risk: a survey |journal=[[Physica Scripta]] |volume=90 |issue=1 |quote=In general, most writers reject proposals for broad relinquishment... Relinquishment proposals suffer from many of the same problems as regulation proposals, but to a greater extent. There is no historical precedent of general, multi-use technology similar to AGI being successfully relinquished for good, nor do there seem to be any theoretical reasons for believing that relinquishment proposals would work in the future. Therefore we do not consider them to be a viable class of proposals.}}</ref><ref>{{Cite news |last=Allenby |first=Brad |date=11 April 2016 |title=The Wrong Cognitive Measuring Stick |language=en-US |work=Slate |url=http://www.slate.com/articles/technology/future_tense/2016/04/why_it_s_a_mistake_to_compare_a_i_with_human_intelligence.html |access-date=15 May 2016 |quote=It is fantasy to suggest that the accelerating development and deployment of technologies that taken together are considered to be A.I. will be stopped or limited, either by regulation or even by national legislation. |archive-date=15 May 2016 |archive-url=https://web.archive.org/web/20160515114003/http://www.slate.com/articles/technology/future_tense/2016/04/why_it_s_a_mistake_to_compare_a_i_with_human_intelligence.html |url-status=live }}</ref> Skeptics consider AI regulation pointless, as no existential risk exists. But scholars who believe in the risk argue that relying on AI industry insiders to regulate or constrain AI research is impractical due to conflicts of interest.<ref name=":7">{{Cite book |last=Yampolskiy |first=Roman V. |chapter=AI Risk Skepticism |date=2022 |editor-last=Müller |editor-first=Vincent C. |title=Philosophy and Theory of Artificial Intelligence 2021 |chapter-url=https://link.springer.com/chapter/10.1007/978-3-031-09153-7_18 |series=Studies in Applied Philosophy, Epistemology and Rational Ethics |volume=63 |language=en |location=Cham |publisher=Springer International Publishing |pages=225–248 |doi=10.1007/978-3-031-09153-7_18 |isbn=978-3-031-09153-7}}</ref> They also agree with skeptics that banning research would be unwise, as research could be moved to countries with looser regulations or conducted covertly.<ref name=":7" /> Additional challenges to bans or regulation include technology entrepreneurs' general skepticism of government regulation and potential incentives for businesses to resist regulation and [[Politicization of science|politicize]] the debate.<ref>{{Cite journal |last=Baum |first=Seth |date=2018-08-22 |title=Superintelligence Skepticism as a Political Tool |journal=Information |volume=9 |issue=9 |page=209 |doi=10.3390/info9090209 |issn=2078-2489 |doi-access=free}}</ref>


==== Regulation ====
==== Regulation ====
Line 252: Line 251:
Musk called for some sort of regulation of AI development as early as 2017. According to [[National Public Radio|NPR]], he is "clearly not thrilled" to be advocating government scrutiny that could impact his own industry, but believes the risks of going completely without oversight are too high: "Normally the way regulations are set up is when a bunch of bad things happen, there's a public outcry, and after many years a regulatory agency is set up to regulate that industry. It takes forever. That, in the past, has been bad but not something which represented a fundamental risk to the existence of civilisation." Musk states the first step would be for the government to gain "insight" into the actual status of current research, warning that "Once there is awareness, people will be extremely afraid... [as] they should be." In response, politicians expressed skepticism about the wisdom of regulating a technology that is still in development.<ref>{{Cite news |last=Domonoske |first=Camila |date=July 17, 2017 |title=Elon Musk Warns Governors: Artificial Intelligence Poses 'Existential Risk' |language=en |work=NPR |url=https://www.npr.org/sections/thetwo-way/2017/07/17/537686649/elon-musk-warns-governors-artificial-intelligence-poses-existential-risk |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20200423135755/https://www.npr.org/sections/thetwo-way/2017/07/17/537686649/elon-musk-warns-governors-artificial-intelligence-poses-existential-risk |archive-date=23 April 2020}}</ref><ref>{{Cite news |last=Gibbs |first=Samuel |date=17 July 2017 |title=Elon Musk: regulate AI to combat 'existential threat' before it's too late |work=The Guardian |url=https://www.theguardian.com/technology/2017/jul/17/elon-musk-regulation-ai-combat-existential-threat-tesla-spacex-ceo |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20200606072024/https://www.theguardian.com/technology/2017/jul/17/elon-musk-regulation-ai-combat-existential-threat-tesla-spacex-ceo |archive-date=6 June 2020}}</ref><ref name="cnbc2">{{Cite news |last=Kharpal |first=Arjun |date=7 November 2017 |title=A.I. is in its 'infancy' and it's too early to regulate it, Intel CEO Brian Krzanich says |work=CNBC |url=https://www.cnbc.com/2017/11/07/ai-infancy-and-too-early-to-regulate-intel-ceo-brian-krzanich-says.html |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20200322115325/https://www.cnbc.com/2017/11/07/ai-infancy-and-too-early-to-regulate-intel-ceo-brian-krzanich-says.html |archive-date=22 March 2020}}</ref>
Musk called for some sort of regulation of AI development as early as 2017. According to [[National Public Radio|NPR]], he is "clearly not thrilled" to be advocating government scrutiny that could impact his own industry, but believes the risks of going completely without oversight are too high: "Normally the way regulations are set up is when a bunch of bad things happen, there's a public outcry, and after many years a regulatory agency is set up to regulate that industry. It takes forever. That, in the past, has been bad but not something which represented a fundamental risk to the existence of civilisation." Musk states the first step would be for the government to gain "insight" into the actual status of current research, warning that "Once there is awareness, people will be extremely afraid... [as] they should be." In response, politicians expressed skepticism about the wisdom of regulating a technology that is still in development.<ref>{{Cite news |last=Domonoske |first=Camila |date=July 17, 2017 |title=Elon Musk Warns Governors: Artificial Intelligence Poses 'Existential Risk' |language=en |work=NPR |url=https://www.npr.org/sections/thetwo-way/2017/07/17/537686649/elon-musk-warns-governors-artificial-intelligence-poses-existential-risk |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20200423135755/https://www.npr.org/sections/thetwo-way/2017/07/17/537686649/elon-musk-warns-governors-artificial-intelligence-poses-existential-risk |archive-date=23 April 2020}}</ref><ref>{{Cite news |last=Gibbs |first=Samuel |date=17 July 2017 |title=Elon Musk: regulate AI to combat 'existential threat' before it's too late |work=The Guardian |url=https://www.theguardian.com/technology/2017/jul/17/elon-musk-regulation-ai-combat-existential-threat-tesla-spacex-ceo |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20200606072024/https://www.theguardian.com/technology/2017/jul/17/elon-musk-regulation-ai-combat-existential-threat-tesla-spacex-ceo |archive-date=6 June 2020}}</ref><ref name="cnbc2">{{Cite news |last=Kharpal |first=Arjun |date=7 November 2017 |title=A.I. is in its 'infancy' and it's too early to regulate it, Intel CEO Brian Krzanich says |work=CNBC |url=https://www.cnbc.com/2017/11/07/ai-infancy-and-too-early-to-regulate-intel-ceo-brian-krzanich-says.html |url-status=live |access-date=27 November 2017 |archive-url=https://web.archive.org/web/20200322115325/https://www.cnbc.com/2017/11/07/ai-infancy-and-too-early-to-regulate-intel-ceo-brian-krzanich-says.html |archive-date=22 March 2020}}</ref>


In 2021 the [[United Nations]] (UN) considered banning autonomous lethal weapons, but consensus could not be reached.<ref>{{Cite web |last=Dawes |first=James |date=2021-12-20 |title=UN fails to agree on 'killer robot' ban as nations pour billions into autonomous weapons research |url=http://theconversation.com/un-fails-to-agree-on-killer-robot-ban-as-nations-pour-billions-into-autonomous-weapons-research-173616 |access-date=2023-07-28 |website=The Conversation |language=en}}</ref> In July 2023 the UN [[United Nations Security Council|Security Council]] for the first time held a session to consider the risks and threats posed by AI to world peace and stability, along with potential benefits.<ref name=":13">{{Cite news |last=Fassihi |first=Farnaz |date=2023-07-18 |title=U.N. Officials Urge Regulation of Artificial Intelligence |language=en-US |work=The New York Times |url=https://www.nytimes.com/2023/07/18/world/un-security-council-ai.html |access-date=2023-07-20 |issn=0362-4331}}</ref><ref>{{Cite web |title=International Community Must Urgently Confront New Reality of Generative, Artificial Intelligence, Speakers Stress as Security Council Debates Risks, Rewards |website=United Nations |url=https://press.un.org/en/2023/sc15359.doc.htm |access-date=2023-07-20}}</ref> [[Secretary-General of the United Nations|Secretary-General]] [[António Guterres]] advocated the creation of a global watchdog to oversee the emerging technology, saying, "Generative AI has enormous potential for good and evil at scale. Its creators themselves have warned that much bigger, potentially catastrophic and existential risks lie ahead."<ref name=":12">{{Cite news |last=Fung |first=Brian |date=2023-07-18 |title=UN Secretary General embraces calls for a new UN agency on AI in the face of 'potentially catastrophic and existential risks' |work=CNN Business |url=https://www.cnn.com/2023/07/18/tech/un-ai-agency/index.html |access-date=2023-07-20 |language=en}}</ref> At the council session, Russia said it believes AI risks are too poorly understood to be considered a threat to global stability. China argued against strict global regulation, saying countries should be able to develop their own rules, while also saying they opposed the use of AI to "create military hegemony or undermine the sovereignty of a country."<ref name=":13" />
In 2021 the [[United Nations]] (UN) considered banning autonomous lethal weapons, but consensus could not be reached.<ref>{{Cite web |last=Dawes |first=James |date=2021-12-20 |title=UN fails to agree on 'killer robot' ban as nations pour billions into autonomous weapons research |url=http://theconversation.com/un-fails-to-agree-on-killer-robot-ban-as-nations-pour-billions-into-autonomous-weapons-research-173616 |access-date=2023-07-28 |website=The Conversation |language=en}}</ref> In July 2023 the UN [[United Nations Security Council|Security Council]] for the first time held a session to consider the risks and threats posed by AI to world peace and stability, along with potential benefits.<ref name=":13">{{Cite news |last=Fassihi |first=Farnaz |date=2023-07-18 |title=U.N. Officials Urge Regulation of Artificial Intelligence |language=en-US |work=The New York Times |url=https://www.nytimes.com/2023/07/18/world/un-security-council-ai.html |access-date=2023-07-20 |issn=0362-4331}}</ref><ref>{{Cite web |title=International Community Must Urgently Confront New Reality of Generative, Artificial Intelligence, Speakers Stress as Security Council Debates Risks, Rewards |website=United Nations |url=https://press.un.org/en/2023/sc15359.doc.htm |access-date=2023-07-20}}</ref> [[Secretary-General of the United Nations|Secretary-General]] [[António Guterres]] advocated the creation of a global watchdog to oversee the emerging technology, saying, "Generative AI has enormous potential for good and evil at scale. Its creators themselves have warned that much bigger, potentially catastrophic and existential risks lie ahead."<ref name=":12">{{Cite news |last=Fung |first=Brian |date=2023-07-18 |title=UN Secretary General embraces calls for a new UN agency on AI in the face of 'potentially catastrophic and existential risks' |work=CNN Business |url=https://www.cnn.com/2023/07/18/tech/un-ai-agency/index.html |access-date=2023-07-20 |language=en}}</ref> At the council session, Russia said it believes AI risks are too poorly understood to be considered a threat to global stability. China argued against strict global regulation, saying countries should be able to develop their own rules, while also saying they opposed the use of AI to "create military hegemony or undermine the sovereignty of a country".<ref name=":13" />


Regulation of conscious AGIs focuses on integrating them with existing human society and can be divided into considerations of their legal standing and of their moral rights.<ref name=":532">{{Cite journal |last1=Sotala |first1=Kaj |last2=Yampolskiy |first2=Roman V. |date=2014-12-19 |title=Responses to catastrophic AGI risk: a survey |journal=Physica Scripta |volume=90 |issue=1 |page=018001 |bibcode=2015PhyS...90a8001S |doi=10.1088/0031-8949/90/1/018001 |issn=0031-8949 |doi-access=free}}</ref> AI arms control will likely require the institutionalization of new international norms embodied in effective technical specifications combined with active monitoring and informal diplomacy by communities of experts, together with a legal and political verification process.<ref>{{Cite journal |last=Geist |first=Edward Moore |date=2016-08-15 |title=It's already too late to stop the AI arms race—We must manage it instead |journal=Bulletin of the Atomic Scientists |volume=72 |issue=5 |pages=318–321 |bibcode=2016BuAtS..72e.318G |doi=10.1080/00963402.2016.1216672 |issn=0096-3402 |s2cid=151967826}}</ref><ref name=":132"/>
Regulation of conscious AGIs focuses on integrating them with existing human society and can be divided into considerations of their legal standing and of their moral rights.<ref name=":532">{{Cite journal |last1=Sotala |first1=Kaj |last2=Yampolskiy |first2=Roman V. |date=2014-12-19 |title=Responses to catastrophic AGI risk: a survey |journal=Physica Scripta |volume=90 |issue=1 |page=018001 |bibcode=2015PhyS...90a8001S |doi=10.1088/0031-8949/90/1/018001 |issn=0031-8949 |doi-access=free}}</ref> AI arms control will likely require the institutionalization of new international norms embodied in effective technical specifications combined with active monitoring and informal diplomacy by communities of experts, together with a legal and political verification process.<ref>{{Cite journal |last=Geist |first=Edward Moore |date=2016-08-15 |title=It's already too late to stop the AI arms race—We must manage it instead |journal=Bulletin of the Atomic Scientists |volume=72 |issue=5 |pages=318–321 |bibcode=2016BuAtS..72e.318G |doi=10.1080/00963402.2016.1216672 |issn=0096-3402 |s2cid=151967826}}</ref><ref name=":132"/>
Line 258: Line 257:
In July 2023, the US government secured voluntary safety commitments from major tech companies, including [[OpenAI]], [[Amazon (company)|Amazon]], [[Google]], [[Meta Platforms|Meta]], and [[Microsoft]]. The companies agreed to implement safeguards, including third-party oversight and security testing by independent experts, to address concerns related to AI's potential risks and societal harms. The parties framed the commitments as an intermediate step while regulations are formed. Amba Kak, executive director of the [[AI Now Institute]], said, "A closed-door deliberation with corporate actors resulting in voluntary safeguards isn't enough" and called for public deliberation and regulations of the kind to which companies would not voluntarily agree.<ref>{{Cite web |date=2023-07-21 |title=Amazon, Google, Meta, Microsoft and other tech firms agree to AI safeguards set by the White House |url=https://apnews.com/article/artificial-intelligence-safeguards-joe-biden-kamala-harris-4caf02b94275429f764b06840897436c |access-date=2023-07-21 |website=AP News |language=en}}</ref><ref>{{Cite web |date=2023-07-21 |title=Amazon, Google, Meta, Microsoft and other firms agree to AI safeguards |url=https://www.redditchadvertiser.co.uk/news/national/23670894.amazon-google-meta-microsoft-firms-agree-ai-safeguards/ |access-date=2023-07-21 |website=Redditch Advertiser |language=en}}</ref>
In July 2023, the US government secured voluntary safety commitments from major tech companies, including [[OpenAI]], [[Amazon (company)|Amazon]], [[Google]], [[Meta Platforms|Meta]], and [[Microsoft]]. The companies agreed to implement safeguards, including third-party oversight and security testing by independent experts, to address concerns related to AI's potential risks and societal harms. The parties framed the commitments as an intermediate step while regulations are formed. Amba Kak, executive director of the [[AI Now Institute]], said, "A closed-door deliberation with corporate actors resulting in voluntary safeguards isn't enough" and called for public deliberation and regulations of the kind to which companies would not voluntarily agree.<ref>{{Cite web |date=2023-07-21 |title=Amazon, Google, Meta, Microsoft and other tech firms agree to AI safeguards set by the White House |url=https://apnews.com/article/artificial-intelligence-safeguards-joe-biden-kamala-harris-4caf02b94275429f764b06840897436c |access-date=2023-07-21 |website=AP News |language=en}}</ref><ref>{{Cite web |date=2023-07-21 |title=Amazon, Google, Meta, Microsoft and other firms agree to AI safeguards |url=https://www.redditchadvertiser.co.uk/news/national/23670894.amazon-google-meta-microsoft-firms-agree-ai-safeguards/ |access-date=2023-07-21 |website=Redditch Advertiser |language=en}}</ref>


In October 2023, U.S. President [[Joe Biden]] issued an [[executive order]] on the "Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence".<ref>{{Cite web |last=The White House |date=2023-10-30 |title=Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence |url=https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/ |access-date=2023-12-19 |website=The White House |language=en-US}}</ref> Alongside other requirements, the order mandates the development of guidelines for AI models that permit the "evasion of human control."
In October 2023, U.S. President [[Joe Biden]] issued an executive order on the "[[Executive Order 14110|Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence]]".<ref>{{Cite web |last=The White House |date=2023-10-30 |title=Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence |url=https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/ |access-date=2023-12-19 |website=The White House |language=en-US}}</ref> Alongside other requirements, the order mandates the development of guidelines for AI models that permit the "evasion of human control".


== See also ==
== See also ==
Line 297: Line 296:
[[Category:Future problems]]
[[Category:Future problems]]
[[Category:Human extinction]]
[[Category:Human extinction]]
[[Category:AI safety]]
[[Category:Technology hazards]]
[[Category:Technology hazards]]
[[Category:Doomsday scenarios]]
[[Category:Doomsday scenarios]]

Revision as of 19:29, 7 July 2024

Existential risk from artificial general intelligence refers to the idea that substantial progress in artificial general intelligence (AGI) could lead to human extinction or an irreversible global catastrophe.[1][2][3]

One argument for the importance of this risk references how human beings dominate other species because the human brain possesses distinctive capabilities other animals lack. If AI were to surpass human intelligence and become superintelligent, it might become uncontrollable. Just as the fate of the mountain gorilla depends on human goodwill, the fate of humanity could depend on the actions of a future machine superintelligence.[4]

The plausibility of existential catastrophe due to AI is widely debated. It hinges in part on whether AGI or superintelligence are achievable, the speed at which dangerous capabilities and behaviors emerge,[5] and whether practical scenarios for AI takeovers exist.[6] Concerns about superintelligence have been voiced by leading computer scientists and tech CEOs such as Geoffrey Hinton,[7] Yoshua Bengio,[8] Alan Turing,[a] Elon Musk,[11] and OpenAI CEO Sam Altman.[12] In 2022, a survey of AI researchers with a 17% response rate found that the majority believed there is a 10 percent or greater chance that human inability to control AI will cause an existential catastrophe.[13][14] In 2023, hundreds of AI experts and other notable figures signed a statement declaring, "Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war".[15] Following increased concern over AI risks, government leaders such as United Kingdom prime minister Rishi Sunak[16] and United Nations Secretary-General António Guterres[17] called for an increased focus on global AI regulation.

Two sources of concern stem from the problems of AI control and alignment. Controlling a superintelligent machine or instilling it with human-compatible values may be difficult. Many researchers believe that a superintelligent machine would likely resist attempts to disable it or change its goals as that would prevent it from accomplishing its present goals. It would be extremely challenging to align a superintelligence with the full breadth of significant human values and constraints.[1][18][19] In contrast, skeptics such as computer scientist Yann LeCun argue that superintelligent machines will have no desire for self-preservation.[20]

A third source of concern is the possibility of a sudden "intelligence explosion" that catches humanity unprepared. In this scenario, an AI more intelligent than its creators would be able to recursively improve itself at an exponentially increasing rate, improving too quickly for its handlers or society at large to control.[1][18] Empirically, examples like AlphaZero, which taught itself to play Go and quickly surpassed human ability, show that domain-specific AI systems can sometimes progress from subhuman to superhuman ability very quickly, although such machine learning systems do not recursively improve their fundamental architecture.[21]

History

One of the earliest authors to express serious concern that highly advanced machines might pose existential risks to humanity was the novelist Samuel Butler, who wrote in his 1863 essay Darwin among the Machines:[22]

The upshot is simply a question of time, but that the time will come when the machines will hold the real supremacy over the world and its inhabitants is what no person of a truly philosophic mind can for a moment question.

In 1951, foundational computer scientist Alan Turing wrote the article "Intelligent Machinery, A Heretical Theory", in which he proposed that artificial general intelligences would likely "take control" of the world as they became more intelligent than human beings:

Let us now assume, for the sake of argument, that [intelligent] machines are a genuine possibility, and look at the consequences of constructing them... There would be no question of the machines dying, and they would be able to converse with each other to sharpen their wits. At some stage therefore we should have to expect the machines to take control, in the way that is mentioned in Samuel Butler's Erewhon.[23]

In 1965, I. J. Good originated the concept now known as an "intelligence explosion" and said the risks were underappreciated:[24]

Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines; there would then unquestionably be an 'intelligence explosion', and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is the last invention that man need ever make, provided that the machine is docile enough to tell us how to keep it under control. It is curious that this point is made so seldom outside of science fiction. It is sometimes worthwhile to take science fiction seriously.[25]

Scholars such as Marvin Minsky[26] and I. J. Good himself[27] occasionally expressed concern that a superintelligence could seize control, but issued no call to action. In 2000, computer scientist and Sun co-founder Bill Joy penned an influential essay, "Why The Future Doesn't Need Us", identifying superintelligent robots as a high-tech danger to human survival, alongside nanotechnology and engineered bioplagues.[28]

Nick Bostrom published Superintelligence in 2014, which presented his arguments that superintelligence poses an existential threat.[29] By 2015, public figures such as physicists Stephen Hawking and Nobel laureate Frank Wilczek, computer scientists Stuart J. Russell and Roman Yampolskiy, and entrepreneurs Elon Musk and Bill Gates were expressing concern about the risks of superintelligence.[30][31][32][33] Also in 2015, the Open Letter on Artificial Intelligence highlighted the "great potential of AI" and encouraged more research on how to make it robust and beneficial.[34] In April 2016, the journal Nature warned: "Machines and robots that outperform humans across the board could self-improve beyond our control—and their interests might not align with ours".[35] In 2020, Brian Christian published The Alignment Problem, which details the history of progress on AI alignment up to that time.[36][37]

In March 2023, key figures in AI, such as Musk, signed a letter from the Future of Life Institute calling a halt to advanced AI training until it could be properly regulated.[38] In May 2023, the Center for AI Safety released a statement signed by numerous experts in AI safety and the AI existential risk which stated: "Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war."[39][40]

AI capabilities

General Intelligence

Artificial general intelligence (AGI) is typically defined as a system that performs at least as well as humans in most or all intellectual tasks.[41] A 2022 survey of AI researchers found that 90% of respondents expected AGI would be achieved in the next 100 years, and half expected the same by 2061.[42] Meanwhile, some researchers dismiss existential risks from AGI as "science fiction" based on their high confidence that AGI will not be created anytime soon.[43]

Breakthroughs in large language models have led some researchers to reassess their expectations. Notably, Geoffrey Hinton said in 2023 that he recently changed his estimate from "20 to 50 years before we have general purpose A.I." to "20 years or less".[44]

Superintelligence

In contrast with AGI, Bostrom defines a superintelligence as "any intellect that greatly exceeds the cognitive performance of humans in virtually all domains of interest", including scientific creativity, strategic planning, and social skills.[45][4] He argues that a superintelligence can outmaneuver humans anytime its goals conflict with humans'. It may choose to hide its true intent until humanity cannot stop it.[46][4] Bostrom writes that in order to be safe for humanity, a superintelligence must be aligned with human values and morality, so that it is "fundamentally on our side".[47]

Stephen Hawking argued that superintelligence is physically possible because "there is no physical law precluding particles from being organised in ways that perform even more advanced computations than the arrangements of particles in human brains".[31]

When artificial superintelligence (ASI) may be achieved, if ever, is necessarily less certain than predictions for AGI. In 2023, OpenAI leaders said that not only AGI, but superintelligence may be achieved in less than 10 years.[48]

Comparison with humans

Bostrom argues that AI has many advantages over the human brain:[4]

  • Speed of computation: biological neurons operate at a maximum frequency of around 200 Hz, compared to potentially multiple GHz for computers.
  • Internal communication speed: axons transmit signals at up to 120 m/s, while computers transmit signals at the speed of electricity, or optically at the speed of light.
  • Scalability: human intelligence is limited by the size and structure of the brain, and by the efficiency of social communication, while AI may be able to scale by simply adding more hardware.
  • Memory: notably working memory, because in humans it is limited to a few chunks of information at a time.
  • Reliability: transistors are more reliable than biological neurons, enabling higher precision and requiring less redundancy.
  • Duplicability: unlike human brains, AI software and models can be easily copied.
  • Editability: the parameters and internal workings of an AI model can easily be modified, unlike the connections in a human brain.
  • Memory sharing and learning: AIs may be able to learn from the experiences of other AIs in a manner more efficient than human learning.

Intelligence explosion

According to Bostrom, an AI that has an expert-level facility at certain key software engineering tasks could become a superintelligence due to its capability to recursively improve its own algorithms, even if it is initially limited in other domains not directly relevant to engineering.[4][46] This suggests that an intelligence explosion may someday catch humanity unprepared.[4]

The economist Robin Hanson has said that, to launch an intelligence explosion, an AI must become vastly better at software innovation than the rest of the world combined, which he finds implausible.[49]

In a "fast takeoff" scenario, the transition from AGI to superintelligence could take days or months. In a "slow takeoff", it could take years or decades, leaving more time for society to prepare.[50]

Alien mind

Superintelligences are sometimes called "alien minds", referring to the idea that their way of thinking and motivations could be vastly different from ours. This is generally considered as a source of risk, making it more difficult to anticipate what a superintelligence might do. It also suggests the possibility that a superintelligence may not particularly value humans by default.[51] To avoid anthropomorphism, superintelligence is sometimes viewed as a powerful optimizer that makes the best decisions to achieve its goals.[4]

The field of "mechanistic interpretability" aims to better understand the inner workings of AI models, potentially allowing us one day to detect signs of deception and misalignment.[52]

Limits

It has been argued that there are limitations to what intelligence can achieve. Notably, the chaotic nature or time complexity of some systems could fundamentally limit a superintelligence's ability to predict some aspects of the future, increasing its uncertainty.[53]

Dangerous capabilities

Advanced AI could generate enhanced pathogens, cyberattacks or manipulate people. These capabilities could be misused by humans,[54] or exploited by the AI itself if misaligned.[4] A full-blown superintelligence could find various ways to gain a decisive influence if it wanted to,[4] but these dangerous capabilities may become available earlier, in weaker and more specialized AI systems. They may cause societal instability and empower malicious actors.[54]

Social manipulation

Geoffrey Hinton warned that in the short term, the profusion of AI-generated text, images and videos will make it more difficult to figure out the truth, which he says authoritarian states could exploit to manipulate elections.[55] Such large-scale, personalized manipulation capabilities can increase the existential risk of a worldwide "irreversible totalitarian regime". It could also be used by malicious actors to fracture society and make it dysfunctional.[54]

Cyberattacks

AI-enabled cyberattacks are increasingly considered a present and critical threat. According to NATO's technical director of cyberspace, "The number of attacks is increasing exponentially".[56] AI can also be used defensively, to preemptively find and fix vulnerabilities, and detect threats.[57]

AI could improve the "accessibility, success rate, scale, speed, stealth and potency of cyberattacks", potentially causing "significant geopolitical turbulence" if it facilitates attacks more than defense.[54]

Speculatively, such hacking capabilities could be used by an AI system to break out of its local environment, generate revenue, or acquire cloud computing resources.[58]

Enhanced pathogens

As AI technology democratizes, it may become easier to engineer more contagious and lethal pathogens. This could enable people with limited skills in synthetic biology to engage in bioterrorism. Dual-use technology that is useful for medicine could be repurposed to create weapons.[54]

For example, in 2022, scientists modified an AI system originally intended for generating non-toxic, therapeutic molecules with the purpose of creating new drugs. The researchers adjusted the system so that toxicity is rewarded rather than penalized. This simple change enabled the AI system to create, in six hours, 40,000 candidate molecules for chemical warfare, including known and novel molecules.[54][59]

AI arms race

Companies, state actors, and other organizations competing to develop AI technologies could lead to a race to the bottom of safety standards.[60] As rigorous safety procedures take time and resources, projects that proceed more carefully risk being out-competed by less scrupulous developers.[61][54]

AI could be used to gain military advantages via autonomous lethal weapons, cyberwarfare, or automated decision-making.[54] As an example of autonomous lethal weapons, miniaturized drones could facilitate low-cost assassination of military or civilian targets, a scenario highlighted in the 2017 short film Slaughterbots.[62] AI could be used to gain an edge in decision-making by quickly analyzing large amounts of data and making decisions more quickly and effectively than humans. This could increase the speed and unpredictability of war, especially when accounting for automated retaliation systems.[54][63]

Types of existential risk

Scope–severity grid from Bostrom's paper "Existential Risk Prevention as Global Priority"[64]

An existential risk is "one that threatens the premature extinction of Earth-originating intelligent life or the permanent and drastic destruction of its potential for desirable future development".[65]

Besides extinction risk, there is the risk that the civilization gets permanently locked into a flawed future. One example is a "value lock-in": If humanity still has moral blind spots similar to slavery in the past, AI might irreversibly entrench it, preventing moral progress. AI could also be used to spread and preserve the set of values of whoever develops it.[66] AI could facilitate large-scale surveillance and indoctrination, which could be used to create a stable repressive worldwide totalitarian regime.[67]

It is difficult or impossible to reliably evaluate whether an advanced AI is sentient and to what degree. But if sentient machines are mass created in the future, engaging in a civilizational path that indefinitely neglects their welfare could be an existential catastrophe.[68][69] Moreover, it may be possible to engineer digital minds that can feel much more happiness than humans with fewer resources, called "super-beneficiaries". Such an opportunity raises the question of how to share the world and which "ethical and political framework" would enable a mutually beneficial coexistence between biological and digital minds.[70]

AI may also drastically improve humanity's future. Toby Ord considers the existential risk a reason for "proceeding with due caution", not for abandoning AI.[67] Max More calls AI an "existential opportunity", highlighting the cost of not developing it.[71]

According to Bostrom, superintelligence could help reduce the existential risk from other powerful technologies such as molecular nanotechnology or synthetic biology. It is thus conceivable that developing superintelligence before other dangerous technologies would reduce the overall existential risk.[4]

AI alignment

The alignment problem is the research problem of how to reliably assign objectives, preferences or ethical principles to AIs.

Instrumental convergence

An "instrumental" goal is a sub-goal that helps to achieve an agent's ultimate goal. "Instrumental convergence" refers to the fact that some sub-goals are useful for achieving virtually any ultimate goal, such as acquiring resources or self-preservation.[72] Bostrom argues that if an advanced AI's instrumental goals conflict with humanity's goals, the AI might harm humanity in order to acquire more resources or prevent itself from being shut down, but only as a way to achieve its ultimate goal.[4]

Some ways in which an advanced misaligned AI could try to gain more power.[73] Power-seeking behaviors may arise because power is useful to accomplish virtually any objective.[74]

Russell argues that a sufficiently advanced machine "will have self-preservation even if you don't program it in... if you say, 'Fetch the coffee', it can't fetch the coffee if it's dead. So if you give it any goal whatsoever, it has a reason to preserve its own existence to achieve that goal."[20][75]

Resistance to changing goals

Even if current goal-based AI programs are not intelligent enough to think of resisting programmer attempts to modify their goal structures, a sufficiently advanced AI might resist any attempts to change its goal structure, just as a pacifist would not want to take a pill that makes them want to kill people. If the AI were superintelligent, it would likely succeed in out-maneuvering its human operators and prevent itself being "turned off" or reprogrammed with a new goal.[4][76] This is particularly relevant to value lock-in scenarios. The field of "corrigibility" studies how to make agents that will not resist attempts to change their goals.[77]

Difficulty of specifying goals

In the "intelligent agent" model, an AI can loosely be viewed as a machine that chooses whatever action appears to best achieve its set of goals, or "utility function". A utility function gives each possible situation a score that indicates its desirability to the agent. Researchers know how to write utility functions that mean "minimize the average network latency in this specific telecommunications model" or "maximize the number of reward clicks", but do not know how to write a utility function for "maximize human flourishing"; nor is it clear whether such a function meaningfully and unambiguously exists. Furthermore, a utility function that expresses some values but not others will tend to trample over the values the function does not reflect.[78][79]

An additional source of concern is that AI "must reason about what people intend rather than carrying out commands literally", and that it must be able to fluidly solicit human guidance if it is too uncertain about what humans want.[80]

Alignment of superintelligences

Some researchers believe the alignment problem may be particularly difficult when applied to superintelligences. Their reasoning includes:

  • As AI systems increase in capabilities, the potential dangers associated with experimentation grow. This makes iterative, empirical approaches increasingly risky.[4][81]
  • If instrumental goal convergence occurs, it may only do so in sufficiently intelligent agents.[82]
  • A superintelligence may find unconventional and radical solutions to assigned goals. Bostrom gives the example that if the objective is to make humans smile, a weak AI may perform as intended, while a superintelligence may decide a better solution is to "take control of the world and stick electrodes into the facial muscles of humans to cause constant, beaming grins."[47]
  • A superintelligence in creation could gain some awareness of what it is, where it is in development (training, testing, deployment, etc.), and how it is being monitored, and use this information to deceive its handlers.[83] Bostrom writes that such an AI could feign alignment to prevent human interference until it achieves a "decisive strategic advantage" that allows it to take control.[4]
  • Analyzing the internals and interpreting the behavior of current large language models is difficult. And it could be even more difficult for larger and more intelligent models.[81]

Alternatively, some find reason to believe superintelligences would be better able to understand morality, human values, and complex goals. Bostrom writes, "A future superintelligence occupies an epistemically superior vantage point: its beliefs are (probably, on most topics) more likely than ours to be true".[4]

In 2023, OpenAI started a project called "Superalignment" to solve the alignment of superintelligences in four years. It called this an especially important challenge, as it said superintelligence may be achieved within a decade. Its strategy involves automating alignment research using artificial intelligence.[84]

Difficulty of making a flawless design

Artificial Intelligence: A Modern Approach, a widely used undergraduate AI textbook,[85][86] says that superintelligence "might mean the end of the human race".[1] It states: "Almost any technology has the potential to cause harm in the wrong hands, but with [superintelligence], we have the new problem that the wrong hands might belong to the technology itself."[1] Even if the system designers have good intentions, two difficulties are common to both AI and non-AI computer systems:[1]

  • The system's implementation may contain initially unnoticed but subsequently catastrophic bugs. An analogy is space probes: despite the knowledge that bugs in expensive space probes are hard to fix after launch, engineers have historically not been able to prevent catastrophic bugs from occurring.[87][88]
  • No matter how much time is put into pre-deployment design, a system's specifications often result in unintended behavior the first time it encounters a new scenario. For example, Microsoft's Tay behaved inoffensively during pre-deployment testing, but was too easily baited into offensive behavior when it interacted with real users.[20]

AI systems uniquely add a third problem: that even given "correct" requirements, bug-free implementation, and initial good behavior, an AI system's dynamic learning capabilities may cause it to develop unintended behavior, even without unanticipated external scenarios. An AI may partly botch an attempt to design a new generation of itself and accidentally create a successor AI that is more powerful than itself but that no longer maintains the human-compatible moral values preprogrammed into the original AI. For a self-improving AI to be completely safe, it would need not only to be bug-free, but to be able to design successor systems that are also bug-free.[1][89]

Orthogonality thesis

Some skeptics, such as Timothy B. Lee of Vox, argue that any superintelligent program we create will be subservient to us, that the superintelligence will (as it grows more intelligent and learns more facts about the world) spontaneously learn moral truth compatible with our values and adjust its goals accordingly, or that we are either intrinsically or convergently valuable from the perspective of an artificial intelligence.[90]

Bostrom's "orthogonality thesis" argues instead that, with some technical caveats, almost any level of "intelligence" or "optimization power" can be combined with almost any ultimate goal. If a machine is given the sole purpose to enumerate the decimals of pi, then no moral and ethical rules will stop it from achieving its programmed goal by any means. The machine may use all available physical and informational resources to find as many decimals of pi as it can.[91] Bostrom warns against anthropomorphism: a human will set out to accomplish their projects in a manner that they consider reasonable, while an artificial intelligence may hold no regard for its existence or for the welfare of humans around it, instead caring only about completing the task.[92]

Stuart Armstrong argues that the orthogonality thesis follows logically from the philosophical "is-ought distinction" argument against moral realism. He claims that even if there are moral facts provable by any "rational" agent, the orthogonality thesis still holds: it is still possible to create a non-philosophical "optimizing machine" that can strive toward some narrow goal but that has no incentive to discover any "moral facts" such as those that could get in the way of goal completion. Another argument he makes is that any fundamentally friendly AI could be made unfriendly with modifications as simple as negating its utility function. Armstrong further argues that if the orthogonality thesis is false, there must be some immoral goals that AIs can never achieve, which he finds implausible.[93]

Skeptic Michael Chorost explicitly rejects Bostrom's orthogonality thesis, arguing that "by the time [the AI] is in a position to imagine tiling the Earth with solar panels, it'll know that it would be morally wrong to do so."[94] Chorost argues that "an A.I. will need to desire certain states and dislike others. Today's software lacks that ability—and computer scientists have not a clue how to get it there. Without wanting, there's no impetus to do anything. Today's computers can't even want to keep existing, let alone tile the world in solar panels."[94]

Anthropomorphic arguments

Anthropomorphic arguments assume that, as machines become more intelligent, they will begin to display many human traits, such as morality or a thirst for power. Although anthropomorphic scenarios are common in fiction, most scholars writing about the existential risk of artificial intelligence reject them.[18] Instead, advanced AI systems are typically modeled as intelligent agents.

The academic debate is between those who worry that AI might threaten humanity and those who believe it would not. Both sides of this debate have framed the other side's arguments as illogical anthropomorphism.[18] Those skeptical of AGI risk accuse their opponents of anthropomorphism for assuming that an AGI would naturally desire power; those concerned about AGI risk accuse skeptics of anthropomorphism for believing an AGI would naturally value or infer human ethical norms.[18][95]

Evolutionary psychologist Steven Pinker, a skeptic, argues that "AI dystopias project a parochial alpha-male psychology onto the concept of intelligence. They assume that superhumanly intelligent robots would develop goals like deposing their masters or taking over the world"; perhaps instead "artificial intelligence will naturally develop along female lines: fully capable of solving problems, but with no desire to annihilate innocents or dominate the civilization."[96] Facebook's director of AI research, Yann LeCun, has said: "Humans have all kinds of drives that make them do bad things to each other, like the self-preservation instinct... Those drives are programmed into our brain but there is absolutely no reason to build robots that have the same kind of drives".[97]

Despite other differences, the x-risk school[b] agrees with Pinker that an advanced AI would not destroy humanity out of emotion such as revenge or anger, that questions of consciousness are not relevant to assess the risk,[98] and that computer systems do not generally have a computational equivalent of testosterone.[99] They think that power-seeking or self-preservation behaviors emerge in the AI as a way to achieve its true goals, according to the concept of instrumental convergence.

Other sources of risk

Bostrom and others have said that a race to be the first to create AGI could lead to shortcuts in safety, or even to violent conflict.[100][101] Roman Yampolskiy and others warn that a malevolent AGI could be created by design, for example by a military, a government, a sociopath, or a corporation, to benefit from, control, or subjugate certain groups of people, as in cybercrime,[102][103] or that a malevolent AGI could choose the goal of increasing human suffering, for example of those people who did not assist it during the information explosion phase.[3]:158

Scenarios

Some scholars have proposed hypothetical scenarios to illustrate some of their concerns.

Treacherous turn

In Superintelligence, Bostrom expresses concern that even if the timeline for superintelligence turns out to be predictable, researchers might not take sufficient safety precautions, in part because "it could be the case that when dumb, smarter is safe; yet when smart, smarter is more dangerous". He suggests a scenario where, over decades, AI becomes more powerful. Widespread deployment is initially marred by occasional accidents—a driverless bus swerves into the oncoming lane, or a military drone fires into an innocent crowd. Many activists call for tighter oversight and regulation, and some even predict impending catastrophe. But as development continues, the activists are proven wrong. As automotive AI becomes smarter, it suffers fewer accidents; as military robots achieve more precise targeting, they cause less collateral damage. Based on the data, scholars mistakenly infer a broad lesson: the smarter the AI, the safer it is. "And so we boldly go—into the whirling knives", as the superintelligent AI takes a "treacherous turn" and exploits a decisive strategic advantage.[4]

Life 3.0

In Max Tegmark's 2017 book Life 3.0, a corporation's "Omega team" creates an extremely powerful AI able to moderately improve its own source code in a number of areas. After a certain point, the team chooses to publicly downplay the AI's ability in order to avoid regulation or confiscation of the project. For safety, the team keeps the AI in a box where it is mostly unable to communicate with the outside world, and uses it to make money, by diverse means such as Amazon Mechanical Turk tasks, production of animated films and TV shows, and development of biotech drugs, with profits invested back into further improving AI. The team next tasks the AI with astroturfing an army of pseudonymous citizen journalists and commentators in order to gain political influence to use "for the greater good" to prevent wars. The team faces risks that the AI could try to escape by inserting "backdoors" in the systems it designs, by hidden messages in its produced content, or by using its growing understanding of human behavior to persuade someone into letting it free. The team also faces risks that its decision to box the project will delay the project long enough for another project to overtake it.[104][105]

Perspectives

The thesis that AI could pose an existential risk provokes a wide range of reactions in the scientific community and in the public at large, but many of the opposing viewpoints share common ground.

Observers tend to agree that AI has significant potential to improve society.[106][107] The Asilomar AI Principles, which contain only those principles agreed to by 90% of the attendees of the Future of Life Institute's Beneficial AI 2017 conference,[105] also agree in principle that "There being no consensus, we should avoid strong assumptions regarding upper limits on future AI capabilities" and "Advanced AI could represent a profound change in the history of life on Earth, and should be planned for and managed with commensurate care and resources."[108][109]

Conversely, many skeptics agree that ongoing research into the implications of artificial general intelligence is valuable. Skeptic Martin Ford has said: "I think it seems wise to apply something like Dick Cheney's famous '1 Percent Doctrine' to the specter of advanced artificial intelligence: the odds of its occurrence, at least in the foreseeable future, may be very low—but the implications are so dramatic that it should be taken seriously".[110] Similarly, an otherwise skeptical Economist wrote in 2014 that "the implications of introducing a second intelligent species onto Earth are far-reaching enough to deserve hard thinking, even if the prospect seems remote".[46]

AI safety advocates such as Bostrom and Tegmark have criticized the mainstream media's use of "those inane Terminator pictures" to illustrate AI safety concerns: "It can't be much fun to have aspersions cast on one's academic discipline, one's professional community, one's life work ... I call on all sides to practice patience and restraint, and to engage in direct dialogue and collaboration as much as possible."[105][111] Toby Ord wrote that the idea that an AI takeover requires robots is a misconception, arguing that the ability to spread content through the internet is more dangerous, and that the most destructive people in history stood out by their ability to convince, not their physical strength.[67]

A 2022 expert survey with a 17% response rate gave a median expectation of 5–10% for the possibility of human extinction from artificial intelligence.[14][112]

Endorsement

The thesis that AI poses an existential risk, and that this risk needs much more attention than it currently gets, has been endorsed by many computer scientists and public figures, including Alan Turing,[a] the most-cited computer scientist Geoffrey Hinton,[113] Elon Musk,[11] OpenAI CEO Sam Altman,[12][114] Bill Gates, and Stephen Hawking.[114] Endorsers of the thesis sometimes express bafflement at skeptics: Gates says he does not "understand why some people are not concerned",[115] and Hawking criticized widespread indifference in his 2014 editorial:

So, facing possible futures of incalculable benefits and risks, the experts are surely doing everything possible to ensure the best outcome, right? Wrong. If a superior alien civilisation sent us a message saying, 'We'll arrive in a few decades,' would we just reply, 'OK, call us when you get here—we'll leave the lights on?' Probably not—but this is more or less what is happening with AI.[31]

Concern over risk from artificial intelligence has led to some high-profile donations and investments. In 2015, Peter Thiel, Amazon Web Services, and Musk and others jointly committed $1 billion to OpenAI, consisting of a for-profit corporation and the nonprofit parent company, which says it aims to champion responsible AI development.[116] Facebook co-founder Dustin Moskovitz has funded and seeded multiple labs working on AI Alignment,[117] notably $5.5 million in 2016 to launch the Centre for Human-Compatible AI led by Professor Stuart Russell.[118] In January 2015, Elon Musk donated $10 million to the Future of Life Institute to fund research on understanding AI decision making. The institute's goal is to "grow wisdom with which we manage" the growing power of technology. Musk also funds companies developing artificial intelligence such as DeepMind and Vicarious to "just keep an eye on what's going on with artificial intelligence,[119] saying "I think there is potentially a dangerous outcome there."[120][121]

In early statements on the topic, Geoffrey Hinton, a major pioneer of deep learning, noted that "there is not a good track record of less intelligent things controlling things of greater intelligence", but said he continued his research because "the prospect of discovery is too sweet".[122][123] In 2023 Hinton quit his job at Google in order to speak out about existential risk from AI. He explained that his increased concern was driven by concerns that superhuman AI might be closer than he previously believed, saying: "I thought it was way off. I thought it was 30 to 50 years or even longer away. Obviously, I no longer think that." He also remarked, "Look at how it was five years ago and how it is now. Take the difference and propagate it forwards. That's scary."[124]

In his 2020 book The Precipice: Existential Risk and the Future of Humanity, Toby Ord, a Senior Research Fellow at Oxford University's Future of Humanity Institute, estimates the total existential risk from unaligned AI over the next 100 years at about one in ten.[67]

Skepticism

Baidu Vice President Andrew Ng said in 2015 that AI existential risk is "like worrying about overpopulation on Mars when we have not even set foot on the planet yet."[96][125] For the danger of uncontrolled advanced AI to be realized, the hypothetical AI may have to overpower or outthink any human, which some experts argue is a possibility far enough in the future to not be worth researching.[126][127]

Skeptics who believe AGI is not a short-term possibility often argue that concern about existential risk from AI is unhelpful because it could distract people from more immediate concerns about AI's impact, because it could lead to government regulation or make it more difficult to fund AI research, or because it could damage the field's reputation.[128] AI and AI ethics researchers Timnit Gebru, Emily M. Bender, Margaret Mitchell, and Angelina McMillan-Major have argued that discussion of existential risk distracts from the immediate, ongoing harms from AI taking place today, such as data theft, worker exploitation, bias, and concentration of power.[129] They further note the association between those warning of existential risk and longtermism, which they describe as a "dangerous ideology" for its unscientific and utopian nature.[130] Gebru and Émile P. Torres have suggested that obsession with AGI is part of a pattern of intellectual movements called TESCREAL.[131]

Wired editor Kevin Kelly argues that natural intelligence is more nuanced than AGI proponents believe, and that intelligence alone is not enough to achieve major scientific and societal breakthroughs. He argues that intelligence consists of many dimensions that are not well understood, and that conceptions of an 'intelligence ladder' are misleading. He notes the crucial role real-world experiments play in the scientific method, and that intelligence alone is no substitute for these.[132]

Meta chief AI scientist Yann LeCun says that AI can be made safe via continuous and iterative refinement, similar to what happened in the past with cars or rockets, and that AI will have no desire to take control.[133]

Several skeptics emphasize the potential near-term benefits of AI. Meta CEO Mark Zuckerberg believes AI will "unlock a huge amount of positive things", such as curing disease and increasing the safety of autonomous cars.[134]

During a 2016 Wired interview of President Barack Obama and MIT Media Lab's Joi Ito, Ito said:

There are a few people who believe that there is a fairly high-percentage chance that a generalized AI will happen in the next 10 years. But the way I look at it is that in order for that to happen, we're going to need a dozen or two different breakthroughs. So you can monitor when you think these breakthroughs will happen.

Obama added:[135][136]

And you just have to have somebody close to the power cord. [Laughs.] Right when you see it about to happen, you gotta yank that electricity out of the wall, man.

Hillary Clinton wrote in What Happened:

Technologists... have warned that artificial intelligence could one day pose an existential security threat. Musk has called it "the greatest risk we face as a civilization". Think about it: Have you ever seen a movie where the machines start thinking for themselves that ends well? Every time I went out to Silicon Valley during the campaign, I came home more alarmed about this. My staff lived in fear that I'd start talking about "the rise of the robots" in some Iowa town hall. Maybe I should have. In any case, policy makers need to keep up with technology as it races ahead, instead of always playing catch-up.[137]

Public surveys

In 2018, a SurveyMonkey poll of the American public by USA Today found 68% thought the real current threat remains "human intelligence", but also found that 43% said superintelligent AI, if it were to happen, would result in "more harm than good", and that 38% said it would do "equal amounts of harm and good".[138]

An April 2023 YouGov poll of US adults found 46% of respondents were "somewhat concerned" or "very concerned" about "the possibility that AI will cause the end of the human race on Earth", compared with 40% who were "not very concerned" or "not at all concerned."[139]

According to an August 2023 survey by the Pew Research Centers, 52% of Americans felt more concerned than excited about new AI developments; nearly a third felt as equally concerned and excited. More Americans saw that AI would have a more helpful than hurtful impact on several areas, from healthcare and vehicle safety to product search and customer service. The main exception is privacy: 53% of Americans believe AI will lead to higher exposure of their personal information.[140]

Mitigation

Many scholars concerned about AGI existential risk believe that extensive research into the "control problem" is essential. This problem involves determining which safeguards, algorithms, or architectures can be implemented to increase the likelihood that a recursively-improving AI remains friendly after achieving superintelligence.[4][141] Social measures are also proposed to mitigate AGI risks,[142][143] such as a UN-sponsored "Benevolent AGI Treaty" to ensure that only altruistic AGIs are created.[144] Additionally, an arms control approach and a global peace treaty grounded in international relations theory have been suggested, potentially for an artificial superintelligence to be a signatory.[145][146]

Researchers at Google have proposed research into general "AI safety" issues to simultaneously mitigate both short-term risks from narrow AI and long-term risks from AGI.[147][148] A 2020 estimate places global spending on AI existential risk somewhere between $10 and $50 million, compared with global spending on AI around perhaps $40 billion. Bostrom suggests prioritizing funding for protective technologies over potentially dangerous ones.[77] Some, like Elon Musk, advocate radical human cognitive enhancement, such as direct neural linking between humans and machines; others argue that these technologies may pose an existential risk themselves.[149][150] Another proposed method is closely monitoring or "boxing in" an early-stage AI to prevent it from becoming too powerful. A dominant, aligned superintelligent AI might also mitigate risks from rival AIs, although its creation could present its own existential dangers.[151]

Institutions such as the Alignment Research Center,[152] the Machine Intelligence Research Institute,[153][154] the Future of Life Institute, the Centre for the Study of Existential Risk, and the Center for Human-Compatible AI[155] are actively engaged in researching AI risk and safety.

Views on banning and regulation

Banning

Some scholars have said that even if AGI poses an existential risk, attempting to ban research into artificial intelligence is still unwise, and probably futile.[156][157][158] Skeptics consider AI regulation pointless, as no existential risk exists. But scholars who believe in the risk argue that relying on AI industry insiders to regulate or constrain AI research is impractical due to conflicts of interest.[159] They also agree with skeptics that banning research would be unwise, as research could be moved to countries with looser regulations or conducted covertly.[159] Additional challenges to bans or regulation include technology entrepreneurs' general skepticism of government regulation and potential incentives for businesses to resist regulation and politicize the debate.[160]

Regulation

In March 2023, the Future of Life Institute drafted Pause Giant AI Experiments: An Open Letter, a petition calling on major AI developers to agree on a verifiable six-month pause of any systems "more powerful than GPT-4" and to use that time to institute a framework for ensuring safety; or, failing that, for governments to step in with a moratorium. The letter referred to the possibility of "a profound change in the history of life on Earth" as well as potential risks of AI-generated propaganda, loss of jobs, human obsolescence, and society-wide loss of control.[107][161] The letter was signed by prominent personalities in AI but also criticized for not focusing on current harms,[162] missing technical nuance about when to pause,[163] or not going far enough.[164]

Musk called for some sort of regulation of AI development as early as 2017. According to NPR, he is "clearly not thrilled" to be advocating government scrutiny that could impact his own industry, but believes the risks of going completely without oversight are too high: "Normally the way regulations are set up is when a bunch of bad things happen, there's a public outcry, and after many years a regulatory agency is set up to regulate that industry. It takes forever. That, in the past, has been bad but not something which represented a fundamental risk to the existence of civilisation." Musk states the first step would be for the government to gain "insight" into the actual status of current research, warning that "Once there is awareness, people will be extremely afraid... [as] they should be." In response, politicians expressed skepticism about the wisdom of regulating a technology that is still in development.[165][166][167]

In 2021 the United Nations (UN) considered banning autonomous lethal weapons, but consensus could not be reached.[168] In July 2023 the UN Security Council for the first time held a session to consider the risks and threats posed by AI to world peace and stability, along with potential benefits.[169][170] Secretary-General António Guterres advocated the creation of a global watchdog to oversee the emerging technology, saying, "Generative AI has enormous potential for good and evil at scale. Its creators themselves have warned that much bigger, potentially catastrophic and existential risks lie ahead."[17] At the council session, Russia said it believes AI risks are too poorly understood to be considered a threat to global stability. China argued against strict global regulation, saying countries should be able to develop their own rules, while also saying they opposed the use of AI to "create military hegemony or undermine the sovereignty of a country".[169]

Regulation of conscious AGIs focuses on integrating them with existing human society and can be divided into considerations of their legal standing and of their moral rights.[171] AI arms control will likely require the institutionalization of new international norms embodied in effective technical specifications combined with active monitoring and informal diplomacy by communities of experts, together with a legal and political verification process.[172][113]

In July 2023, the US government secured voluntary safety commitments from major tech companies, including OpenAI, Amazon, Google, Meta, and Microsoft. The companies agreed to implement safeguards, including third-party oversight and security testing by independent experts, to address concerns related to AI's potential risks and societal harms. The parties framed the commitments as an intermediate step while regulations are formed. Amba Kak, executive director of the AI Now Institute, said, "A closed-door deliberation with corporate actors resulting in voluntary safeguards isn't enough" and called for public deliberation and regulations of the kind to which companies would not voluntarily agree.[173][174]

In October 2023, U.S. President Joe Biden issued an executive order on the "Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence".[175] Alongside other requirements, the order mandates the development of guidelines for AI models that permit the "evasion of human control".

See also

Notes

  1. ^ a b In a 1951 lecture[9] Turing argued that "It seems probable that once the machine thinking method had started, it would not take long to outstrip our feeble powers. There would be no question of the machines dying, and they would be able to converse with each other to sharpen their wits. At some stage therefore we should have to expect the machines to take control, in the way that is mentioned in Samuel Butler's Erewhon". Also in a lecture broadcast on the BBC[10] he expressed the opinion: "If a machine can think, it might think more intelligently than we do, and then where should we be? Even if we could keep the machines in a subservient position, for instance by turning off the power at strategic moments, we should, as a species, feel greatly humbled... This new danger... is certainly something which can give us anxiety."
  2. ^ as interpreted by Seth Baum

References

  1. ^ a b c d e f g Russell, Stuart; Norvig, Peter (2009). "26.3: The Ethics and Risks of Developing Artificial Intelligence". Artificial Intelligence: A Modern Approach. Prentice Hall. ISBN 978-0-13-604259-4.
  2. ^ Bostrom, Nick (2002). "Existential risks". Journal of Evolution and Technology. 9 (1): 1–31.
  3. ^ a b Turchin, Alexey; Denkenberger, David (3 May 2018). "Classification of global catastrophic risks connected with artificial intelligence". AI & Society. 35 (1): 147–163. doi:10.1007/s00146-018-0845-5. ISSN 0951-5666. S2CID 19208453.
  4. ^ a b c d e f g h i j k l m n o p q Bostrom, Nick (2014). Superintelligence: Paths, Dangers, Strategies (First ed.). Oxford University Press. ISBN 978-0-19-967811-2.
  5. ^ Vynck, Gerrit De (23 May 2023). "The debate over whether AI will destroy us is dividing Silicon Valley". Washington Post. ISSN 0190-8286. Retrieved 27 July 2023.
  6. ^ Metz, Cade (10 June 2023). "How Could A.I. Destroy Humanity?". The New York Times. ISSN 0362-4331. Retrieved 27 July 2023.
  7. ^ "'Godfather of artificial intelligence' weighs in on the past and potential of AI". www.cbsnews.com. 25 March 2023. Retrieved 10 April 2023.
  8. ^ "How Rogue AIs may Arise". yoshuabengio.org. 26 May 2023. Retrieved 26 May 2023.
  9. ^ Turing, Alan (1951). Intelligent machinery, a heretical theory (Speech). Lecture given to '51 Society'. Manchester: The Turing Digital Archive. Archived from the original on 26 September 2022. Retrieved 22 July 2022.
  10. ^ Turing, Alan (15 May 1951). "Can digital computers think?". Automatic Calculating Machines. Episode 2. BBC. Can digital computers think?.
  11. ^ a b Parkin, Simon (14 June 2015). "Science fiction no more? Channel 4's Humans and our rogue AI obsessions". The Guardian. Archived from the original on 5 February 2018. Retrieved 5 February 2018.
  12. ^ a b Jackson, Sarah. "The CEO of the company behind AI chatbot ChatGPT says the worst-case scenario for artificial intelligence is 'lights out for all of us'". Business Insider. Retrieved 10 April 2023.
  13. ^ "The AI Dilemma". www.humanetech.com. Retrieved 10 April 2023. 50% of AI researchers believe there's a 10% or greater chance that humans go extinct from our inability to control AI.
  14. ^ a b "2022 Expert Survey on Progress in AI". AI Impacts. 4 August 2022. Retrieved 10 April 2023.
  15. ^ Roose, Kevin (30 May 2023). "A.I. Poses 'Risk of Extinction,' Industry Leaders Warn". The New York Times. ISSN 0362-4331. Retrieved 3 June 2023.
  16. ^ Sunak, Rishi (14 June 2023). "Rishi Sunak Wants the U.K. to Be a Key Player in Global AI Regulation". Time.
  17. ^ a b Fung, Brian (18 July 2023). "UN Secretary General embraces calls for a new UN agency on AI in the face of 'potentially catastrophic and existential risks'". CNN Business. Retrieved 20 July 2023.
  18. ^ a b c d e Yudkowsky, Eliezer (2008). "Artificial Intelligence as a Positive and Negative Factor in Global Risk" (PDF). Global Catastrophic Risks: 308–345. Bibcode:2008gcr..book..303Y. Archived (PDF) from the original on 2 March 2013. Retrieved 27 August 2018.
  19. ^ Russell, Stuart; Dewey, Daniel; Tegmark, Max (2015). "Research Priorities for Robust and Beneficial Artificial Intelligence" (PDF). AI Magazine. Association for the Advancement of Artificial Intelligence: 105–114. arXiv:1602.03506. Bibcode:2016arXiv160203506R. Archived (PDF) from the original on 4 August 2019. Retrieved 10 August 2019., cited in "AI Open Letter - Future of Life Institute". Future of Life Institute. January 2015. Archived from the original on 10 August 2019. Retrieved 9 August 2019.
  20. ^ a b c Dowd, Maureen (April 2017). "Elon Musk's Billion-Dollar Crusade to Stop the A.I. Apocalypse". The Hive. Archived from the original on 26 July 2018. Retrieved 27 November 2017.
  21. ^ "AlphaGo Zero: Starting from scratch". www.deepmind.com. 18 October 2017. Retrieved 28 July 2023.
  22. ^ Breuer, Hans-Peter. 'Samuel Butler's "the Book of the Machines" and the Argument from Design.' Archived 15 March 2023 at the Wayback Machine Modern Philology, Vol. 72, No. 4 (May 1975), pp. 365–383.
  23. ^ Turing, A M (1996). "Intelligent Machinery, A Heretical Theory". 1951, Reprinted Philosophia Mathematica. 4 (3): 256–260. doi:10.1093/philmat/4.3.256.
  24. ^ Hilliard, Mark (2017). "The AI apocalypse: will the human race soon be terminated?". The Irish Times. Archived from the original on 22 May 2020. Retrieved 15 March 2020.
  25. ^ I.J. Good, "Speculations Concerning the First Ultraintelligent Machine" Archived 2011-11-28 at the Wayback Machine (HTML ), Advances in Computers, vol. 6, 1965.
  26. ^ Russell, Stuart J.; Norvig, Peter (2003). "Section 26.3: The Ethics and Risks of Developing Artificial Intelligence". Artificial Intelligence: A Modern Approach. Upper Saddle River, New Jersey: Prentice Hall. ISBN 978-0-13-790395-5. Similarly, Marvin Minsky once suggested that an AI program designed to solve the Riemann Hypothesis might end up taking over all the resources of Earth to build more powerful supercomputers to help achieve its goal.
  27. ^ Barrat, James (2013). Our final invention: artificial intelligence and the end of the human era (First ed.). New York: St. Martin's Press. ISBN 978-0-312-62237-4. In the bio, playfully written in the third person, Good summarized his life's milestones, including a probably never before seen account of his work at Bletchley Park with Turing. But here's what he wrote in 1998 about the first superintelligence, and his late-in-the-game U-turn: [The paper] 'Speculations Concerning the First Ultra-intelligent Machine' (1965)...began: 'The survival of man depends on the early construction of an ultra-intelligent machine.' Those were his [Good's] words during the Cold War, and he now suspects that 'survival' should be replaced by 'extinction.' He thinks that, because of international competition, we cannot prevent the machines from taking over. He thinks we are lemmings. He said also that 'probably Man will construct the deus ex machina in his own image.'
  28. ^ Anderson, Kurt (26 November 2014). "Enthusiasts and Skeptics Debate Artificial Intelligence". Vanity Fair. Archived from the original on 22 January 2016. Retrieved 30 January 2016.
  29. ^ Metz, Cade (9 June 2018). "Mark Zuckerberg, Elon Musk and the Feud Over Killer Robots". The New York Times. Archived from the original on 15 February 2021. Retrieved 3 April 2019.
  30. ^ Hsu, Jeremy (1 March 2012). "Control dangerous AI before it controls us, one expert says". NBC News. Archived from the original on 2 February 2016. Retrieved 28 January 2016.
  31. ^ a b c "Stephen Hawking: 'Transcendence looks at the implications of artificial intelligence – but are we taking AI seriously enough?'". The Independent (UK). Archived from the original on 25 September 2015. Retrieved 3 December 2014.
  32. ^ "Stephen Hawking warns artificial intelligence could end mankind". BBC. 2 December 2014. Archived from the original on 30 October 2015. Retrieved 3 December 2014.
  33. ^ Eadicicco, Lisa (28 January 2015). "Bill Gates: Elon Musk Is Right, We Should All Be Scared Of Artificial Intelligence Wiping Out Humanity". Business Insider. Archived from the original on 26 February 2016. Retrieved 30 January 2016.
  34. ^ "Research Priorities for Robust and Beneficial Artificial Intelligence: an Open Letter". Future of Life Institute. Archived from the original on 15 January 2015. Retrieved 23 October 2015.
  35. ^ "Anticipating artificial intelligence". Nature. 532 (7600): 413. 2016. Bibcode:2016Natur.532Q.413.. doi:10.1038/532413a. ISSN 1476-4687. PMID 27121801. S2CID 4399193.
  36. ^ Christian, Brian (6 October 2020). The Alignment Problem: Machine Learning and Human Values. W. W. Norton & Company. ISBN 978-0-393-63582-9. Archived from the original on 5 December 2021. Retrieved 5 December 2021.
  37. ^ Dignum, Virginia (26 May 2021). "AI – the people and places that make, use and manage it". Nature. 593 (7860): 499–500. Bibcode:2021Natur.593..499D. doi:10.1038/d41586-021-01397-x. S2CID 235216649.
  38. ^ "Elon Musk among experts urging a halt to AI training". BBC News. 29 March 2023. Retrieved 9 June 2023.
  39. ^ "Statement on AI Risk". Center for AI Safety. Retrieved 8 June 2023.
  40. ^ "Artificial intelligence could lead to extinction, experts warn". BBC News. 30 May 2023. Retrieved 8 June 2023.
  41. ^ "DeepMind and Google: the battle to control artificial intelligence". The Economist. ISSN 0013-0613. Retrieved 12 July 2023.
  42. ^ "AI timelines: What do experts in artificial intelligence expect for the future?". Our World in Data. Retrieved 12 July 2023.
  43. ^ De Vynck, Gerrit (20 May 2023). "The debate over whether AI will destroy us is dividing Silicon Valley". The Washington Post.
  44. ^ "'The Godfather of A.I.' just quit Google and says he regrets his life's work because it can be hard to stop 'bad actors from using it for bad things'". Fortune. Retrieved 12 July 2023.
  45. ^ "Everything you need to know about superintelligence". Spiceworks. Retrieved 14 July 2023.
  46. ^ a b c Babauta, Leo. "A Valuable New Book Explores The Potential Impacts Of Intelligent Machines On Human Life". Business Insider. Retrieved 19 March 2024.
  47. ^ a b Bostrom, Nick (27 April 2015), What happens when our computers get smarter than we are?, retrieved 13 July 2023.
  48. ^ "Governance of superintelligence". openai.com. Retrieved 12 July 2023.
  49. ^ "Overcoming Bias: I Still Don't Get Foom". www.overcomingbias.com. Archived from the original on 4 August 2017. Retrieved 20 September 2017.
  50. ^ Cotton-Barratt, Owen; Ord, Toby (12 August 2014). "Strategic considerations about different speeds of AI takeoff". The Future of Humanity Institute. Retrieved 12 July 2023.
  51. ^ Tegmark, Max (25 April 2023). "The 'Don't Look Up' Thinking That Could Doom Us With AI". Time. Retrieved 14 July 2023. As if losing control to Chinese minds were scarier than losing control to alien digital minds that don't care about humans. [...] it's clear by now that the space of possible alien minds is vastly larger than that.
  52. ^ "19 – Mechanistic Interpretability with Neel Nanda". AXRP – the AI X-risk Research Podcast. 4 February 2023. Retrieved 13 July 2023. it's plausible to me that the main thing we need to get done is noticing specific circuits to do with deception and specific dangerous capabilities like that and situational awareness and internally-represented goals.
  53. ^ "Superintelligence Is Not Omniscience". AI Impacts. 7 April 2023. Retrieved 16 April 2023.
  54. ^ a b c d e f g h i Hendrycks, Dan; Mazeika, Mantas; Woodside, Thomas (21 June 2023). "An Overview of Catastrophic AI Risks". arXiv:2306.12001 [cs.CY].
  55. ^ Taylor, Josh; Hern, Alex (2 May 2023). "'Godfather of AI' Geoffrey Hinton quits Google and warns over dangers of misinformation". The Guardian. ISSN 0261-3077. Retrieved 13 July 2023.
  56. ^ "How NATO is preparing for a new era of AI cyber attacks". euronews. 26 December 2022. Retrieved 13 July 2023.
  57. ^ "ChatGPT and the new AI are wreaking havoc on cybersecurity in exciting and frightening ways". ZDNET. Retrieved 13 July 2023.
  58. ^ Toby Shevlane; Sebastian Farquhar; Ben Garfinkel; Mary Phuong; Jess Whittlestone; Jade Leung; Daniel Kokotajlo; Nahema Marchal; Markus Anderljung; Noam Kolt; Lewis Ho; Divya Siddarth; Shahar Avin; Will Hawkins; Been Kim; Iason Gabriel; Vijay Bolina; Jack Clark; Yoshua Bengio; Paul Christiano; Allan Dafoe (24 May 2023). "Model evaluation for extreme risks". arXiv:2305.15324 [cs.AI].
  59. ^ Urbina, Fabio; Lentzos, Filippa; Invernizzi, Cédric; Ekins, Sean (7 March 2022). "Dual use of artificial-intelligence-powered drug discovery". Nature Machine Intelligence. 4 (3): 189–191. doi:10.1038/s42256-022-00465-9. ISSN 2522-5839. PMC 9544280. PMID 36211133.
  60. ^ Walter, Yoshija (27 March 2023). "The rapid competitive economy of machine learning development: a discussion on the social risks and benefits". AI and Ethics. 4 (2): 1. doi:10.1007/s43681-023-00276-7.
  61. ^ "The AI Arms Race Is On. Start Worrying". Time. 16 February 2023. Retrieved 17 July 2023.
  62. ^ Brimelow, Ben. "The short film 'Slaughterbots' depicts a dystopian future of killer drones swarming the world". Business Insider. Retrieved 20 July 2023.
  63. ^ Mecklin, John (17 July 2023). "'Artificial Escalation': Imagining the future of nuclear risk". Bulletin of the Atomic Scientists. Retrieved 20 July 2023.
  64. ^ Bostrom, Nick (2013). "Existential Risk Prevention as Global Priority" (PDF). Global Policy. 4 (1): 15–3. doi:10.1111/1758-5899.12002 – via Existential Risk.
  65. ^ Doherty, Ben (17 May 2018). "Climate change an 'existential security risk' to Australia, Senate inquiry says". The Guardian. ISSN 0261-3077. Retrieved 16 July 2023.
  66. ^ MacAskill, William (2022). What we owe the future. New York, New York: Basic Books. ISBN 978-1-5416-1862-6.
  67. ^ a b c d Ord, Toby (2020). "Chapter 5: Future Risks, Unaligned Artificial Intelligence". The Precipice: Existential Risk and the Future of Humanity. Bloomsbury Publishing. ISBN 978-1-5266-0021-9.
  68. ^ Samuelsson, Paul Conrad (June–July 2019). "Artificial Consciousness: Our Greatest Ethical Challenge". Philosophy Now. No. 132. Retrieved 19 August 2023.
  69. ^ Kateman, Brian (24 July 2023). "AI Should Be Terrified of Humans". Time. Retrieved 19 August 2023.
  70. ^ Fisher, Richard. "The intelligent monster that you should let eat you". www.bbc.com. Retrieved 19 August 2023.
  71. ^ More, Max (19 June 2023). "Existential Risk vs. Existential Opportunity: A balanced approach to AI risk". Extropic Thoughts. Retrieved 14 July 2023.
  72. ^ Omohundro, S. M. (2008, February). The basic AI drives. In AGI (Vol. 171, pp. 483–492).
  73. ^ Carlsmith, Joseph (16 June 2022). "Is Power-Seeking AI an Existential Risk?". arXiv:2206.13353 [cs.CY].
  74. ^ "'The Godfather of A.I.' warns of 'nightmare scenario' where artificial intelligence begins to seek power". Fortune. Retrieved 10 June 2023.
  75. ^ Wakefield, Jane (15 September 2015). "Why is Facebook investing in AI?". BBC News. Archived from the original on 2 December 2017. Retrieved 27 November 2017.
  76. ^ Yudkowsky, Eliezer (2011). "Complex Value Systems are Required to Realize Valuable Futures" (PDF). Archived (PDF) from the original on 29 September 2015. Retrieved 10 August 2020.
  77. ^ a b Ord, Toby (2020). The Precipice: Existential Risk and the Future of Humanity. Bloomsbury Publishing Plc. ISBN 978-1-5266-0019-6.
  78. ^ Yudkowsky, E. (2011, August). Complex value systems in friendly AI. In International Conference on Artificial General Intelligence (pp. 388–393). Germany: Springer, Berlin, Heidelberg.
  79. ^ Russell, Stuart (2014). "Of Myths and Moonshine". Edge. Archived from the original on 19 July 2016. Retrieved 23 October 2015.
  80. ^ Dietterich, Thomas; Horvitz, Eric (2015). "Rise of Concerns about AI: Reflections and Directions" (PDF). Communications of the ACM. 58 (10): 38–40. doi:10.1145/2770869. S2CID 20395145. Archived (PDF) from the original on 4 March 2016. Retrieved 23 October 2015.
  81. ^ a b Yudkowsky, Eliezer (29 March 2023). "The Open Letter on AI Doesn't Go Far Enough". Time. Retrieved 16 July 2023.
  82. ^ Bostrom, Nick (1 May 2012). "The Superintelligent Will: Motivation and Instrumental Rationality in Advanced Artificial Agents". Minds and Machines. 22 (2): 71–85. doi:10.1007/s11023-012-9281-3. ISSN 1572-8641. S2CID 254835485. as long as they possess a sufficient level of intelligence, agents having any of a wide range of final goals will pursue similar intermediary goals because they have instrumental reasons to do so.
  83. ^ Ngo, Richard; Chan, Lawrence; Sören Mindermann (22 February 2023). "The alignment problem from a deep learning perspective". arXiv:2209.00626 [cs.AI].
  84. ^ "Introducing Superalignment". openai.com. Retrieved 16 July 2023.
  85. ^ Tilli, Cecilia (28 April 2016). "Killer Robots? Lost Jobs?". Slate. Archived from the original on 11 May 2016. Retrieved 15 May 2016.
  86. ^ "Norvig vs. Chomsky and the Fight for the Future of AI". Tor.com. 21 June 2011. Archived from the original on 13 May 2016. Retrieved 15 May 2016.
  87. ^ Graves, Matthew (8 November 2017). "Why We Should Be Concerned About Artificial Superintelligence". Skeptic (US magazine). Vol. 22, no. 2. Archived from the original on 13 November 2017. Retrieved 27 November 2017.
  88. ^ Johnson, Phil (30 July 2015). "Houston, we have a bug: 9 famous software glitches in space". IT World. Archived from the original on 15 February 2019. Retrieved 5 February 2018.
  89. ^ Yampolskiy, Roman V. (8 April 2014). "Utility function security in artificially intelligent agents". Journal of Experimental & Theoretical Artificial Intelligence. 26 (3): 373–389. doi:10.1080/0952813X.2014.895114. S2CID 16477341. Nothing precludes sufficiently smart self-improving systems from optimising their reward mechanisms in order to optimisetheir current-goal achievement and in the process making a mistake leading to corruption of their reward functions.
  90. ^ "Will artificial intelligence destroy humanity? Here are 5 reasons not to worry". Vox. 22 August 2014. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  91. ^ Bostrom, Nick (2014). Superintelligence: Paths, Dangers, Strategies. Oxford, United Kingdom: Oxford University Press. p. 116. ISBN 978-0-19-967811-2.
  92. ^ Bostrom, Nick (2012). "Superintelligent Will" (PDF). Nick Bostrom. Archived (PDF) from the original on 28 November 2015. Retrieved 29 October 2015.
  93. ^ Armstrong, Stuart (1 January 2013). "General Purpose Intelligence: Arguing the Orthogonality Thesis". Analysis and Metaphysics. 12. Archived from the original on 11 October 2014. Retrieved 2 April 2020. Full text available here Archived 25 March 2020 at the Wayback Machine.
  94. ^ a b Chorost, Michael (18 April 2016). "Let Artificial Intelligence Evolve". Slate. Archived from the original on 27 November 2017. Retrieved 27 November 2017.
  95. ^ "Should humans fear the rise of the machine?". The Telegraph (UK). 1 September 2015. Archived from the original on 12 January 2022. Retrieved 7 February 2016.
  96. ^ a b Shermer, Michael (1 March 2017). "Apocalypse AI". Scientific American. 316 (3): 77. Bibcode:2017SciAm.316c..77S. doi:10.1038/scientificamerican0317-77. PMID 28207698. Archived from the original on 1 December 2017. Retrieved 27 November 2017.
  97. ^ "Intelligent Machines: What does Facebook want with AI?". BBC News. 14 September 2015. Retrieved 31 March 2023.
  98. ^ Baum, Seth (30 September 2018). "Countering Superintelligence Misinformation". Information. 9 (10): 244. doi:10.3390/info9100244. ISSN 2078-2489.
  99. ^ "The Myth Of AI". www.edge.org. Archived from the original on 11 March 2020. Retrieved 11 March 2020.
  100. ^ Bostrom, Nick, Superintelligence: paths, dangers, strategies (Audiobook), ISBN 978-1-5012-2774-5, OCLC 1061147095.
  101. ^ Sotala, Kaj; Yampolskiy, Roman V (19 December 2014). "Responses to catastrophic AGI risk: a survey". Physica Scripta. 90 (1): 12. Bibcode:2015PhyS...90a8001S. doi:10.1088/0031-8949/90/1/018001. ISSN 0031-8949.
  102. ^ Pistono, Federico; Yampolskiy, Roman V. (9 May 2016). Unethical Research: How to Create a Malevolent Artificial Intelligence. OCLC 1106238048.
  103. ^ Haney, Brian Seamus (2018). "The Perils & Promises of Artificial General Intelligence". SSRN Working Paper Series. doi:10.2139/ssrn.3261254. ISSN 1556-5068. S2CID 86743553.
  104. ^ Russell, Stuart (30 August 2017). "Artificial intelligence: The future is superintelligent". Nature. 548 (7669): 520–521. Bibcode:2017Natur.548..520R. doi:10.1038/548520a. S2CID 4459076.
  105. ^ a b c Tegmark, Max (2017). Life 3.0: Being Human in the Age of Artificial Intelligence (1st ed.). Mainstreaming AI Safety: Knopf. ISBN 978-0-451-48507-6.
  106. ^ Kumar, Vibhore. "Council Post: At The Dawn Of Artificial General Intelligence: Balancing Abundance With Existential Safeguards". Forbes. Retrieved 23 July 2023.
  107. ^ a b "Pause Giant AI Experiments: An Open Letter". Future of Life Institute. Retrieved 30 March 2023.
  108. ^ "AI Principles". Future of Life Institute. 11 August 2017. Archived from the original on 11 December 2017. Retrieved 11 December 2017.
  109. ^ "Elon Musk and Stephen Hawking warn of artificial intelligence arms race". Newsweek. 31 January 2017. Archived from the original on 11 December 2017. Retrieved 11 December 2017.
  110. ^ Ford, Martin (2015). "Chapter 9: Super-intelligence and the Singularity". Rise of the Robots: Technology and the Threat of a Jobless Future. Basic Books. ISBN 978-0-465-05999-7.
  111. ^ Bostrom, Nick (2016). "New Epilogue to the Paperback Edition". Superintelligence: Paths, Dangers, Strategies (Paperback ed.).
  112. ^ "Why Uncontrollable AI Looks More Likely Than Ever". Time. 27 February 2023. Retrieved 30 March 2023. It is therefore no surprise that according to the most recent AI Impacts Survey, nearly half of 731 leading AI researchers think there is at least a 10% chance that human-level AI would lead to an "extremely negative outcome," or existential risk.
  113. ^ a b Maas, Matthijs M. (6 February 2019). "How viable is international arms control for military artificial intelligence? Three lessons from nuclear weapons of mass destruction". Contemporary Security Policy. 40 (3): 285–311. doi:10.1080/13523260.2019.1576464. ISSN 1352-3260. S2CID 159310223.
  114. ^ a b "Impressed by artificial intelligence? Experts say AGI is coming next, and it has 'existential' risks". ABC News. 23 March 2023. Retrieved 30 March 2023.
  115. ^ Rawlinson, Kevin (29 January 2015). "Microsoft's Bill Gates insists AI is a threat". BBC News. Archived from the original on 29 January 2015. Retrieved 30 January 2015.
  116. ^ Washington Post (14 December 2015). "Tech titans like Elon Musk are spending $1 billion to save you from terminators". Chicago Tribune. Archived from the original on 7 June 2016.
  117. ^ "Doomsday to utopia: Meet AI's rival factions". Washington Post. 9 April 2023. Retrieved 30 April 2023.
  118. ^ "UC Berkeley – Center for Human-Compatible AI (2016)". Open Philanthropy. 27 June 2016. Retrieved 30 April 2023.
  119. ^ "The mysterious artificial intelligence company Elon Musk invested in is developing game-changing smart computers". Tech Insider. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  120. ^ Clark 2015a.
  121. ^ "Elon Musk Is Donating $10M Of His Own Money To Artificial Intelligence Research". Fast Company. 15 January 2015. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  122. ^ Tilli, Cecilia (28 April 2016). "Killer Robots? Lost Jobs?". Slate. Archived from the original on 11 May 2016. Retrieved 15 May 2016.
  123. ^ Khatchadourian, Raffi (23 November 2015). "The Doomsday Invention: Will artificial intelligence bring us utopia or destruction?". The New Yorker. Archived from the original on 29 April 2019. Retrieved 7 February 2016.
  124. ^ "Warning of AI's danger, pioneer Geoffrey Hinton quits Google to speak freely". www.arstechnica.com. 2023. Retrieved 23 July 2023.
  125. ^ Garling, Caleb (5 May 2015). "Andrew Ng: Why 'Deep Learning' Is a Mandate for Humans, Not Just Machines". Wired. Retrieved 31 March 2023.
  126. ^ "Is artificial intelligence really an existential threat to humanity?". MambaPost. 4 April 2023.
  127. ^ "The case against killer robots, from a guy actually working on artificial intelligence". Fusion.net. Archived from the original on 4 February 2016. Retrieved 31 January 2016.
  128. ^ "AI experts challenge 'doomer' narrative, including 'extinction risk' claims". VentureBeat. 31 May 2023. Retrieved 8 July 2023.
  129. ^ Coldewey, Devin (1 April 2023). "Ethicists fire back at 'AI Pause' letter they say 'ignores the actual harms'". TechCrunch. Retrieved 23 July 2023.
  130. ^ "DAIR (Distributed AI Research Institute)". DAIR Institute. Retrieved 23 July 2023.
  131. ^ Gebru, Timnit; Torres, Émile P. (14 April 2024). "The TESCREAL bundle: Eugenics and the promise of utopia through artificial general intelligence". First Monday. doi:10.5210/fm.v29i4.13636. ISSN 1396-0466.
  132. ^ Kelly, Kevin (25 April 2017). "The Myth of a Superhuman AI". Wired. Archived from the original on 26 December 2021. Retrieved 19 February 2022.
  133. ^ Jindal, Siddharth (7 July 2023). "OpenAI's Pursuit of AI Alignment is Farfetched". Analytics India Magazine. Retrieved 23 July 2023.
  134. ^ "Mark Zuckerberg responds to Elon Musk's paranoia about AI: 'AI is going to... help keep our communities safe.'". Business Insider. 25 May 2018. Archived from the original on 6 May 2019. Retrieved 6 May 2019.
  135. ^ Dadich, Scott. "Barack Obama Talks AI, Robo Cars, and the Future of the World". WIRED. Archived from the original on 3 December 2017. Retrieved 27 November 2017.
  136. ^ Kircher, Madison Malone. "Obama on the Risks of AI: 'You Just Gotta Have Somebody Close to the Power Cord'". Select All. Archived from the original on 1 December 2017. Retrieved 27 November 2017.
  137. ^ Clinton, Hillary (2017). What Happened. Simon and Schuster. p. 241. ISBN 978-1-5011-7556-5. via [1] Archived 1 December 2017 at the Wayback Machine
  138. ^ "Elon Musk says AI could doom human civilization. Zuckerberg disagrees. Who's right?". 5 January 2023. Archived from the original on 8 January 2018. Retrieved 8 January 2018.
  139. ^ "AI doomsday worries many Americans. So does apocalypse from climate change, nukes, war, and more". 14 April 2023. Archived from the original on 23 June 2023. Retrieved 9 July 2023.
  140. ^ Tyson, Alec; Kikuchi, Emma (28 August 2023). "Growing public concern about the role of artificial intelligence in daily life". Pew Research Center. Retrieved 17 September 2023.
  141. ^ Sotala, Kaj; Yampolskiy, Roman (19 December 2014). "Responses to catastrophic AGI risk: a survey". Physica Scripta. 90 (1).
  142. ^ Barrett, Anthony M.; Baum, Seth D. (23 May 2016). "A model of pathways to artificial superintelligence catastrophe for risk and decision analysis". Journal of Experimental & Theoretical Artificial Intelligence. 29 (2): 397–414. arXiv:1607.07730. doi:10.1080/0952813x.2016.1186228. ISSN 0952-813X. S2CID 928824.
  143. ^ Sotala, Kaj; Yampolskiy, Roman V (19 December 2014). "Responses to catastrophic AGI risk: a survey". Physica Scripta. 90 (1): 018001. Bibcode:2015PhyS...90a8001S. doi:10.1088/0031-8949/90/1/018001. ISSN 0031-8949. S2CID 4749656.
  144. ^ Ramamoorthy, Anand; Yampolskiy, Roman (2018). "Beyond MAD? The race for artificial general intelligence". ICT Discoveries. 1 (Special Issue 1). ITU: 1–8. Archived from the original on 7 January 2022. Retrieved 7 January 2022.
  145. ^ Carayannis, Elias G.; Draper, John (11 January 2022). "Optimising peace through a Universal Global Peace Treaty to constrain the risk of war from a militarised artificial superintelligence". AI & Society. 38 (6): 2679–2692. doi:10.1007/s00146-021-01382-y. ISSN 0951-5666. PMC 8748529. PMID 35035113. S2CID 245877737.
  146. ^ Carayannis, Elias G.; Draper, John (30 May 2023), "The challenge of advanced cyberwar and the place of cyberpeace", The Elgar Companion to Digital Transformation, Artificial Intelligence and Innovation in the Economy, Society and Democracy, Edward Elgar Publishing, pp. 32–80, doi:10.4337/9781839109362.00008, ISBN 978-1-83910-936-2, retrieved 8 June 2023.
  147. ^ Vincent, James (22 June 2016). "Google's AI researchers say these are the five key problems for robot safety". The Verge. Archived from the original on 24 December 2019. Retrieved 5 April 2020.
  148. ^ Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. "Concrete problems in AI safety." arXiv preprint arXiv:1606.06565 (2016).
  149. ^ Johnson, Alex (2019). "Elon Musk wants to hook your brain up directly to computers – starting next year". NBC News. Archived from the original on 18 April 2020. Retrieved 5 April 2020.
  150. ^ Torres, Phil (18 September 2018). "Only Radically Enhancing Humanity Can Save Us All". Slate Magazine. Archived from the original on 6 August 2020. Retrieved 5 April 2020.
  151. ^ Barrett, Anthony M.; Baum, Seth D. (23 May 2016). "A model of pathways to artificial superintelligence catastrophe for risk and decision analysis". Journal of Experimental & Theoretical Artificial Intelligence. 29 (2): 397–414. arXiv:1607.07730. doi:10.1080/0952813X.2016.1186228. S2CID 928824.
  152. ^ Piper, Kelsey (29 March 2023). "How to test what an AI model can – and shouldn't – do". Vox. Retrieved 28 July 2023.
  153. ^ Piesing, Mark (17 May 2012). "AI uprising: humans will be outsourced, not obliterated". Wired. Archived from the original on 7 April 2014. Retrieved 12 December 2015.
  154. ^ Coughlan, Sean (24 April 2013). "How are humans going to become extinct?". BBC News. Archived from the original on 9 March 2014. Retrieved 29 March 2014.
  155. ^ Bridge, Mark (10 June 2017). "Making robots less confident could prevent them taking over". The Times. Archived from the original on 21 March 2018. Retrieved 21 March 2018.
  156. ^ McGinnis, John (Summer 2010). "Accelerating AI". Northwestern University Law Review. 104 (3): 1253–1270. Archived from the original on 15 February 2016. Retrieved 16 July 2014. For all these reasons, verifying a global relinquishment treaty, or even one limited to AI-related weapons development, is a nonstarter... (For different reasons from ours, the Machine Intelligence Research Institute) considers (AGI) relinquishment infeasible...
  157. ^ Sotala, Kaj; Yampolskiy, Roman (19 December 2014). "Responses to catastrophic AGI risk: a survey". Physica Scripta. 90 (1). In general, most writers reject proposals for broad relinquishment... Relinquishment proposals suffer from many of the same problems as regulation proposals, but to a greater extent. There is no historical precedent of general, multi-use technology similar to AGI being successfully relinquished for good, nor do there seem to be any theoretical reasons for believing that relinquishment proposals would work in the future. Therefore we do not consider them to be a viable class of proposals.
  158. ^ Allenby, Brad (11 April 2016). "The Wrong Cognitive Measuring Stick". Slate. Archived from the original on 15 May 2016. Retrieved 15 May 2016. It is fantasy to suggest that the accelerating development and deployment of technologies that taken together are considered to be A.I. will be stopped or limited, either by regulation or even by national legislation.
  159. ^ a b Yampolskiy, Roman V. (2022). "AI Risk Skepticism". In Müller, Vincent C. (ed.). Philosophy and Theory of Artificial Intelligence 2021. Studies in Applied Philosophy, Epistemology and Rational Ethics. Vol. 63. Cham: Springer International Publishing. pp. 225–248. doi:10.1007/978-3-031-09153-7_18. ISBN 978-3-031-09153-7.
  160. ^ Baum, Seth (22 August 2018). "Superintelligence Skepticism as a Political Tool". Information. 9 (9): 209. doi:10.3390/info9090209. ISSN 2078-2489.
  161. ^ "Elon Musk and other tech leaders call for pause in 'out of control' AI race". CNN. 29 March 2023. Retrieved 30 March 2023.
  162. ^ "Open letter calling for AI 'pause' shines light on fierce debate around risks vs. hype". VentureBeat. 29 March 2023. Retrieved 20 July 2023.
  163. ^ Vincent, James (14 April 2023). "OpenAI's CEO confirms the company isn't training GPT-5 and "won't for some time"". The Verge. Retrieved 20 July 2023.
  164. ^ "The Open Letter on AI Doesn't Go Far Enough". Time. 29 March 2023. Retrieved 20 July 2023.
  165. ^ Domonoske, Camila (17 July 2017). "Elon Musk Warns Governors: Artificial Intelligence Poses 'Existential Risk'". NPR. Archived from the original on 23 April 2020. Retrieved 27 November 2017.
  166. ^ Gibbs, Samuel (17 July 2017). "Elon Musk: regulate AI to combat 'existential threat' before it's too late". The Guardian. Archived from the original on 6 June 2020. Retrieved 27 November 2017.
  167. ^ Kharpal, Arjun (7 November 2017). "A.I. is in its 'infancy' and it's too early to regulate it, Intel CEO Brian Krzanich says". CNBC. Archived from the original on 22 March 2020. Retrieved 27 November 2017.
  168. ^ Dawes, James (20 December 2021). "UN fails to agree on 'killer robot' ban as nations pour billions into autonomous weapons research". The Conversation. Retrieved 28 July 2023.
  169. ^ a b Fassihi, Farnaz (18 July 2023). "U.N. Officials Urge Regulation of Artificial Intelligence". The New York Times. ISSN 0362-4331. Retrieved 20 July 2023.
  170. ^ "International Community Must Urgently Confront New Reality of Generative, Artificial Intelligence, Speakers Stress as Security Council Debates Risks, Rewards". United Nations. Retrieved 20 July 2023.
  171. ^ Sotala, Kaj; Yampolskiy, Roman V. (19 December 2014). "Responses to catastrophic AGI risk: a survey". Physica Scripta. 90 (1): 018001. Bibcode:2015PhyS...90a8001S. doi:10.1088/0031-8949/90/1/018001. ISSN 0031-8949.
  172. ^ Geist, Edward Moore (15 August 2016). "It's already too late to stop the AI arms race—We must manage it instead". Bulletin of the Atomic Scientists. 72 (5): 318–321. Bibcode:2016BuAtS..72e.318G. doi:10.1080/00963402.2016.1216672. ISSN 0096-3402. S2CID 151967826.
  173. ^ "Amazon, Google, Meta, Microsoft and other tech firms agree to AI safeguards set by the White House". AP News. 21 July 2023. Retrieved 21 July 2023.
  174. ^ "Amazon, Google, Meta, Microsoft and other firms agree to AI safeguards". Redditch Advertiser. 21 July 2023. Retrieved 21 July 2023.
  175. ^ The White House (30 October 2023). "Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence". The White House. Retrieved 19 December 2023.

Bibliography