Friendly artificial intelligence: Difference between revisions

Content deleted Content added
m removed Category:Futures studies using HotCat redundant
Citation bot (talk | contribs)
Add: s2cid. | Use this bot. Report bugs. | Suggested by Neko-chan | Category:Philosophy of artificial intelligence | #UCB_Category 35/47
Line 27:
Luke Muehlhauser, writing for the [[Machine Intelligence Research Institute]], recommends that [[machine ethics]] researchers adopt what [[Bruce Schneier]] has called the "security mindset": Rather than thinking about how a system will work, imagine how it could fail. For instance, he suggests even an AI that only makes accurate predictions and communicates via a text interface might cause unintended harm.<ref name=MuehlhauserSecurity2013>{{cite web|last1=Muehlhauser|first1=Luke|title=AI Risk and the Security Mindset|url=http://intelligence.org/2013/07/31/ai-risk-and-the-security-mindset/|website=Machine Intelligence Research Institute|access-date=15 July 2014|date=31 Jul 2013}}</ref>
 
In 2014, Luke Muehlhauser and Nick Bostrom underlined the need for 'friendly AI';<ref>{{Cite journal|last1=Muehlhauser|first1=Luke|last2=Bostrom|first2=Nick|title=Why We Need Friendly AI|date=2013-12-17|journal=Think|volume=13|issue=36|pages=41–47|doi=10.1017/s1477175613000316|s2cid=143657841|issn=1477-1756}}</ref> nonetheless, the difficulties in designing a 'friendly' superintelligence, for instance via programming counterfactual moral thinking, are considerable.<ref>{{Cite journal|last1=Boyles|first1=Robert James M.|last2=Joaquin|first2=Jeremiah Joven|date=2019-07-23|title=Why friendly AIs won't be that friendly: a friendly reply to Muehlhauser and Bostrom|journal=AI & Society|volume=35|issue=2|pages=505–507|doi=10.1007/s00146-019-00903-0|s2cid=198190745|issn=0951-5666}}</ref><ref>{{Cite journal|last=Chan|first=Berman|date=2020-03-04|title=The rise of artificial intelligence and the crisis of moral passivity|journal=AI & Society|volume=35|issue=4|pages=991–993|language=en|doi=10.1007/s00146-020-00953-9|s2cid=212407078|issn=1435-5655}}</ref>
 
==Coherent extrapolated volition==
Line 60:
{{see also|Technological singularity#Criticisms}}
 
Some critics believe that both human-level AI and superintelligence are unlikely, and that therefore friendly AI is unlikely. Writing in ''[[The Guardian]]'', Alan Winfield compares human-level artificial intelligence with faster-than-light travel in terms of difficulty, and states that while we need to be "cautious and prepared" given the stakes involved, we "don't need to be obsessing" about the risks of superintelligence.<ref>{{cite news|last1=Winfield|first1=Alan|title=Artificial intelligence will not turn into a Frankenstein's monster|url=https://www.theguardian.com/technology/2014/aug/10/artificial-intelligence-will-not-become-a-frankensteins-monster-ian-winfield|access-date=17 September 2014|work=[[The Guardian]]}}</ref> Boyles and Joaquin, on the other hand, argue that Luke Muehlhauser and [[Nick Bostrom]]’s proposal to create friendly AIs appear to be bleak. This is because Muehlhauser and Bostrom seem to hold the idea that intelligent machines could be programmed to think counterfactually about the moral values that humans beings would have had.<ref>{{cite journal | last1 = Muehlhauser | first1 = Luke | last2 = Bostrom | first2 = Nick | year = 2014 | title = Why we need friendly AI | journal = Think | volume = 13 | issue = 36 | pages = 41–47 | doi = 10.1017/S1477175613000316| s2cid = 143657841 }}</ref> In an article in ''[[AI & Society]]'', Boyles and Joaquin maintain that such AIs would not be that friendly considering the following: the infinite amount of antecedent counterfactual conditions that would have to be programmed into a machine, the difficulty of cashing out the set of moral values—that is, those that a more ideal than the ones human beings possess at present, and the apparent disconnect between counterfactual antecedents and ideal value consequent.<ref>{{cite journal | last1 = Boyles | first1 = Robert James M. | last2 = Joaquin | first2 = Jeremiah Joven | year = 2019 | title = Why Friendly AIs won't be that Friendly: A Friendly Reply to Muehlhauser and Bostrom | journal = AI & Society | volume = 35 | issue = 2 | pages = 505–507 | doi = 10.1007/s00146-019-00903-0| s2cid = 198190745 }}</ref>
 
Some philosophers claim that any truly "rational" agent, whether artificial or human, will naturally be benevolent; in this view, deliberate safeguards designed to produce a friendly AI could be unnecessary or even harmful.<ref>Kornai, András. "[http://www.kornai.com/Papers/agi12.pdf Bounding the impact of AGI]". Journal of Experimental & Theoretical Artificial Intelligence ahead-of-print (2014): 1-22. "...the essence of AGIs is their reasoning facilities, and it is the very logic of their being that will compel them to behave in a moral fashion... The real nightmare scenario (is one where) humans find it advantageous to strongly couple themselves to AGIs, with no guarantees against self-deception."</ref> Other critics question whether it is possible for an artificial intelligence to be friendly. Adam Keiper and Ari N. Schulman, editors of the technology journal ''[[The New Atlantis (journal)|The New Atlantis]]'', say that it will be impossible to ever guarantee "friendly" behavior in AIs because problems of ethical complexity will not yield to software advances or increases in computing power. They write that the criteria upon which friendly AI theories are based work "only when one has not only great powers of prediction about the likelihood of myriad possible outcomes, but certainty and consensus on how one values the different outcomes.<ref>{{cite web|url=http://www.thenewatlantis.com/publications/the-problem-with-friendly-artificial-intelligence|author=Adam Keiper and Ari N. Schulman|title=The Problem with 'Friendly' Artificial Intelligence|publisher=The New Atlantis|access-date = 2012-01-16}}</ref>