Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Friendly artificial intelligence
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
== Other approaches == {{See also|AI control problem#Alignment|AI safety}} [[Steve Omohundro]] has proposed a "scaffolding" approach to [[AI safety]], in which one provably safe AI generation helps build the next provably safe generation.<ref name=Hendry2014>{{cite news|last1=Hendry|first1=Erica R.|title=What Happens When Artificial Intelligence Turns On Us?|url=http://www.smithsonianmag.com/innovation/what-happens-when-artificial-intelligence-turns-us-180949415/|access-date=15 July 2014|work=Smithsonian Magazine|date=21 Jan 2014|archive-date=19 July 2014|archive-url=https://web.archive.org/web/20140719142131/http://www.smithsonianmag.com/innovation/what-happens-when-artificial-intelligence-turns-us-180949415/|url-status=live}}</ref> [[Seth Baum]] argues that the development of safe, socially beneficial artificial intelligence or artificial general intelligence is a function of the social psychology of AI research communities and so can be constrained by extrinsic measures and motivated by intrinsic measures. Intrinsic motivations can be strengthened when messages resonate with AI developers; Baum argues that, in contrast, "existing messages about beneficial AI are not always framed well". Baum advocates for "cooperative relationships, and positive framing of AI researchers" and cautions against characterizing AI researchers as "not want(ing) to pursue beneficial designs".<ref>{{Cite journal|last=Baum|first=Seth D.|date=2016-09-28|title=On the promotion of safe and socially beneficial artificial intelligence|journal=AI & Society|volume=32|issue=4|pages=543β551|doi=10.1007/s00146-016-0677-0|s2cid=29012168|issn=0951-5666}}</ref> In his book ''[[Human Compatible]]'', AI researcher [[Stuart J. Russell]] lists three principles to guide the development of beneficial machines. He emphasizes that these principles are not meant to be explicitly coded into the machines; rather, they are intended for the human developers. The principles are as follows:<ref name="HC">{{cite book |last=Russell |first=Stuart |date=October 8, 2019 |title=Human Compatible: Artificial Intelligence and the Problem of Control |url=https://archive.org/details/humancompatiblea0000russ |location=United States |publisher=Viking |isbn=978-0-525-55861-3 |author-link=Stuart J. Russell |oclc=1083694322 |url-access=registration }}</ref>{{rp|173}} {{quote| # The machine's only objective is to maximize the realization of human preferences. # The machine is initially uncertain about what those preferences are. # The ultimate source of information about human preferences is human behavior.}} The "preferences" Russell refers to "are all-encompassing; they cover everything you might care about, arbitrarily far into the future."<ref name="HC"/>{{rp|173}} Similarly, "behavior" includes any choice between options,<ref name="HC"/>{{rp|177}} and the uncertainty is such that some probability, which may be quite small, must be assigned to every logically possible human preference.<ref name="HC"/>{{rp|201}}
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)