HOME

TheInfoList



OR:

''Human Compatible: Artificial Intelligence and the Problem of Control'' is a 2019 non-fiction book by computer scientist Stuart J. Russell. It asserts that risk to humanity from advanced
artificial intelligence Artificial intelligence (AI) is intelligence—perceiving, synthesizing, and inferring information—demonstrated by machine A machine is a physical system using Power (physics), power to apply Force, forces and control Motion, moveme ...
(AI) is a serious concern despite the uncertainty surrounding future progress in AI. It also proposes an approach to the AI control problem.


Summary

Russell begins by asserting that the standard model of AI research, in which the primary definition of success is getting better and better at achieving rigid human-specified goals, is dangerously misguided. Such goals may not actually reflect what human designers intend, such as by failing to take into account any human values not included in the goals. If an AI developed according to the standard model were to become
superintelligent A superintelligence is a hypothetical agent that possesses intelligence far surpassing that of the brightest and most gifted human minds. "Superintelligence" may also refer to a property of problem-solving systems (e.g., superintelligent language ...
, it would likely not fully reflect human values and could be catastrophic to humanity. Russell asserts that precisely because the timeline for developing human-level or superintelligent AI is highly uncertain, safety research should be begun as soon as possible, as it is also highly uncertain how long it would take to complete such research. Russell argues that continuing progress in AI capability is inevitable because of economic pressures. Such pressures can already be seen in the development of existing AI technologies such as
self-driving car A self-driving car, also known as an autonomous car, driver-less car, or robotic car (robo-car), is a car that is capable of traveling without human input.Xie, S.; Hu, J.; Bhowmick, P.; Ding, Z.; Arvin, F.,Distributed Motion Planning for S ...
s and personal assistant software. Moreover, human-level AI could be worth many trillions of dollars. Russell then examines the current debate surrounding AI risk. He offers refutations to a number of common arguments dismissing AI risk and attributes much of their persistence to tribalism—AI researchers may see AI risk concerns as an "attack" on their field. However, Russell reiterates that there are legitimate reasons to take AI risk concerns seriously and that economic pressures make continued innovation in AI inevitable. Russell then proposes an approach to developing provably beneficial machines that focuses on deference to humans. Unlike in the standard model of AI, where the objective is rigid and certain, this approach would have the AI's true objective remain uncertain, with the AI only approaching certainty about it as it gains more information about humans and the world. This uncertainty would, ideally, prevent catastrophic misunderstandings of human preferences and encourage cooperation and communication with humans. Russell concludes by calling for tighter governance of AI research and development as well as cultural introspection about the appropriate amount of autonomy to retain in an AI-dominated world.


Russell's three principles

Russell lists three principles to guide the development of beneficial machines. He emphasizes that these principles are not meant to be explicitly coded into the machines; rather, they are intended for the human developers. The principles are as follows: The "preferences" Russell refers to "are all-encompassing; they cover everything you might care about, arbitrarily far into the future." Similarly, "behavior" includes any choice between options, and the uncertainty is such that some probability, which may be quite small, must be assigned to every logically possible human preference. Russell explores inverse reinforcement learning, in which a machine infers a reward function from observed behavior, as a possible basis for a mechanism for learning human preferences.


Reception

Several reviewers agreed with the book's arguments. Ian Sample in ''
The Guardian ''The Guardian'' is a British daily newspaper A newspaper is a periodical publication containing written information about current events and is often typed in black ink with a white or gray background. Newspapers can cover a wide ...
'' called it "convincing" and "the most important book on AI this year". Richard Waters of the ''
Financial Times The ''Financial Times'' (''FT'') is a British daily newspaper printed in broadsheet and published digitally that focuses on business and economic current affairs. Based in London, England, the paper is owned by a Japanese holding company, Nikke ...
'' praised the book's "bracing intellectual rigour". '' Kirkus Reviews'' endorsed it as "a strong case for planning for the day when machines can outsmart us". The same reviewers characterized the book as "wry and witty", or "accessible" due to its "laconic style and dry humour". Matthew Hutson of the ''
Wall Street Journal ''The Wall Street Journal'' is an American business-focused, international daily newspaper based in New York City, with international editions also available in Chinese and Japanese. The ''Journal'', along with its Asian editions, is published ...
'' said "Mr. Russell's exciting book goes deep, while sparkling with dry witticisms". A '' Library Journal'' reviewer called it "The right guide at the right time". James McConnachie of ''
The Times ''The Times'' is a British daily national newspaper based in London. It began in 1785 under the title ''The Daily Universal Register'', adopting its current name on 1 January 1788. ''The Times'' and its sister paper '' The Sunday Times'' ...
'' wrote "This is not quite the popular book that AI urgently needs. Its technical parts are too difficult, its philosophical ones too easy. But it is fascinating, and significant." By contrast, ''Human Compatible'' was criticized in its ''
Nature Nature, in the broadest sense, is the physical world or universe. "Nature" can refer to the phenomena of the physical world, and also to life in general. The study of nature is a large, if not the only, part of science. Although humans ar ...
'' review by David Leslie, an Ethics Fellow at the
Alan Turing Institute The Alan Turing Institute is the United Kingdom's national institute for data science and artificial intelligence, founded in 2015 and largely funded by the UK government. It is named after Alan Turing, the British mathematician and computing p ...
; and similarly in a ''
New York Times ''The New York Times'' (''the Times'', ''NYT'', or the Gray Lady) is a daily newspaper based in New York City with a worldwide readership reported in 2020 to comprise a declining 840,000 paid print subscribers, and a growing 6 million paid ...
'' opinion essay by Melanie Mitchell. One point of contention was whether
superintelligence A superintelligence is a hypothetical agent that possesses intelligence far surpassing that of the brightest and most gifted human minds. "Superintelligence" may also refer to a property of problem-solving systems (e.g., superintelligent language ...
is possible. Leslie states Russell "fails to convince that we will ever see the arrival of a 'second intelligent species'", and Mitchell doubts a machine could ever "surpass the generality and flexibility of human intelligence" without losing "the speed, precision and programmability of a computer". A second disagreement was whether intelligent machines would naturally tend to adopt so-called "common sense" moral values. On Russell's thought experiment about a geoengineering robot that "asphyxiates humanity to deacidify the oceans", Leslie "struggles to identify any intelligence". Similarly, Mitchell believes an intelligent robot would naturally tend to be "tempered by the common sense, values and social judgment without which general intelligence cannot exist". The book was longlisted for the 2019 ''Financial Times''/McKinsey award.


See also

* '' Artificial Intelligence: A Modern Approach'' *
Center for Human-Compatible Artificial Intelligence The Center for Human-Compatible Artificial Intelligence (CHAI) is a research center at University of California, Berkeley (UC Berkeley) focusing on advanced artificial intelligence (AI) safety methods. CHAI was founded in 2016 by a group of aca ...
* '' The Precipice: Existential Risk and the Future of Humanity'' * '' Slaughterbots'' * '' Superintelligence: Paths, Dangers, Strategies''


References


External links


Interview with Stuart J. Russell
{{Existential risk from artificial intelligence, state=expanded American non-fiction books 2019 non-fiction books English-language books Existential risk from artificial general intelligence Futurology books Technology books Viking Press books