Human Compatible: Artificial Intelligence and the Problem of Control
Human Compatible: Artificial Intelligence and the Problem of Control book cover

Human Compatible: Artificial Intelligence and the Problem of Control

Hardcover – October 8, 2019

Price
$13.35
Format
Hardcover
Pages
352
Publisher
Viking
Publication Date
ISBN-13
978-0525558613
Dimensions
6.19 x 1.15 x 9.26 inches
Weight
1.2 pounds

Description

Praise for Human Compatible : “This is the most important book I have read in quite some time.xa0 It lucidly explains how the coming age of artificial super-intelligence threatens human control. Crucially, it also introduces a novel solution and a reason for hope.” —Daniel Kahneman, winner of the Nobel Prize and author of Thinking, Fast and Slow “A must-read: this intellectual tour-de-force by one of AI's true pioneers not only explains the risks of ever more powerful artificial intelligence in axa0captivating and persuasivexa0way, but also proposes a concrete and promising solution.”xa0 — Max Tegmark, author of Life 3.0 “A thought-provoking and highly readable account of the past, present and future of AI . . . Russell is grounded in the realities of the technology, including its many limitations, and isn’t one to jump at the overheated language of sci-fi . . .xa0If you are looking for a serious overview to the subject that doesn’t talk down to its non-technical readers, this is a good place to start . . .xa0[Russell] deploys a bracing intellectual rigour . . . But a laconic style and dry humour keep his book accessible to the lay reader.” — Financial Times “ A carefully written explanation of the concepts underlying AI as well as the history of their development. If you want to understand how fast AI is developing and why the technology is so dangerous, Human Compatible is your guide.” — TechCrunch “Sound[s] an important alarm bell . . . Human Compatible marks a major stride in AI studies, not least in its emphasis on ethics. At the book’s heart, Russell incisively discusses the misuses of AI.” — Nature “An AI expert’s chilling warning . . . Fascinating, and significant . . . Russell is not warning of the dangers of conscious machines, just that superintelligent ones might be misused or might misuse themselves.” — The Times (UK) “An excellent, nuanced history of the field.” — The Telegraph (UK) “A brillantly clear and fascinating exposition of the history of computing thus far, and how very difficult true AI will be to build.” — The Spectator (UK) “ Human Compatible made me a convert to Russell's concerns with our ability to control our upcoming creation—super-intelligent machines. Unlike outside alarmists and futurists,xa0Russell is a leading authority on AI. His new bookxa0will educate the public about AI more than any book I can think of, and is a delightful and uplifting read.” — Judea Pearl, Turing Award-winner and author of The Book of Why “Stuart Russell has long been the most sensible voice in computer science on the topic of AI risk. And he has now writtenxa0the book we've all been waiting for -- axa0brilliant and utterly accessible guidexa0to what will be either the best or worst technological development in human history.” — Sam Harris, author of Waking Up and host of the Making Sense podcast “This beautifully written book addresses a fundamental challenge for humanity: increasingly intelligent machines that do what we ask but not what we really intend.xa0 Essential reading if you care about our future.” —Yoshua Bengio, winner of the 2019 Turing Award and co-author of Deep Learning “Authoritative [and] accessible . . . A strong case for planning for the day when machines can outsmart us.” — Kirkus Reviews “The right guide at the right time for technology enthusiasts seeking to explore the primary concepts of what makes AI valuable while simultaneously examining the disconcerting aspects of AI misuse.” — Library Journal “The same mix ofxa0de-mystifying authority and practical advicexa0that Dr. Benjamin Spock once brought to the care and raising of children, Dr. Stuart Russell now brings to the care, raising, and yes, disciplining of machines.xa0He has written the book that most—but perhaps not all—machines would like you to read.” — George Dyson, author of Turing's Cathedral “Persuasively argued and lucidly imagined, Human Compatible offers anxa0unflinching, incisivexa0look at what awaits us in the decades ahead. No researcher has argued more persuasively about the risks of AI or shown more clearly the way forward.xa0Anyone who takes the future seriously should pay attention.” — Brian Christian, author of Algorithms to Live By “A book that charts humanity's quest to understand intelligence, pinpoints why it became unsafe, and shows how to course-correct if we want to survive as a species. Stuart Russell, author of the leading AI textbook, can do all that with the wealth of knowledge of a prominent AI researcher and the persuasive clarity and wit of a brilliant educator.” —Jann Tallinn, co-founder of Skype “Can we coexist happily with the intelligent machines that humans will create? ‘Yes,’ answers Human Compatible , ‘but first . . .’ Through a brilliant reimagining of the foundations of artificial intelligence, Russell takes you on a journey from the very beginning, explaining the questions raised by an AI-driven society and beautifully making the case for how to ensure machines remain beneficial to humans. A totally readable and crucially important guide to the future from one of the world's leading experts.” —Tabitha Goldstaub, co-founder of CognitionX and Head of the UK Government's AI Council “Stuart Russell, one of the most important AI scientists of the last 25 years, may have written the most important book about AI so far, on one of the most important questions of the 21st century: How to build AI to be compatible with us. The book proposes a novel and intriguing solution for this problem, while offering many thought-provoking ideas and insights about AI along the way. An accessible and engaging must-read for the developers of AI and the users of AI—that is, forxa0all of us.” —James Manyika, chairman and director of McKinsey Global Institute “In clear and compelling language, Stuart Russell describes the huge potential benefits of artificial Intelligence, as well as the hazards and ethical challenges. It's especially welcome that a respected leading authority should offer this balanced appraisal, avoiding both hype and scaremongering.” —Lord Martin Rees, Astronomer Royal and former President of the Royal Society Stuart Russell is a professor of Computer Science and holder of the Smith-Zadeh Chair in Engineering at the University of California, Berkeley. He has served as the Vice-Chair of the World Economic Forum's Council on AI and Robotics and as an advisor to the United Nations on arms control. He is a Fellow of the American Association for Artificial Intelligence, the Association for Computing Machinery, and the American Association for the Advancement of Science. He is the author (with Peter Norvig) of the definitive and universally acclaimed textbook on AI, Artificial Intelligence: A Modern Approach . Excerpt. © Reprinted by permission. All rights reserved. 1 If We Succeed A long time ago, my parents lived in Birmingham, England, in a house near the university. They decided to move out of the city and sold the house to David Lodge, a professor of English literature. Lodge was by that time already a well-known novelist. I never met him, but I decided to read some of his books: Changing Places and Small World. Among the principal characters were fictional academics moving from a fictional version of Birmingham to a fictional version of Berkeley, California. As I was an actual academic from the actual Birmingham who had just moved to the actual Berkeley, it seemed that someone in the Department of Coincidences was telling me to pay attention. One particular scene from Small World struck me: The protagonist, an aspiring literary theorist, attends a major international conference and asks a panel of leading figures, "What follows if everyone agrees with you?" The question causes consternation, because the panelists had been more concerned with intellectual combat than ascertaining truth or attaining understanding. It occurred to me then that an analogous question could be asked of the leading figures in AI: "What if you succeed?" The field's goal had always been to create human-level or superhuman AI, but there was little or no consideration of what would happen if we did. A few years later, Peter Norvig and I began work on a new AI textbook, whose first edition appeared in 1995. The book's final section is titled "What If We Do Succeed?" The section points to the possibility of good and bad outcomes but reaches no firm conclusions. By the time of the third edition in 2010, many people had finally begun to consider the possibility that superhuman AI might not be a good thing-but these people were mostly outsiders rather than mainstream AI researchers. By 2013, I became convinced that the issue not only belonged in the mainstream but was possibly the most important question facing humanity. In November 2013, I gave a talk at the Dulwich Picture Gallery, a venerable art museum in south London. The audience consisted mostly of retired people-nonscientists with a general interest in intellectual matters-so I had to give a completely nontechnical talk. It seemed an appropriate venue to try out my ideas in public for the first time. After explaining what AI was about, I nominated five candidates for "biggest event in the future of humanity": 1. We all die (asteroid impact, climate catastrophe, pandemic, etc.). 2. We all live forever (medical solution to aging). 3. We invent faster-than-light travel and conquer the universe. 4. We are visited by a superior alien civilization. 5. We invent superintelligent AI. I suggested that the fifth candidate, superintelligent AI, would be the winner, because it would help us avoid physical catastrophes and achieve eternal life and faster-than-light travel, if those were indeed possible. It would represent a huge leap-a discontinuity-in our civilization. The arrival of superintelligent AI is in many ways analogous to the arrival of a superior alien civilization but much more likely to occur. Perhaps most important, AI, unlike aliens, is something over which we have some say. Then I asked the audience to imagine what would happen if we received notice from a superior alien civilization that they would arrive on Earth in thirty to fifty years. The word pandemonium doesn't begin to describe it. Yet our response to the anticipated arrival of superintelligent AI has been . . . well, underwhelming begins to describe it. (In a later talk, I illustrated this in the form of the email exchange shown in figure 1.) Finally, I explained the significance of superintelligent AI as follows: "Success would be the biggest event in human history . . . and perhaps the last event in human history." From: Superior Alien Civilization To: [email protected] Subject: Contact Be warned: we shall arrive in 30-50 years From: [email protected] To: Superior Alien Civilization Subject: Out of office: Re: Contact Humanity is currently out of the office. We will respond to your message when we return. Figure 1: Probably not the email exchange that would follow the first contact by a superior alien civilization. A few months later, in April 2014, I was at a conference in Iceland and got a call from National Public Radio asking if they could interview me about the movie Transcendence, which had just been released in the United States. Although I had read the plot summaries and reviews, I hadn't seen it because I was living in Paris at the time, and it would not be released there until June. It so happened, however, that I had just added a detour to Boston on the way home from Iceland, so that I could participate in a Defense Department meeting. So, after arriving at Boston's Logan Airport, I took a taxi to the nearest theater showing the movie. I sat in the second row and watched as a Berkeley AI professor, played by Johnny Depp, was gunned down by anti-AI activists worried about, yes, superintelligent AI. Involuntarily, I shrank down in my seat. (Another call from the Department of Coincidences?) Before Johnny Depp's character dies, his mind is uploaded to a quantum supercomputer and quickly outruns human capabilities, threatening to take over the world. On April 19, 2014, a review of Transcendence, co-authored with physicists Max Tegmark, Frank Wilczek, and Stephen Hawking, appeared in the Huffington Post. It included the sentence from my Dulwich talk about the biggest event in human history. From then on, I would be publicly committed to the view that my own field of research posed a potential risk to my own species. How Did We Get Here? The roots of AI stretch far back into antiquity, but its "official" beginning was in 1956. Two young mathematicians, John McCarthy and Marvin Minsky, had persuaded Claude Shannon, already famous as the inventor of information theory, and Nathaniel Rochester, the designer of IBM's first commercial computer, to join them in organizing a summer program at Dartmouth College. The goal was stated as follows: The study is to proceed on the basis of the conjecture that every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it. An attempt will be made to find how to make machines use language, form abstractions and concepts, solve kinds of problems now reserved for humans, and improve themselves. We think that a significant advance can be made in one or more of these problems if a carefully selected group of scientists work on it together for a summer. Needless to say, it took much longer than a summer: we are still working on all these problems. In the first decade or so after the Dartmouth meeting, AI had several major successes, including Alan Robinson's algorithm for general-purpose logical reasoning and Arthur Samuel's checker-playing program, which taught itself to beat its creator. The first AI bubble burst in the late 1960s, when early efforts at machine learning and machine translation failed to live up to expectations. A report commissioned by the UK government in 1973 concluded, "In no part of the field have the discoveries made so far produced the major impact that was then promised." In other words, the machines just weren't smart enough. My eleven-year-old self was, fortunately, unaware of this report. Two years later, when I was given a Sinclair Cambridge Programmable calculator, I just wanted to make it intelligent. With a maximum program size of thirty-six keystrokes, however, the Sinclair was not quite big enough for human-level AI. Undeterred, I gained access to the giant CDC 6600 supercomputer at Imperial College London and wrote a chess program-a stack of punched cards two feet high. It wasn't very good, but it didn't matter. I knew what I wanted to do. By the mid-1980s, I had become a professor at Berkeley, and AI was experiencing a huge revival thanks to the commercial potential of so-called expert systems. The second AI bubble burst when these systems proved to be inadequate for many of the tasks to which they were applied. Again, the machines just weren't smart enough. An AI winter ensued. My own AI course at Berkeley, currently bursting with over nine hundred students, had just twenty-five students in 1990. The AI community learned its lesson: smarter, obviously, was better, but we would have to do our homework to make that happen. The field became far more mathematical. Connections were made to the long-established disciplines of probability, statistics, and control theory. The seeds of today's progress were sown during that AI winter, including early work on large-scale probabilistic reasoning systems and what later became known as deep learning. Beginning around 2011, deep learning techniques began to produce dramatic advances in speech recognition, visual object recognition, and machine translation-three of the most important open problems in the field. By some measures, machines now match or exceed human capabilities in these areas. In 2016 and 2017, DeepMind's AlphaGo defeated Lee Sedol, former world Go champion, and Ke Jie, the current champion-events that some experts predicted wouldn't happen until 2097, if ever. Now AI generates front-page media coverage almost every day. Thousands of start-up companies have been created, fueled by a flood of venture funding. Millions of students have taken online AI and machine learning courses, and experts in the area command salaries in the millions of dollars. Investments flowing from venture funds, national governments, and major corporations are in the tens of billions of dollars annually-more money in the last five years than in the entire previous history of the field. Advances that are already in the pipeline, such as self-driving cars and intelligent personal assistants, are likely to have a substantial impact on the world over the next decade or so. The potential economic and social benefits of AI are vast, creating enormous momentum in the AI research enterprise. What Happens Next? Does this rapid rate of progress mean that we are about to be overtaken by machines? No. There are several breakthroughs that have to happen before we have anything resembling machines with superhuman intelligence. Scientific breakthroughs are notoriously hard to predict. To get a sense of just how hard, we can look back at the history of another field with civilization-ending potential: nuclear physics. In the early years of the twentieth century, perhaps no nuclear physicist was more distinguished than Ernest Rutherford, the discoverer of the proton and the "man who split the atom." Like his colleagues, Rutherford had long been aware that atomic nuclei stored immense amounts of energy; yet the prevailing view was that tapping this source of energy was impossible. On September 11, 1933, the British Association for the Advancement of Science held its annual meeting in Leicester. Lord Rutherford addressed the evening session. As he had done several times before, he poured cold water on the prospects for atomic energy: "Anyone who looks for a source of power in the transformation of the atoms is talking moonshine." Rutherford's speech was reported in the Times of London the next morning. Leo Szilard, a Hungarian physicist who had recently fled from Nazi Germany, was staying at the Imperial Hotel on Russell Square in London. He read the Times' report at breakfast. Mulling over what he had read, he went for a walk and invented the neutron-induced nuclear chain reaction. The problem of liberating nuclear energy went from impossible to essentially solved in less than twenty-four hours. Szilard filed a secret patent for a nuclear reactor the following year. The first patent for a nuclear weapon was issued in France in 1939. The moral of this story is that betting against human ingenuity is foolhardy, particularly when our future is at stake. Within the AI community, a kind of denialism is emerging, even going as far as denying the possibility of success in achieving the long-term goals of AI. It's as if a bus driver, with all of humanity as passengers, said, "Yes, I am driving as hard as I can towards a cliff, but trust me, we'll run out of gas before we get there!" I am not saying that success in AI will necessarily happen, and I think it's quite unlikely that it will happen in the next few years. It seems prudent, nonetheless, to prepare for the eventuality. If all goes well, it would herald a golden age for humanity, but we have to face the fact that we are planning to make entities that are far more powerful than humans. How do we ensure that they never, ever have power over us? To get just an inkling of the fire we're playing with, consider how content-selection algorithms function on social media. They aren't particularly intelligent, but they are in a position to affect the entire world because they directly influence billions of people. Typically, such algorithms are designed to maximize click-through, that is, the probability that the user clicks on presented items. The solution is simply to present items that the user likes to click on, right? Wrong. The solution is to change the user's preferences so that they become more predictable. A more predictable user can be fed items that they are likely to click on, thereby generating more revenue. People with more extreme political views tend to be more predictable in which items they will click on. (Possibly there is a category of articles that die-hard centrists are likely to click on, but it's not easy to imagine what this category consists of.) Like any rational entity, the algorithm learns how to modify the state of its environment-in this case, the user's mind-in order to maximize its own reward. The consequences include the resurgence of fascism, the dissolution of the social contract that underpins democracies around the world, and potentially the end of the European Union and NATO. Not bad for a few lines of code, even if it had a helping hand from some humans. Now imagine what a really intelligent algorithm would be able to do. What Went Wrong? The history of AI has been driven by a single mantra: "The more intelligent the better." I am convinced that this is a mistake-not because of some vague fear of being superseded but because of the way we have understood intelligence itself. The concept of intelligence is central to who we are-that's why we call ourselves Homo sapiens, or "wise man." After more than two thousand years of self-examination, we have arrived at a characterization of intelligence that can be boiled down to this: Humans are intelligent to the extent that our actions can be expected to achieve our objectives. All those other characteristics of intelligence-perceiving, thinking, learning, inventing, and so on-can be understood through their contributions to our ability to act successfully. From the very beginnings of AI, intelligence in machines has been defined in the same way: Read more

Features & Highlights

  • "The most important book on AI this year."
  • --The Guardian
  • "Mr. Russell's exciting book goes deep, while sparkling with dry witticisms." --
  • The Wall Street Journal
  • "The most important book I have read in quite some time" (Daniel Kahneman); "A must-read" (Max Tegmark); "The book we've all been waiting for" (Sam Harris)
  • A leading artificial intelligence researcher lays out a new approach to AI that will enable us to coexist successfully with increasingly intelligent machines
  • In the popular imagination, superhuman artificial intelligence is an approaching tidal wave that threatens not just jobs and human relationships, but civilization itself. Conflict between humans and machines is seen as inevitable and its outcome all too predictable.In this groundbreaking book, distinguished AI researcher Stuart Russell argues that this scenario can be avoided, but only if we rethink AI from the ground up. Russell begins by exploring the idea of intelligence in humans and in machines. He describes the near-term benefits we can expect, from intelligent personal assistants to vastly accelerated scientific research, and outlines the AI breakthroughs that still have to happen before we reach superhuman AI. He also spells out the ways humans are already finding to misuse AI, from lethal autonomous weapons to viral sabotage.If the predicted breakthroughs occur and superhuman AI emerges, we will have created entities far more powerful than ourselves. How can we ensure they never, ever, have power over us? Russell suggests that we can rebuild AI on a new foundation, according to which machines are designed to be inherently uncertain about the human preferences they are required to satisfy. Such machines would be humble, altruistic, and committed to pursue our objectives, not theirs. This new foundation would allow us to create machines that are provably deferential and provably beneficial.

Customer Reviews

Rating Breakdown

★★★★★
60%
(375)
★★★★
25%
(156)
★★★
15%
(94)
★★
7%
(44)
-7%
(-44)

Most Helpful Reviews

✓ Verified Purchase

The AI problem and how to solve it – by a world-leading AI researcher

A must-read: this intellectual tour-de-force by one of AI’s true pioneers not only explains the risks of ever more powerful artificial intelligence in a captivating and persuasive way, but also proposes a concrete and promising solution.

Will AI eventually supersede human intelligence at all tasks and, if so, will this be the best thing ever to happen to humanity, or the worst? There have been many thought-provoking books on this topic, including Nick Bostrom's "Superintelligence", but this is the first one written by a world-leading AI researcher. Stuart Russell is the first author of the standard textbook on the subject, "Artificial Intelligence: A Modern Approach". I can personally certify that he remains
at the top of his game, since I've had many opportunities to read his recent peer-reviewed technical AI papers as well as to personally experience his depth and expertise during numerous AI conversations and AI conferences over the years.
The book is loosely speaking organized into two parts: the problem and the solution, and I'll attempt to summarize them below.

1) THE PROBLEM: Russell argues that intelligence isn't something mysterious that can only exist in biological organisms, but instead involves information processing that can in principle be performed even better by future machines. He also argues that this is likely to happen, because curiosity and profit will continue to inexorably drive today's rapid pace of AI development until it eventually reaches the level of Artificial General Intelligence (AGI), defined as AI that can perform all intellectual tasks at least as well as humans. AGI could be great for humanity if used to amplify human intelligence to wisely solve pressing problems that stump us, and to create a world free from disease, poverty and misery, but things could also go terribly wrong. Russell argues that the real risk with AGI isn't malice, like in silly Hollywood movies, but competence: machines that succeed in accomplishing goals that aren't aligned with ours. When the autopilot of a German passenger jet flew into the Alps killing 150 people, the computer didn't do so because it was evil, but because the goal it had been given (to lower its altitude to 100 meters) was misaligned with the goals of the passengers, and nobody had thought of teaching it the goal to never fly into mountains. Russell argues that we can already get eerie premonitions of what's it's like to be up against misaligned intelligent entities from case studies of certain large corporations having goals that don't align with humanity's best interests.

The historical account Russell gives of these ideas provides a fascinating perspective, especially since
he personally knew most of the key players. He describes how early AI pioneers such as Alan Turing, John von Neumann, and Norbert Wiener were acutely aware of the value-alignment problem, and how subsequent generations of AI researchers tended to forget them once short-term applications and business opportunities appeared. Upton Sinclair once quipped "It is difficult to get a man to understand something when his salary depends upon his not understanding it", so it's hardly surprising that today's AI experts in industry are less likely to voice concerns than academics such as Turing, von Neuman, Wiener and Russell. Yet Stuart argues that we need to sound the alarm: if AI research succeeds in its original goal of building AGI, then whoever or whatever controls it may be able to take control of Earth much as Homo Sapiens seized control from other less intelligent mammals, so we better ensure that humanity fares better than the Neanderthals did.

2) THE SOLUTION: What I find truly unique about this book, besides Russell's insider perspective, is that he doesn't merely explain the problem, but also proposes a concrete and promising solution. And not a merely vague slogans such as "let's engage policymakers" or "lets ban X, Y and Z", but a clever nerdy technical solution that redefines the very foundation of machine learning. He explains his solution beautifully in the book, so below I'll merely attempt to give a rough sense of the key idea.
The "standard model" of AI is to give a machine learning system a goal, and then train it using lots of data to get as good as possible at accomplishing that goal. That's much of what my grad students and I do in my MIT research group, and that's what Facebook did when they trained an AI system to maximize the amount of time you spent on their site. Sometimes, you later realize that this goal wasn't exactly what you wanted; for example, Facebook switched off that use-time-maximizing system after the 2016 US and Brexit votes made clear to them that it had created massive online echo chambers that polarized society. But if such a value-misaligned AI is smarter than us and has copied itself all over the internet, it's not easy to switch it off, and it may actively try to thwart you switching it off because that would prevent it from achieving its goal.

Stuart's radical solution is to ditch the standard model altogether for sufficiently powerful AI-systems, training them to accomplish not a fixed goal they've been given, but instead to accomplish *your* goal, whatever that is. This builds on a technique known by the nerdy name "Inverse Reinforcement Learning" (IRL) that Stuart has pioneered, and completely transforms the AI's incentives: since it can't be sure that it's fully understood your goals, it will actively try to learn more about what you really want, and always be open to you redirecting it or even switching it off.

In summary, this book is a captivating page-turner on what is arguably the most important conversation of our time: the fate of humanity when faced with machines that can outsmart us. Thanks in large part to Russell, IRL is now a blossoming sub-field of AI research, and if this book motivates more people to deploy it in safety-critical systems, then it will undoubtedly increase the chance that our high-tech future will be a happy one.
146 people found this helpful
✓ Verified Purchase

Robbie and Harriet

This popular book by a well-known artificial intelligence researcher is a promotion of his views. He believes computers with superhuman intelligence will be available by the end of this century. This sets him in opposition to skeptics like psychologist Steven Pinker and in the ranks of futurists like Ray Kurzweil. (This is a conntroversial field.) He frequently cites AlphaGo, a Go playing system, as an example of his point of view. The design team “didn’t design decision procedures that work only for Go. Instead they made improvements to two general purpose techniques-- lookahead search to make decisions and reinforcement learning to learn how to evaluate positions-- so that they were sufficiently effective to play Go at a superhuman level.” The history of Artificial Intelligence (AI) has until now been dominated by machines whose rigid objectives come entirely from their designers. He wants machines which build their own objectives, relying on human input to tell them in a casual way what they are doing is right or wrong. They don’t exactly reach goals, but rather maximize by trial and error the values of utility functions that represent the benefit of their actions.

The author shows that ordinary evolution has contributed a sort of intelligence to an organism as simple as the bacterium E. coli, as simply as a preference for solutions containing glucose. Of course it has no nervous system or brain, so its seeking to maximize the utility function, “glucose concentration” is undeveloped at best. Another example is the Baldwin effect, in which the processes of learning and evolution jointly produce seemingly intelligent behavior, shows how they work together. The more intelligent organism learns faster.

The world of AI is dominated by the idea of intelligence as a matter of making choices, so the author looks at measures to evaluate the payoff, generally called 'utility', resulting from a choice as to how to act. He considers the complications involved in games where two or more people are in competition. (The players perform opposing utility calculations.) He weaves into this a sketch of computer history and the techniques of AI, explaining the concept of a machine (an 'intelligent agent') with a number of 'internal states'. In response to inputs from an external world, the machine acts over time, emitting outputs, to survive and thrive in a possibly unfavorable environment. 'Lookahead search', which previews the effects of possible actions, is essential here, as is the building of 'knowledge based systems'. A difficulty here is that very little of our foreknowledge is absolutely certain. To take account of this we need 'Bayesian networks' in which a probability is associated with an action. The modern ideal of rationality is to maximize of expected utility. Frequently the number of inputs the system may encounter is enormous and also the number of states, with a final reward coming only a long time in the future; then a technique of 'reinforcement learning' is useful. The final goal is to create a system that is more intelligent than a human and might even produce a version of itself that would be more intelligent than its parent.

The techniques outlined above enable us to build specialized systems for dealing with complex board games, recognizing handwritten numerals, controlling cars (somewhat), and acting as personal assistants or domestic robots (also somewhat). An enormous step forward would be general-purpose AI, one general-purpose program that does everything, but this is at present quite impossible, so instead we attempt to build different types of agent programs for different types of problems. This is not explained in any detail. Language contains a huge amount of human knowledge. It would be much too wasteful for the machine to “rediscover it”. Probably there a bootstrapping process, but it’s not as easy as it looks. An example is the NELL project at Carnegie-Mellon University, which seems to misunderstand much of its input and puts out many nonsensical phrases. He comments “Reading requires knowledge and knowledge (largely) comes from reading. Again, language is the only way out but apparently an insurmountable problem.”

There is a widespread view today, shared by the author, that computers will in a hundred years or so be able to think for themselves. The most interesting question of superintelligent machines was raised in a serious way in 1950 by A. M. Turing himself, who suggested “turning off the power at strategic moments” when we find the behavior of the machine disconcerting. What if we just pull out the plug? The author (Russell) replies that “a superintelligent entity will have already thought of that possibility and taken steps to prevent it. And it will do that not because it wants to stay alive but because it is pursuing whatever objective we gave it and knows that it will fail if it is switched off.” (p. 161) It necessarily will possess vast knowledge to fulfill its “prime directive” of doing anything possible. But an un-powered machine can’t do anything. Therefore, the structure of the machine will dictate that it do everything in its vast array of powers to stay powered. There is an assumption here: the AI has all the knowledge and all the powers of the world at its disposal, so that no matter what the circumstances it can devise some sequence of actions to frustrate any attempt to cut off the power. Therefore it will not let anyone unplug it or even go near it. But the AI has all the knowledge of the world at its disposal. I don’t buy this argument, because the idea of a computer with the ability to foresee a possible human intention or block it in the near future is preposterous. Does the computer possess extra-sensory perception?

The author is an advocate of personally oriented devices that are devoted to humans specifically, that are humble about what they understand about their users, and try to predict human preferences. (His anthropomorphism) It’s very tiresome to wade through the Robbie-and-Harriet vignettes illustrating the petty misunderstandings between a woman and her superintelligent digital personal assistant. The problem as he himself admits is getting machines to understand natural language, not to mention understanding how humans express their preferences. When he depicts future personal digital assistants as if they were already people, the reader naturally feels suspicious.
12 people found this helpful
✓ Verified Purchase

The Best Treatment of Latest in AI and How it could evolve

Mr Russell is a very good writer. The book starts with list of five candidates for "biggest event in the future of humanity" - Catastrophe (e.g., Ebola), Methuselah (immortality), #BreakingPhysics (faster than light travel), Discovering Alien Civilization and "Superintelligent AI". Author emphasizes why the last would be most impactful out of the five. It makes intuitive sense - among state of the art broom, cleaning solution and a robot that can clean everything, the last one will always win the vote.

The author recalls how the problem of liberating nuclear energy went from impossible to solved in under a day. It is an interesting analogy for all complex problems, including that of superintelligent AI.

Some insights from the book -

* Uncertainty about objectives in humans is a feature, not a bug. We should follow the same while building AI. Logic requires certainty. Early "Expert Systems" tried to wrap logic onto AI. As AI community embraced uncertainty and created measure to incorporate it in design (e.g., Judea Pearl's Bayes Net), it evolved fast out of "ice age".
* Neurons carry signals rapidly at a rate of 270 miles-per-hour
* Jellyfish have no brains - they move thanks to a "nerve net"
* Our brains have 10^11 neurons and 10^15 synapses and "cycle time" of a few milliseconds per state change. It is slow compared to a computer, but extremely power efficient.
* While Moore's law has hit a physical constraint - circuit dimensions are already just a few atoms wide - newer "ways", especially, TPU (Tensor Processing Units), Quantum Hardware should keep scaling computing powers at same or higher rate.
* Research and progress toward "tool AI" (or, narrow/specific solutions, e.g., Playing Go) often makes great progress towards general-purpose AI.
* "Theory of probabilities is just common sense reduced to calculus" - Laplace
* First successful Reinforcement learning system was Arthur Samuel's Checkers program
* Reflex agents (like emergency braking in autopilot) "implement a designer's objective but do not know what the objective is or why they are acting so". In other words, they are taking actions as proxy of the designers, not as themselves! This makes reflex agents extremely inflexible and hard to use outside a very narrow band of use cases.
* Loss functions are often "mono dimensional", i.e., they assign the same weight to every type of error. The author sites an unfortunate case where Google image-labeling service labeled a human and his friends as Gorillas. i.e., the loss function assumed the cost of misclassifying a human as Gorilla is same as misclassifying a dog as one!
* The first smart home controller - ECHO - was built in 1966 by James Sutherland. It weighed 800 lbs, consumed 3.5kW and managed 3 digital clocks and TV antenna.
* Dexterity is hard - "Most robots can't pick up most objects most of the time". e.g., shake exactly 2 pills out of a bottle
* It's said about Da Vinci that he never learned to paint. He painted to learn. Deep Reinforcement Learning needs to crack the problem of "commonsense knowledge" in language by reading/conversing. CMU's NELL has acquired 120 million language beliefs in last 9 years and yet has confidence only on 3% of it
* In Science, discovery of new concepts are generally attributed to 3 Is - Intuition, Insight and Inspiration
* "Civilization advanced by extending the number of important operations which we can perform without thinking about them" - Alfred N Whitehead (1911). AI systems could use such abstractions (e.g., take a driving decision based on time of the day, weather and the tire pressure - without any hard-coded) to take a decision
* Humans have a far more flexible computational architecture to discover and use high-level actions (e.g., divide a Go board into segments and think in parallel on each), but have a tiny short-term memory and slow hardware that " severely limit out ability to look into future, handle large number of contingencies and consider large number of alternative plans"
* Search engines add economic value of about $17,500 per user per year!
* Collection of N humans is far less efficient than collection of N machines because information in one brain can only be transferred to another by buggy language and a low bandwidth channel. That's why N humans "spend most of their time in meetings"!
* General purpose AI would be "EaaS" (Everything as a Service)
* The author's suggestion for a AI-ready society (which has to be more equitable) "our cultures to gradually down-weight pride and envy as central elements of perceived self-worth".
* Generally, automation increases the share of income going to capital and decreases the share going to labor.
* The Great decoupling - after 1973, there is a huge chasm between growth in productivity and growth in wages.
* Mechanical transportation became cheaper than upkeep of a horse, so horses became pet food!
* On the advice "Learn to code" and worst case scenario of AI impacting jobs - "Data Science is a very tiny lifeboat for a giant cruise ship"
* A machine that is uncertain about "true objective" will exhibit a kind of humility. e.g., it will allow itself to be switched off.
* In one of the last chapters, author introduces IRL (Inverse Reinforcement Learning) - "while reinforcement learning generates behavior from rewards, we actually wanted the opposite: to learn the rewards given the behavior".
* First principle of Beneficial AI: a machine's only purpose is the realization of human preferences.
* Implicit rules are hard, e.g., self-driving cars are terrible at handling 4-way stop signs when it's not clear who has right of way

Overall, "Human Compatible" is not only a highly readable book but it also does the best job among the recent pop-AI books to (a) go over the technical and algorithmic foundations of AI, (b) ethical, economic and on-premise challenges for and from AI, and (c) introduces a new paradigm, framework and philosophy that could change the way such systems are designed.
10 people found this helpful
✓ Verified Purchase

Outdated by years already. Written in a dull pedantic style.

Useless knowledge that the author gleaned from other sources in my opinion. The material is dated don't waste your money.
6 people found this helpful
✓ Verified Purchase

Nice review of AI -- Unfortunately the authors solution is not reasonable

Stuart Russell has written 1/2 of a good book. It is a review of the development and capabilities of AI (artificial intelligence). Unfortunately, the second half is an argument for how to make AI safe for humans, and the solution, that the AI device act for maximum utility of the human race (and that is basically what it comes down to) is impossible.

The issue that the author is worried about is what if there is a super-AI computer with access to all the knowledge in the world (impossible in and of itself, but let's stipulate that). And, say that you want the computer to do something for you, but the solution that the computer finds to do it is harmful to humans. Presumably, if that situation occurred, you would turn the machine off, but, unfortunately, the computer knows this and kills you instead so that it can accomplish its goal.

To solve the problem, just program the computer to do what you want instead of a particular task. And, so that it doesn't harm humanity, it also takes into account the utility functions of everyone in the world and tries to maximize that at the same time as accomplishing your task. (Forget for a moment that maximizing the world's utility and accomplishing your tasks are different objective functions.)

So, what is the problem with this? The first problem is convincing someone to produce a machine that will try to maximize the utility of humanity -- who would buy such a thing? People want to buy a machine to accomplish a specific goal and not to maximize the utility of humanity. The second problem is how to force everyone who creates an AI machine to follow this principle. For example, how about a military organization which wants to create an autonomous soldier -- a pretty extreme case, but how do you manufacture an autonomous AI soldier which is trying to maximize the utility of humanity. The third problem is that maximizing the utility of humanity is an impossible optimization problem. Maybe the solution to the problem is not the current arrangement, but something more like The Matrix. So the computer goes about changing the world into The Matrix. Or maybe the solution is to create some kind of drug that allows people to exist as contented zombies. Even if one restricts oneself to situations "close" to the current situation, the idea that some uber-computer could sit above humanity and manage the economy to the benefit of everyone is a socialist's dream. This has been proven to be impossible by economists such as Hayek. There are just too many variables to deal with. And, how would an AI machine designed to shine one's shoes or cut one's hair, for example, go about maximizing the utility of all of humanity?
4 people found this helpful
✓ Verified Purchase

Nice review of AI -- Unfortunately the authors solution is not reasonable

Stuart Russell has written 1/2 of a good book. It is a review of the development and capabilities of AI (artificial intelligence). Unfortunately, the second half is an argument for how to make AI safe for humans, and the solution, that the AI device act for maximum utility of the human race (and that is basically what it comes down to) is impossible.

The issue that the author is worried about is what if there is a super-AI computer with access to all the knowledge in the world (impossible in and of itself, but let's stipulate that). And, say that you want the computer to do something for you, but the solution that the computer finds to do it is harmful to humans. Presumably, if that situation occurred, you would turn the machine off, but, unfortunately, the computer knows this and kills you instead so that it can accomplish its goal.

To solve the problem, just program the computer to do what you want instead of a particular task. And, so that it doesn't harm humanity, it also takes into account the utility functions of everyone in the world and tries to maximize that at the same time as accomplishing your task. (Forget for a moment that maximizing the world's utility and accomplishing your tasks are different objective functions.)

So, what is the problem with this? The first problem is convincing someone to produce a machine that will try to maximize the utility of humanity -- who would buy such a thing? People want to buy a machine to accomplish a specific goal and not to maximize the utility of humanity. The second problem is how to force everyone who creates an AI machine to follow this principle. For example, how about a military organization which wants to create an autonomous soldier -- a pretty extreme case, but how do you manufacture an autonomous AI soldier which is trying to maximize the utility of humanity. The third problem is that maximizing the utility of humanity is an impossible optimization problem. Maybe the solution to the problem is not the current arrangement, but something more like The Matrix. So the computer goes about changing the world into The Matrix. Or maybe the solution is to create some kind of drug that allows people to exist as contented zombies. Even if one restricts oneself to situations "close" to the current situation, the idea that some uber-computer could sit above humanity and manage the economy to the benefit of everyone is a socialist's dream. This has been proven to be impossible by economists such as Hayek. There are just too many variables to deal with. And, how would an AI machine designed to shine one's shoes or cut one's hair, for example, go about maximizing the utility of all of humanity?
4 people found this helpful
✓ Verified Purchase

Well-written and comprehensive overview of AI

An extremely well-written, comprehensive overview of Artificial Intelligence (AI) — with a focus on the very real risks it poses to the continued viability of the human race and a proposal for how to move forward reaping the benefits of AI without making us “seriously unhappy.”

AI Pioneer Stuart Russell is a Professor of Computer Science at UC Berkeley, has numerous awards, fellowships, chairmanships, etc. and has co-authored a textbook on AI with Peter Norvig. This is a book written by that rare creature — someone who knows his subject thoroughly and can explain it. He does not shy away from the complexity of the topic but breaks it down and explains it, simply making it accessible to anyone who is willing to read and think. He includes short, clear examples from science, philosophy, history, and even science fiction and references current and historical work from academia, research labs, and startups from around the world.

The book is divided into three parts: the concept and definition of intelligence in humans and machines; a set of problems around the control of machines with superhuman intelligence; and a proposal for shifting our approach to AI to prevent these problems from occurring rather than trying to “stuff the genie back into the bottle” once it is too late.

Russell explains the potential problems of unleashing a massively intelligent machine on the world. An AI machine offers incredible scale. Think of an entity that (with the proper sensors) can see the entire physical world at once, that can listen and process all concurrent conversations at once, that can absorb all the documented history of the planet in a single hour. And we plan to control this entity via programming. With a superhuman intelligence, the programming would need to be at the objective level. And yet — specifications — even with every day human programmers — are incredibly hard to get right. Russell uses the example of giving the machine the task to counter the rapid acidification of the oceans resulting from higher carbon dioxide levels. The machine does this in record time, unfortunately depleting the atmosphere of oxygen in the process (and we all die). Remember the old stories about getting three wishes and always screwing it up? This would make those stories look trivial. Russell never uses scare tactics and does not wildly overstate the thesis — instead he uses practical examples and includes one tremendously simple chapter (the not-so-great debate) that lists every argument people have made that we don’t have to worry and rebuts them quickly.

His solution: we should design machines correctly now so we don’t have to try to control them later. He wants to build a “provably beneficial machine” — provably in the mathematical sense. His machine would operate on only three principles: the machine’s only objective is to maximize realization of human preferences; the machine is initially uncertain as to what these preferences are; and the ultimate source of information on human preferences is human behavior. This is interesting — he wants to “steer away from the driving idea of 20th century technology that optimize a given objective” and instead “develop an AI system that defers to humans and gradually align themselves to user preferences and intentions.” There follows an entire chapter devoted to how we can program the machines to determine what those human preferences are, particularly in light of competing preferences, potentially evil preferences, the cognitive limitations of humans to understand their own preferences, behavioral economics, the nature of mind, definitions of altruism — you name it — all the fascinating areas of understanding human behavior become part of the problem. Which, while completely fascinating, strikes me as even more difficult than trying to work out how to define exact specifications in the first place!

I was left with a knot in my gut about how fast AI is moving without much oversight and how suddenly relevant these issues (that I had long relegated to comfortable musings in science fiction) have become. While I find his proposed solution intriguing, it is hard, hard, hard — and expecting random investors and startups to tackle harder design problems instead of racing towards monetization will be tricky. On the other hand, we move forward as a civilization by raising the issues and embedding them in our moral consciousness and Russell has done an excellent job of clearly teeing up a huge number of costs, benefits, and issues from technical to ethical. Highly recommended if you have any interest in the topic.
3 people found this helpful
✓ Verified Purchase

Important to read but may gloss over the critical driver

The crux of Russell's argument and of his proposal for avoiding a nuclear arms-like race pushing AI to the point where it can severely damage us as individuals and as a species is this in Chapter 7: "The first reason for optimism is that there are strong economic incentives to develop AI systems that defer to humans and gradually align themselves to user preferences and intentions."

Perhaps But: In today's world the drive for economic and strategic supremacy is the undeniable reality and the prospects of economic and strategic advantage seem destined to continue to be the forces that govern advancement of AI science and engineering and, especially, determine which applications get developed and deployed.

And unlike nuclear arms technologies, which because of cost and extremely narrow availability we can at least attempt to control by convention and centralized authority, AI technologies are and will continue to be available to pretty much anyone who wants them and, as compute power simultaneously expands and declines in cost, anyone will be able to deploy them.

At the end of the day, because it is and will continue to be impossible to control who builds what AI, the critical driver of their use will be how much AI advances the goals of the individuals, organizations, and nations that use AI.

Is it a certainty that the economics of good-actor AI can be made so strong that bad-actor AI will be unattractive to all but the generally incompetent or easily thwarted?

I don't think it is. Consider for example the Cybersecurity AI arms race.

How AI develops and how it is used across human civilization will depend on how strongly and uniformly human civilization values Share and Share Alike and uniformly rejects Get Everything You Can Just Don't Get Caught.

Anyone considering the recent trajectory of American politics, and for example the predestined outcome of the "battle" over healthcare, can see that the latter is increasingly the order of the day.

As a thought experiment ask yourself this: If a CEO were faced with authorizing the development of AI-1 that would increase market share and profits by 20% but might injure or even kill a few people vs. AI-2 that would increase market share by 2% but wouldn't kill or injure anyone, what would be the decision?

The problem for Russell's economic argument is that some CEOs would almost certainly authorize AI-1, some CEOs might authorize AI-2, and the more important AI-1 is to organizational survival or potential supremacy the more likely the CEO would authorize AI-1. For example premature release of autonomous driving AI to production in order to beat a competitor to market or an end-of-quarter stock analyst assessment.

Although I'd love to see it I'm not optimistic - nor indeed is Russell - that you, I, and well-meaning level headed leaders of the fields that yield increasingly potent AI can have much influence in positioning Share and Share Alike as the primary driver of the direction we steer in discovering, designing, and deploying AI.

Unfortunately the evolution of AI might well be increasingly driven by the same factors that currently appear to be driving the evolution of human civilization.
1 people found this helpful
✓ Verified Purchase

Preface (reproduced here)

"Why This Book? Why Now?

This book is about the past, present, and future of our attempt to understand and create intelligence. This matters, not because AI is rapidly becoming a pervasive aspect of the present but because it is the dominant technology of the future. The world's great powers are waking up to this fact, and the world's largest corporations have known it for some time. We cannot predict exactly how the technology will develop or on what timeline. Nevertheless, we must plan for the possibility that machines will far exceed the human capacity for decision making in the real world. What then?

Everything civilization has to offer is the product of our intelligence; gaining access to considerably greater intelligence would be the biggest even in human history. The purpose of the book is to explain why it might be the last event in human history and how to make sure that it is not.

Overview of the Book

The book has three parts. The first part (Chapters 1 to 3) explores the idea of intelligence in humans and in machines. The material requires no technical background, but for those who are interested, it is supplemented by four appendices that explain some of the core concepts underlying present-day AI systems. The second part (Chapters 4 to 6) discusses some problems arising from imbuing machines with intelligence. I focus in particular on the problem of control: retaining absolute power over machines that are more powerful than us. The third part (Chapters 7 to 10) suggests a new way to think about AI and to ensure that machines remain beneficial to humans, forever. The book is intended for a general audience but will, I hope, be of value in convincing specialists in artificial intelligence to rethink their fundamental assumptions."

-------

I highly recommend this book.

For those who are already familiar with AI safety, a good summary which seems to me to capture the important claims/arguments is Rohin Shah's summary on the AI Alignment Forum / LessWrong "[AN #69] Stuart Russell's new book on why we need to replace the standard model of AI".
1 people found this helpful
✓ Verified Purchase

Must-read exposition of serious challenges in AI alignment

This book will greatly help AI professionals understand key arguments, avoid classic missteps, and appreciate the serious challenge humanity faces in aligning artificial intelligence systems. Russell straightforwardly debunks common objections, writing with both candor and charm.