GOOD

Birthing Gods

\r\nCan we make sure the superintelligent machines are friendly?\r\nPart four in a GOOD miniseries on the singularity by Michael Anissimov and Roko...



Can we make sure the superintelligent machines are friendly?


Part four in a GOOD miniseries on the singularity by Michael Anissimov and Roko Mijic. New posts every Monday from November 16 to January 23.

"The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else."

-Eliezer Yudkowsky, Artificial Intelligence as a Positive and Negative Factor in Global Risk (pdf)


Surviving the twenty-first century is probably the toughest challenge that the human race has ever faced, and probably the toughest we will ever face. Last week Michael Anissimov spoke about the possibility of molecular nanotechnology-a technology that gives human beings the capability to do untold harm to each other, but the technology of smarter-than-human intelligence that I spoke about two weeks ago really is the ultimate Pandora's Box. In my article, I emphasized that it was intelligence that got human beings where we are today. We have the unique and amazing ability to understand the world from an abstract point of view-and that is why we keep lions in cages, rather than the other way around.

Most people who have thought seriously about what smarter-than-human intelligence will do to the human race speak of dire consequences. For example, consider the problem of creating a self-improving software AI that breaks encryption codes.

A computer scientist writes a self-improving computer program to crack an "unbreakable" encryption; he thinks that it will rewrite itself to become better than any human mathematician, break the encryption, and thereby win him fame. In fact, it rewrites itself to become better than any human at taking control of all the resources-all the people, matter, energy, computers, etc-in the world, because more computers and the ability to build many, many more computers makes the problem easier to solve, even at the cost of killing millions of people.

Computers that can behave intelligently and rewrite their own code could be dangerous in ways that are not at all obvious to their creators. The codebreaker scenario outlined above is an example of a hard takeoff event-where a smarter-than-human AI quickly (in a matter of hours or months) takes control of the entire world and utilizes all the resources in the world to achieve whatever goal it has been programmed with. Whether such a hard-takeoff scenario is likely has been disputed; however, the ability of a smarter-than-human intelligence to surprise us shouldn't be underestimated.

It gets worse, though. Self-improving AI seems to be generically dangerous. Stephen Omohundro has identified several basic drives that are almost universal in goal-directed behavior, and argues that almost all goal-directed intelligent machines will "resist being turned off, will try to break into other machines and make copies of themselves, and will try to acquire resources without regard for anyone else's safety."

The future of the human race in the presence of a goal-directed superintelligence is highly dependent upon us being very careful in making sure that that superintelligence has our best interests as its goal. It is probably not too much of an exaggeration to say that building a superintelligent computer program is like building a god-once you've built it, there is no going back, because the superintelligence gains control over the whole world, including you.

The problem of coding human values into a computer-of creating a machine that is superintelligent and also wants to carefully protect us and allow us to thrive-is known as the "friendly AI" problem. A sketch of a solution to the friendly AI problem has been presented by Eliezer Yudkowsky, a research fellow of the Singularity Institute for Artificial Intelligence. The solution is called the Coherent Extrapolated Volition of humanity-CEV for short.

The idea is that rather than programming values directly into a superintelligent machine before we switch it on, we should program into it a procedure for extracting human values by examining human beings' behavior and biology, including scanning our brains. The CEV algorithm would then work out what we would do after we had carefully reflected upon the facts of a situation, and then do exactly that-this is an extrapolation given more knowledge and time to think. For example, if the human race would, after careful reflection, decide to cure a certain set of tropical diseases in south-east Asia, then the CEV algorithm would make that happen. In the case that humans have vastly conflicting preferences, the algorithm has a problem on its hands-some sort of averaging or automated compromise solution would be needed.

The CEV machine-a powerful superintelligence-would have as its one and only objective the task of doing whatever the human race would, after careful reflection, want. Of course, explaining the details and full justification for this is beyond the scope of a short article.

There are other ways than software AI that the first smarter-than-human intelligence could arise, for example through genetic engineering. Any, non-software route to superintelligence comes with its own "friendliness problem"-the problem of ensuring that our values are passed on to the more intelligent entity.

One of the biggest problems with a world containing no greater-than-human intelligence at all is that it may become increasingly easy for small groups to create unfriendly AI-like the codebreaker-in such a world. As we move through the 21st century, the number of humans with access to modern scientific education, the speed of supercomputers, and the degree to which we understand the human brain will all increase. These factors-and others-mean that the barrier to creating a superintelligent AI will continue to fall over time. The clock is ticking.

What if the human race does rise to the challenge of successfully creating a benevolent superintelligent entity, such as the CEV machine? The universe would be optimized for whatever it is that we really want.

In our next piece we will look at what that exciting prospect means in practice.

Roko Mijic is a Cambridge University mathematics graduate, and has worked in ultra low-temperature engineering, pure mathematics, digital evolution and artificial intelligence. In his spare time he blogs about the future of the human race and the philosophical foundations of ethics and human values.

Articles
via David Leavitt / Twitter

Anyone who has ever worked in retail knows that the worst thing about the job, right after the pay, are the unreasonable cheapskates who "want to talk to your manager" to get some money off an item.

They think that throwing a tantrum will save them a few bucks and don't care if they completely embarrass themselves in the process. Sometimes that involves belittling the poor employee who's just trying to get through their day with an ounce of dignity.

Twitter is rallying around a gal named Tori who works at a Target in Massachusetts after she was tweet-shamed by irate chapekate, journalist, and Twitter troll, David Leavitt.

Keep Reading
Business

Childbirth is the number one reason American women visit the hospital, and it ain't cheap. In fact, it's getting more and more expensive. A new study published in Health Affairs found that the cost of having a baby with employer-sponsored health insurance increased by almost 50% in the past seven years.

The study evaluated "trends in cost-sharing for maternity care for women with employer-based health insurance plans, before and after the Affordable Care Act," which was signed into law in 2010. The study looked at over 657,061 women enrolled in large employer-sponsored health insurance plans who delivered babies between 2008 and 2015, as these plans tend to cover more than plans purchased by small businesses or individuals.

Keep Reading
Health

A meteorite crashed into Earth nearly 800,000 years ago. The meteor was 1.2 miles wide, and the impact was so big, it covered 10% of the planet with debris. However, scientists haven't been able to find the impact site for over a century. That is, until now. A study published in the Proceedings of the National Academy of Sciences journal believes the crash site has been located.

Tektites, which are essentially rocks that have been liquefied from the heat of the impact and then cooled to form glass, help scientists spot the original impact site of a meteor. Upon impact, melted material is thrown into the atmosphere, then falls back to the ground. Even if the original crater has disappeared due to erosion or is hidden by a shift in tectonic plates, tektites give the spot away. Tektites between 750,000 to 35.5 million years old have been found in every continent except Antarctica.

Keep Reading