Making Computers Think Like Humans: Time to Panic?

IBM has invented computer chips that learn. It's not scary, we swear.

This is your brain on chips.

Computer chips have followed Moore’s Law since Intel founder Gordon Moore came up with it in 1965: Every two years, the number of transistors engineers can place on a circuit board has roughly doubled, and so has processing speed. But now, chip-makers are running into a wall—physics—and soon (perhaps as soon as 2015) miniaturization will no longer increase computing power.

To defeat that trend, an international team of IBM researchers has announced the development of a new kind of cognitive chip that doesn’t need software: instead, it is modeled on the human brain’s ability to learn, reorganizing itself the same way that neurons do.

Hearing the news, nerds were quick to crack jokes about “Skynet,” the malevolent artificial intelligence from the Terminator movies, or recall The Matrix, Dune, or a hundred other science fiction sagas that rely on the conceit that when computers start learning, humans find they've opened up a Pandora's box of troubles.

Nerves aren’t calmed by the news that some funding for the project comes from the Defense Advanced Research Project Agency, the Pentagon think tank that tends to be the subject of conspiracy theories but has produced such inventions as the internet and GPS.

“There’s certainly nothing to fear here; we’re talking about very, very simple capabilities,” Dr. Chris Kello, a cognitive scientist at the University of California, Merced, who is working on the project. He’s a little miffed that I’ve asked. “Nobody talks about Terminator every time you pick up a phone and talk to a computer when you’re booking a flight.”

On the other hand, those voice recognition computers function so poorly they can’t possibly be a threat, unless they’re just smart enough to maliciously enjoy hoarse, angry people screaming “existing reservation!” again and again.

Regardless, Kello explains that the key to the next generation of computers is creating a processor that handles data en masse with “lots of very small neuron-like assessing units [that] coordinate to solve problems,” rather than one big processor processing data in order.

“If you think about how amazing our visual capabilities are, the ability to take a large amount of data coming in […] that data is very noisy, and represents a very complex world,” he says. “For the human visual system to be able to put together the pieces of the world through this noisy is array is something that current computers really can’t touch.”

Mello’s job is to help test and teach the chip how to learn by applying what we know about how the human brain gathers information and adapts to it. At Merced, they’ve rendered a model of their campus that virtual agents will be asked explore.

Is this where androids dream of electric sheep?

“The tests are much what you might imagine testing animal intelligence, or human intelligence, for that matter, “ he says. “You put organisms in an environment and give them tasks and see how they do.”

Organisms? IBM’s press release notes the chips contain no biological materials, and they don’t take advantage of nanotechnology, either, a frequently-cited path toward brain-like computers. But, Kello says, “in all of those cases you’re dealing with systems where uncertainty is the rule rather than the exception. “

The ability to handle uncertainty and interpret massive amounts of data at the same time will make cognitive computers far more effective at challenges that involve interacting with the real world, from monitoring busy intersections to collecting tsunami warning data.

Right now, the researchers are only at phase two of the project, testing and refining the chips with neuron levels of 10 to the sixth power. Only in phase four, according to a schedule laid out by DARPA, does the team imagine building a robot that contains chips of 10 to the eighth power, just below the 10 to the 10th power neuron density in the human brain. At that point, we at GOOD will happily welcome our new robot overlords.

Or just share in the advantages of a major 21st-century innovation. We’ll see how it turns out.

Photos courtesy IBM, U.C. Merced.

via National Nurses United/Twitter

An estimated eight million people in the U.S. have started a crowdfunding campaign to help pay for their own or a member of their household's healthcare costs, according to a survey released Wednesday.

The poll, which was conducted by the National Opinion Research Center (NORC) at the University of Chicago, also found that in addition to the millions who have launched crowdfunding efforts for themselves or a member of their household, at least 12 million more Americans have started crowdfunding efforts for someone else.

Keep Reading
via Library of Congress

In the months after the Japanese bombed Pearl Harbor in 1941, President Franklin D. Roosevelt authorized the military to move Japanese-Americans into internment camps to defend the West Coast from spies.

From 1942 to 1946, an estimated 120,000 Japanese Americans, of which a vast majority were second- and third-generation citizens, were taken from their homes and forced to live in camps surrounded by armed military and barbed wire.

After the war, the decision was seen as a cruel act of racist paranoia by the American government against its own citizens.

The internment caused most of the Japanese-Americans to lose their money and homes.

Keep Reading

Step by step. 8 million steps actually. That is how recent college graduate and 22-year-old Sam Bencheghib approached his historic run across the United States. That is also how he believes we can all individually and together make a big impact on ridding the world of plastic waste.

Keep Reading
The Planet