IBM researchers get closer to brain-like computing


Credit: DARPA

IBM (s ibm) researchers say they have whipped up a programming language, algorithms and applications to deploy on top of a computing system inspired by the human brain.

This is the latest progress IBM and collaborating groups have made for a little project from the federal Defense Advanced Research Projects Agency (DARPA) dubbed Systems of Neuromorphic Adaptive Plastic Scalable Electronics, or SyNAPSE.

“It’s a very modest goal — it’s to build a brain-like computer,” said Dharmendra Modha, principal investigator and senior manager at IBM Research.

In recent years, Modha and his fellow researchers have simulated a brain larger than a cat cortex and built chips that function like a human brain (pictured), adapting in response to new information in real time.

Simulating a brain in software

Researchers have since developed a massively parallel, multi-threaded software simulator for its brain-like architecture. The simulation covers 2 billion neurosynaptic cores all connected with one another, representing around 100 trillion synapses, which is roughly equivalent to the number inside of a human brain, Modha said.

Each one of those neurosynaptic cores can be thought of as a building block, which contains 256 neurons recreated on chips for computation, 256 replicated axons for bringing information to the core from external neurons and tens of thousands of synapses providing memory.

In addition to the simulation, researchers have also built a neuron mathematical model, or a set of equations, that can be easily implemented with a few transistors in silicon, Modha said.

A language to speak brain tongue

They’ve created a new object-oriented Corelet programming language for designing, developing and debugging corelets, which are abstractions of a network of neurosynaptic cores. The corelet system and language are more appropriate than traditional computing methods for handling lots of simple tasks and processing them at the same time the way a human brain does.

The researchers also formulated an environment in which developers can work with the corelets and wrote a bunch of algorithms and applications — such as predicting sequences, detecting motion or looking at a music score from Bach or Beethoven and identifying the composer. Finally, they created a curriculum, so others can learn about the Corelet language, the library of algorithms and applications, the chip simulator and other parts of the project.

DARPA has given IBM $12 million to take the SyNAPSE forward into the next phase. The money will go toward improving hardware, software and education, a spokesman wrote in an email. Future work could involve packing more neurons into physical gear and testing for more complex perception of the outside environment.

Cognitive computing has been a focus at IBM at least as far back as 2006, and the company was simulating neural nets as early as 1956.

Imagine the possibilities

These latest breakthroughs bring IBM — and DARPA — a few steps closer to being able to run a sensor-equipped device at a very small size and with very low power use. That could lead to lots of custom applications, such as drones and ground-based unmanned vehicles, but it could also be mass-produced to usher in new types of personal and industrial computing down the line.

But Modha didn’t speak of products in the funnel. So we are left to wonder with our brains — and now it’s clearer just how powerful they actually are — what other tiny technology could result from the work and how soon ordinary people will have access to it.

To see Modha talk at length about the SyNAPSE project, check out this 2010 video interview with him and my colleague Stacey Higginbotham:




Brain like computing.

Does that mean it will forget what it was supposed to be working on, screw up calculations, get facts wrong or simply work on whatever the hell it wants to?

John Bailo

Does the human brain use algorithms? Have they proven that?

Nola Garcia

Recent research indicates the brain has both innate and learned algorithms. A comment on this piece: well written and informative. The work being done here is fascinating.

Not T

Mr T, commenting above, twice. Please educate yourself more on the subject. There is much more than the old stuff here that you don’t realize.

What these folks are doing is ground breaking.


Putting in old and limited neural network stuff that does not work well into a big chip will not make it work any better.
Instead new research is squashed, universities are starved, taxpayer money is wasted.
Bravo big corporation, bravo big sales fluff team.


This is BS: putting in old stuff that does work into a big chip will not make it work any better.
Everyone there is just fluffing up their egos and wasting taxpayer’s money.


They should use this with the quantum computers and 23andme to manipulate the foxo gene and turn off aging.

Fredrik Asplund

Very fascinating stuff!

Linked on Sprawler:

Comments are closed.