Cognitive Computing

Deep learning inference possible in embedded systems thanks to TrueNorth

Share this post:

Scientists at IBM Research – Almaden have demonstrated that the TrueNorth brain-inspired computer chip, with its 1 million neurons and 256 million synapses, can efficiently implement inference with deep networks that approach state-of-the-art classification accuracy on several vision and speech datasets. This will open up the possibilities of embedding intelligence in the entire computing stack from the Internet of Things, to smartphones, to robotics, to cars, to cloud computing, and even supercomputing.

TrueNorth data set samples.

TrueNorth data set samples.

The novel architecture of the TrueNorth processor can classify image data at between 1,200 and 2,600 frames per second while using a mere 25 to 275 mW, which is effectively greater than 6,000 fps per Watt. Like that kung fu master in the movies who simultaneously fights assaults from many opponents, this processor can detect patterns in real time from 50-100 cameras at once – each with 32×32 color pixels and streaming information at the standard TV rate of 24 fps – while running on a smartphone battery for days without recharging.

The breakthrough was published this week in the peer-reviewed Proceedings of the National Academy of Sciences (PNAS). The essence of the innovation was a new algorithm for training deep networks to run efficiently on a neuromorphic architecture, such as TrueNorth, by using 1-bit neural spikes, low-precision synapses, and constrained block-wise connectivity—a task that was previously thought to be difficult, if not, impossible. The neuromorphic deep networks can be specified and trained with the same ease-of-use as contemporary deep learning systems such as MatConvNet, thus removing the need for data scientists to learn the intricacies of TrueNorth’s architecture.

“The goal of brain-inspired computing is to deliver a scalable neural network substrate while approaching fundamental limits of time, space, and energy,” said IBM Fellow Dharmendra Modha, chief scientist, Brain-inspired Computing, IBM Research.

“The new milestone provides a palpable proof-of-concept that the efficiency of brain-inspired computing can be merged with the effectiveness of deep learning, paving the path towards a new generation of cognitive computing spanning mobile, cloud, and supercomputers.”

IBM Fellow Dharmendra Modha with the NS16e

IBM Fellow Dharmendra Modha with the NS16e

In March of this year, IBM demonstrated and delivered a scale-up neuromorphic system—the NS16e, which consists of a 16-chip array of TrueNorth processors—to the Department of Energy’s Lawrence Livermore National Laboratory (LLNL); this configuration is designed for efficiently running large-scale networks that do not fit on a single TrueNorth chip. The NS16e System interconnects TrueNorth chips via a built-in chip-to-chip message-passing interface that does not require additional circuitry or firmware, significantly reducing latency and energy of inter-chip communication.

LLNL is investigating classification and recognition uses of the system for embedded, low-power applications and for high-speed supercomputing applications. Specifically, LLNL is investigating how to use the system to detect cars in overhead imagery with context, clutter, and occlusion; to detect defects in additive manufacturing; and to supervise complex dynamical simulations, like physics problems in industrial design, to avoid failures.

Today, the TrueNorth development ecosystem includes not only the TrueNorth brain-inspired processor, the novel algorithm for training deep networks and the scaled-up NS16e System but also a simulator, a programming language, an integrated programming environment, a library of algorithms and applications, firmware, a teaching curriculum, single-chip boards, and scaled-out systems. The resulting ecosystem is in use by more than 130 users at more than 40 universities, government agencies, and national labs on five continents.

The ecosystem was originally developed under the auspices of Defense Advanced Research Projects Agency’s (DARPA) Systems of Neuromorphic Adaptive Plastic Scalable Electronics (SyNAPSE) program in collaboration with Cornell University. IBM has been holding a series of training events (at Telluride 2015, at IBM Almaden in August 2015, at Telluride 2016, and at IBM Research – Almaden in May 2016) to make the technology widely available. For more information and regular updates, Follow Modha’s blog.

More Cognitive Computing stories

Unlocking the Potential of Today’s Noisy Quantum Computers for OLED Applications

Scientists at Mitsubishi Chemical, a member of the IBM Quantum Hub at Keio University in Japan, reached out to our team about experimenting with new approaches to error mitigation and novel quantum algorithms to address these very challenges. In the new arXiv preprint, “Applications of Quantum Computing for Investigations of Electronic Transitions in Phenylsulfonyl-carbazole TADF Emitters,” we – along with collaborators at Keio University and JSR - describe quantum computations of the “excited states,” or high energy states, of industrial chemical compounds that could potentially be used in the fabrication of efficient organic light emitting diode (OLED) devices.

Continue reading

New Macromolecule Could Hold Key to Reversing Antibiotic Resistance

To address the challenge of antibiotic resistance, scientists from IBM and the Agency for Science, Technology and Research and the Singapore-MIT Alliance for Research and Technology have published new findings in Advanced Science, which unveil the effectiveness of a new polymer in the fight against resistant bacteria.

Continue reading

VLSI 2020: IBM Research Highlights Nanosheet, AI Processor and Photonics Advances

At VLSI 2020, IBM Research is spotlighting key developments for hybrid cloud infrastructure and AI, marked by improvements in performance, energy efficiency, area scaling, and new workloads.

Continue reading