Research into more energy efficient data-processing makes computers resemble human brains

0
784

THE RAPID progression of technology has led to a huge increase in energy usage to process the massive troves of data generated by devices.

However, researchers in the Cockrell School of Engineering at The University of Texas at Austin have found a way to make the new generation of smart computers more energy efficient.

Traditionally, silicon chips have formed the building blocks of the infrastructure that powers computers.

This research uses magnetic components instead of silicon and discovers new information about how the physics of the magnetic components can cut energy costs and requirements of training algorithms — neural networks that can think like humans and do things like recognise images and patterns.

Jean Anne Incorvia, an assistant professor in the Cockrell School’s Department of Electrical and Computer Engineering, commented on the research: “Right now, the methods for training your neural networks are very energy-intensive.

“What our work can do is help reduce the training effort and energy costs.”

Assistant Professor Incorvia led the study with first author and second-year graduate student Can Cui.

Asst Professor Incorvia and Ms Cui discovered that spacing magnetic nanowires, acting as artificial neurons, in certain ways naturally increases the ability for the artificial neurons to compete against each other, with the most activated ones winning out.

Achieving this effect, known as “lateral inhibition,” traditionally requires extra circuitry within computers, which increases costs and takes more energy and space.

lateral inhibition magnetic processing

(Image by University of Texas at Austin)

Asst Professor Incorvia said their method provides an energy reduction of 20 to 30 times the amount used by a standard back-propagation algorithm when performing the same learning tasks.

The same way human brains contain neurons, new-era computers have artificial versions of these integral nerve cells.

Lateral inhibition occurs when the neurons firing the fastest are able to prevent slower neurons from firing. In computing, this cuts down on energy use in processing data.

Asst Professor Incorvia explains that the way computers operate is fundamentally changing.

A major trend is the concept of neuromorphic computing, which is essentially designing computers to think like human brains.

Instead of processing tasks one at a time, these smarter devices are meant to analyse huge amounts of data simultaneously.

These innovations have powered the revolution in machine learning and artificial intelligence that has dominated the technology landscape in recent years.

This research focused on interactions between two magnetic neurons and initial results on interactions of multiple neurons.

The next step involves applying the findings to larger sets of multiple neurons as well as experimental verification of their findings.

The research was funded by a National Science Foundation CAREER Award and Sandia National Laboratories, with resources from UT’s Texas Advanced Computing Center.

The researchers’ findings were have been published in IOP Nanotechnology