Skip to main content

This breakthrough memory technology could make AI 1,000 times more efficient

We all know AI has a power problem. On the whole, global AI usage already drew as much energy as the entire nation of Cyprus did in 2021.

But engineering researchers at the University of Minnesota Twin Cities have developed and demonstrated a new computer memory design that could drastically reduce the amount of energy AI systems consume to help temper this problem. Their research was recently published in the journal Nature journal Unconventional Computing.

Recommended Videos

Most modern computing systems are built on what is known as the Von Neumann architecture, where the logic and memory subsystems are separated. During normal operations, data is shuttled back and forth between the memory modules and processors. This is the basic foundation of modern computers operate.

However, as processing speeds rapidly outpace I/O technology, this data transfer becomes a bottleneck both in terms of processing speed (also known as the memory wall problem) and power consumption. As the researchers pointed out, just shuffling the data back and forth consumers as much as 200 times the amount of power that the computations themselves do.

Developers have sought to work around this issue by bringing the logic and memory physically closer together with “near-memory” and “in-memory” computing designs. Near-memory systems stack the logic and memory on top of one another in a 3D array, they’re layered PB&J-style, while in-memory systems intersperse clusters of logic throughout the memory on a single chip, more like a peanut butter and banana sandwich.

The Twin Cities research team’s solution is a novel, fully digital, in-memory design, dubbed computational random-access memory (CRAM), wherein, “logic is performed natively by the memory cells; the data for logic operations never has to leave the memory,” per the researchers. The team achieved this by integrating a reconfigurable spintronic compute substrate directly into the memory cell, an advance that the researchers found could reduce an AI operation’s energy consumption by an “order of 1,000x over a state-of-the-art solution.”

And that 1,000x improvement could just be the baseline. The research team tested CRAM on an MNIST handwritten digit classifier task and found it to be “2,500× and 1,700× less in energy and time, respectively, compared to a near-memory processing system at the 16 nm technology node.”

The emerging AI industry is already facing significant resource issues. The ever faster, ever more powerful and capable GPUs that underpin AI software are immensely energy hungry. NVIDIA‘s newest top-of-the-line Blackwell B200 consumes up to 1,200W, for example, and generates so much waste heat that it requires liquid cooling, another resource-intensive operation.

With hyperscalers like Google, Amazon, and Microsoft all scrambling to build out the physical infrastructure necessary to power the oncoming AI revolution — ie gigawatt-sized data centers, some with their own attached nuclear power plants — creating more energy-efficient compute and memory resources will become increasingly critical to the long-term viability of AI technology.

Andrew Tarantola
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
AI is making a long-running scam even more effective
An elderly person holding a phone.

You’ve no doubt heard of the scam where the perpetrator calls up an elderly person and pretends to be their grandchild or some other close relative. The usual routine is to act in a distressed state, pretend they’re in a sticky situation, and ask for an urgent cash transfer to resolve the situation. While many grandparents will realize the voice isn’t that of their grandchild and hang up, others won’t notice and, only too keen to help their anxious relative, go ahead and send money to the caller’s account.

A Washington Post report on Sunday reveals that some scammers have taken the con to a whole new level by deploying AI technology capable of cloning voices, making it even more likely that the target will fall for the ruse.

Read more
Are gaming PCs more expensive today? Here’s what $1,000 bought you 10 years ago
A close-up image of Nvidia's RTX 3080 Ti graphics card.

Say it with me: "Building a gaming PC is getting more expensive." Price is top of mind when building a gaming PC in 2022, and why wouldn't it be? Today, the best graphics cards will cost you well over $1,000, DDR5 is ungodly expensive, and CPU prices are double or even triple what they were a decade ago.

It's easy to add up the numbers and come to a conclusion, but that ignores game optimizations, falling prices of other components, and the various upscaling tools players have to squeeze extra performance out of their PCs. Instead of adding up what you could spend on a gaming PC, I added up what you would spend.

Read more
Brain-like chips may make PCs 1 million times more efficient
New brain-like computer chip.

Future computers may be modeled to simulate and resemble the human brain, and however futuristic that might sound, this technology might be closer than we think.

A new scientific report shows that a new material could potentially help in the production of PC chips that will be 1 million times more efficient than current technologies, and this could happen before the end of this decade.

Read more