Skip to main content

Researchers put A.I. inside a camera lens to compute ‘at the speed of light’

Stanford

The camera is the eye for many automated devices and the computer is the brain — but researchers at Stanford University recently combined the two in an attempt to make smart cameras more compact. A team of graduate students recently created an artificially intelligent camera that doesn’t need a large, separate computer to process all the data — because it’s built into the optics itself.

Current object recognition technology uses A.I. on a separate computer to run the images or footage through algorithms to identify objects. As the Stanford researchers explain, driverless cars have a large computer in the trunk in order to recognize when a pedestrian steps out in front of the car’s path. Those computers are big, require lots of energy and are often slow.

The team of Stanford researchers behind the AI camera. Stanford

The researchers instead worked to build the A.I. directly into the camera, both to create smaller systems as well as faster ones. The camera has two layers — the first is called an optical computer. As light passes through the camera, the built-in computer pre-processes that data. The optical computer filters out unnecessary data, reducing the number of necessary calculations. The second layer is a traditional computer and imaging sensor that handles the remaining calculations. The science behind the A.I. camera is complex, but the result could lead to significant advances for devices with built-in cameras like self-driving cars and drones.

Recommended Videos

By processing the scene as the light hits the camera instead of after the fact, the researchers said they were able to create a camera that didn’t require input power, reducing the typically intense computing power required for object recognition A.I. “We’ve outsourced some of the math of artificial intelligence into the optics,” graduate student Julie Chang said.

Please enable Javascript to view this content

Gordon Wetzstein, the assistant professor that led the research, says that moving the computing to the optics allows for much faster data processing, along with reducing the computing required. “Millions of calculations are circumvented and it all happens at the speed of light,” he said.

While one of the goals is to reduce the size of different devices, the research hasn’t quite reached that point yet. The camera takes up a lab bench, but the researchers suggest that further research will help the concept to decrease in size, eventually making the AI camera small enough for more portable devices. The team suggests several different potential uses, from autonomous cameras and drones to handheld medical imaging. The processing at the speed of light could be a big advantage in self-driving cars that rely on cameras to see and avoid potential collisions.

The device’s size suggests the technology needs some time to mature before actually being integrated into actual products, but the research takes a significant step forward for artificially intelligent cameras. The team recently published the full research in Nature’s Scientific Reports.

Hillary K. Grigonis
Hillary never planned on becoming a photographer—and then she was handed a camera at her first writing job and she's been…
Algorithmic architecture: Should we let A.I. design buildings for us?
Generated Venice cities

Designs iterate over time. Architecture designed and built in 1921 won’t look the same as a building from 1971 or from 2021. Trends change, materials evolve, and issues like sustainability gain importance, among other factors. But what if this evolution wasn’t just about the types of buildings architects design, but was, in fact, key to how they design? That’s the promise of evolutionary algorithms as a design tool.

While designers have long since used tools like Computer Aided Design (CAD) to help conceptualize projects, proponents of generative design want to go several steps further. They want to use algorithms that mimic evolutionary processes inside a computer to help design buildings from the ground up. And, at least when it comes to houses, the results are pretty darn interesting.
Generative design
Celestino Soddu has been working with evolutionary algorithms for longer than most people working today have been using computers. A contemporary Italian architect and designer now in his mid-70s, Soddu became interested in the technology’s potential impact on design back in the days of the Apple II. What interested him was the potential for endlessly riffing on a theme. Or as Soddu, who is also professor of generative design at the Polytechnic University of Milan in Italy, told Digital Trends, he liked the idea of “opening the door to endless variation.”

Read more
Emotion-sensing A.I. is here, and it could be in your next job interview
man speaking into phone

I vividly remember witnessing speech recognition technology in action for the first time. It was in the mid-1990s on a Macintosh computer in my grade school classroom. The science fiction writer Arthur C. Clarke once wrote that “any sufficiently advanced technology is indistinguishable from magic” -- and this was magical all right, seeing spoken words appearing on the screen without anyone having to physically hammer them out on a keyboard.

Jump forward another couple of decades, and now a large (and rapidly growing) number of our devices feature A.I. assistants like Apple’s Siri or Amazon’s Alexa. These tools, built using the latest artificial intelligence technology, aren’t simply able to transcribe words -- they are able to make sense of their contents to carry out actions.

Read more
Language supermodel: How GPT-3 is quietly ushering in the A.I. revolution
Profile of head on computer chip artificial intelligence.

OpenAI’s GPT-2 text-generating algorithm was once considered too dangerous to release. Then it got released -- and the world kept on turning.

In retrospect, the comparatively small GPT-2 language model (a puny 1.5 billion parameters) looks paltry next to its sequel, GPT-3, which boasts a massive 175 billion parameters, was trained on 45 TB of text data, and cost a reported $12 million (at least) to build.

Read more