Skip to main content

Robots are learning navigation skills by being trained to hunt prey

robot hunt prey 13900953  male lion panthera leo
Image used with permission by copyright holder
The newest king of the jungle? It’s not a big cat. It’s a robot. Scientists at the Institute of Neuroinformatics at the University of Zurich in Switzerland are teaching robots how to behave like predators and hunt their prey by way of a specialized camera and software that allows for deadly precision.

Don’t worry — the end goal isn’t actually predatory in nature. Rather, scientists say, they’re just trying to get robots to navigate their surroundings more accurately and efficiently, perhaps by finding a mark and following it around.

“Following [in large groups of self-driving cars or drones] is the obvious application, but one could imagine future luggage or shopping carts that follow you,” Tobi Delbruck, a professor at the Institute of Neuroinformatics, told Motherboard via email. “This way, the problem is less like a predator and its prey and more like herding, or a parent and child.”

But regardless of what you call it, the concept behind the new technology is decidedly, well, animalistic. The robot’s hardware is modeled after animal behavior, and depends largely on a silicon retina that behaves much like the human eye to quickly process visual data. A normal camera wouldn’t suffice in this situation because a slower frame rate doesn’t allow for the robot to see an actual movement path, especially if its “prey” (or whatever) is moving quickly.

The data obtained by way of the retina is then processed using a deep learning neural network that becomes “smarter” the more it’s used. Because in both the wild and in the lab, practice makes perfect.

So watch out, world. You may soon be followed around by robots who are watching you, quite literally, like a hawk. But whether or not that’s actually a bad thing has yet to be determined.

Editors' Recommendations

Lulu Chang
Former Digital Trends Contributor
Fascinated by the effects of technology on human interaction, Lulu believes that if her parents can use your new app…
Japanese researchers use deep learning A.I. to get driftwood robots moving
driftwood ai robots move mzmzmjaymw

walk

Did you ever make sculptures out of found objects like driftwood? Researchers at the University of Tokyo have taken this same idea and applied it to robots. In doing so, they’ve figured out a way to take everyday natural objects like pieces of wood and get deep reinforcement learning algorithms to figure out how to make them move. Using just a few basic servos, they’ve opened up a whole new way of building robots -- and it’s pretty darn awesome.

Read more
Digital Trends’ Tech For Change CES 2023 Awards
Digital Trends CES 2023 Tech For Change Award Winners Feature

CES is more than just a neon-drenched show-and-tell session for the world’s biggest tech manufacturers. More and more, it’s also a place where companies showcase innovations that could truly make the world a better place — and at CES 2023, this type of tech was on full display. We saw everything from accessibility-minded PS5 controllers to pedal-powered smart desks. But of all the amazing innovations on display this year, these three impressed us the most:

Samsung's Relumino Mode
Across the globe, roughly 300 million people suffer from moderate to severe vision loss, and generally speaking, most TVs don’t take that into account. So in an effort to make television more accessible and enjoyable for those millions of people suffering from impaired vision, Samsung is adding a new picture mode to many of its new TVs.
[CES 2023] Relumino Mode: Innovation for every need | Samsung
Relumino Mode, as it’s called, works by adding a bunch of different visual filters to the picture simultaneously. Outlines of people and objects on screen are highlighted, the contrast and brightness of the overall picture are cranked up, and extra sharpness is applied to everything. The resulting video would likely look strange to people with normal vision, but for folks with low vision, it should look clearer and closer to "normal" than it otherwise would.
Excitingly, since Relumino Mode is ultimately just a clever software trick, this technology could theoretically be pushed out via a software update and installed on millions of existing Samsung TVs -- not just new and recently purchased ones.

Read more
AI turned Breaking Bad into an anime — and it’s terrifying
Split image of Breaking Bad anime characters.

These days, it seems like there's nothing AI programs can't do. Thanks to advancements in artificial intelligence, deepfakes have done digital "face-offs" with Hollywood celebrities in films and TV shows, VFX artists can de-age actors almost instantly, and ChatGPT has learned how to write big-budget screenplays in the blink of an eye. Pretty soon, AI will probably decide who wins at the Oscars.

Within the past year, AI has also been used to generate beautiful works of art in seconds, creating a viral new trend and causing a boon for fan artists everywhere. TikTok user @cyborgism recently broke the internet by posting a clip featuring many AI-generated pictures of Breaking Bad. The theme here is that the characters are depicted as anime characters straight out of the 1980s, and the result is concerning to say the least. Depending on your viewpoint, Breaking Bad AI (my unofficial name for it) shows how technology can either threaten the integrity of original works of art or nurture artistic expression.
What if AI created Breaking Bad as a 1980s anime?
Playing over Metro Boomin's rap remix of the famous "I am the one who knocks" monologue, the video features images of the cast that range from shockingly realistic to full-on exaggerated. The clip currently has over 65,000 likes on TikTok alone, and many other users have shared their thoughts on the art. One user wrote, "Regardless of the repercussions on the entertainment industry, I can't wait for AI to be advanced enough to animate the whole show like this."

Read more