Electrical engineers at the University of California San Diego have developed a faster collision detection algorithm that uses machine learning to help robots avoid moving objects and weave through complex, rapidly changing environments in real time. The algorithm, dubbed “Fastron,” runs up to 8 times faster than existing collision detection algorithms.
A team of engineers, led by Michael Yip, a professor of electrical and computer engineering and member of the Contextual Robotics Institute at UC San Diego, will present the new algorithm at the first annual Conference on Robot Learning Nov. 13 to 15 at Google headquarters in Mountain View, Calif. The conference brings the top machine learning scientists to an invitation-only event. Yip’s team will deliver one of the long talks during the 3-day conference.
The team envisions that Fastron will be broadly useful for robots that operate in human environments where they must be able to work with moving objects and people fluidly. One application they are exploring in particular is robot-assisted surgeries using the da Vinci Surgical System, in which a robotic arm would autonomously perform assistive tasks (suction, irrigation or pulling tissue back) without getting in the way of the surgeon-controlled arms or the patient’s organs.
“This algorithm could help a robot assistant cooperate in surgery in a safe way,” Yip said.
The team also envisions that Fastron can be used for robots that work at home for assisted living applications, as well as for computer graphics for the gaming and movie industry, where collision checking is often a bottleneck for most algorithms.
A problem with existing collision detection algorithms is that they are very computation-heavy. They spend a lot of time specifying all the points in a given space—the specific 3D geometries of the robot and obstacles—and performing collision checks on every single point to determine whether two bodies are intersecting at any given time. The computation gets even more demanding when obstacles are moving.
To lighten the computational load, Yip and his team in the Advanced Robotics and Controls Lab (ARClab) at UC San Diego developed a minimalistic approach to collision detection. The result was Fastron, an algorithm that uses machine learning strategies—which are traditionally used to classify objects—to classify collisions versus non-collisions in dynamic environments. “We actually don’t need to know all the specific geometries and points. All we need to know is whether the robot’s current position is in collision or not,” said Nikhil Das, an electrical engineering Ph.D. student in Yip’s group and the study’s first author.
Fastron simulation: the autonomous arm (blue arm) reaches the target configuration (wireframe arm) while avoiding the motions of a human-controlled arm (red arm). Image courtesy of ARClab at UC San Diego.
The Fastron algorithm
The name Fastron comes from combining Fast and Perceptron, which is a machine learning technique for performing classification. An important feature of Fastron is that it updates its classification boundaries very quickly to accommodate for moving scenes, something that has been challenging for the machine learning community in general to do.
Fastron’s active learning strategy works using a feedback loop. It starts out by creating a model of the robot’s configuration space, or C-space, which is the space showing all possible positions the robot can attain. Fastron models the C-space using just a sparse set of points, consisting of a small number of so-called collision points and collision-free points. The algorithm then defines a classification boundary between the collision and collision-free points—this boundary is essentially a rough outline of where the abstract obstacles are in the C-space. As obstacles move, the classification boundary changes. Rather than performing collision checks on each point in the C-space, as is done with other algorithms, Fastron intelligently selects checks near the boundaries. Once it classifies the collisions and non-collisions, the algorithm updates its classifier and then continues the cycle.
Because Fastron’s models are more simplistic, the researchers set its collision checks to be more conservative. Since just a few points represent the entire space, Das explained, it’s not always certain what’s happening in the space between two points, so the team developed the algorithm to predict a collision in that space. “We leaned toward making a risk-averse model and essentially padded the workspace obstacles,” Das said. This ensures that the robot can be tuned to be more conservative in sensitive environments like surgery, or for robots that work at home for assisted living.
The team has so far demonstrated the algorithm in computer simulations on robots and obstacles in simulation. Moving forward, the team is working to further improve the speed and accuracy of Fastron. Their goal is to implement Fastron in a robotic surgery and a homecare robot setting.
Learn more: Speedy collision detector could make robots better human assistants
The Latest on: Robots interacting with humans
[google_news title=”” keyword=”robots interacting with humans” num_posts=”10″ blurb_length=”0″ show_thumb=”left”]- Cobot proudly reveals innovative non-humanoid robots, announces $100M Series B.on April 30, 2024 at 9:01 am
The burgeoning field of robotics, traditionally dominated by applications in heavy industry and scientific research, is experiencing a significant shift as startups like Sunnyvale, California’s ...
- WATCH: Chery reveals human-like robot driven by AIon April 30, 2024 at 5:44 am
Chery’s robot will be able to answer questions, respond to instructions and assist people in a natural and human-like way.
- Sophia the Robot Falls in Greeceon April 30, 2024 at 12:00 am
Sophia, the globally renowned humanoid AI robot, suffered a malfunction in Thessaloniki, Greece when she fell down a step at an exhibition center. The robot that is currently touring Greece fell at ...
- Astribot S1 AI Humanoid robot unveiled demonstrating its agility, dexterity and accuracyon April 27, 2024 at 12:55 am
This week the Astribot S1 humanoid robot was unveiled in Shenzhen, China, marking another significant leap forward in autonomous robotics ...
- Robot uses AI to mimic human facial expressionson April 25, 2024 at 6:24 am
A robot capable of mimicking human facial expressions has been developed by engineers at Columbia University in the US.
- This robot can predict a smile before it happenson April 24, 2024 at 11:17 am
STORY: This AI-integrated robotic face can predict a smile before it happens.It's called Emo and it can anticipate and mimic human facial expressions.Engineers at Columbia University’s Creative ...
- People, not design features, make a robot socialon April 22, 2024 at 8:24 am
It takes a village to nurture social robots. Researchers who develop social robots—ones that people interact with—focus too much on design features and not enough on sociological factors, like ...
- How the Optimus Tesla Bot humanoid robot is madeon April 22, 2024 at 5:21 am
Tesla, renowned for its groundbreaking innovations in electric vehicles and sustainable energy, has moving into the realm of humanoid robotics over the ...
- Humanoid Robot with Lost in Space Vibes Looks Ready to Make Humans Its Slaveson April 21, 2024 at 11:44 pm
American robotics company Boston Dynamics released a few details on the newest version of the Atlas humanoid robot, one that is electrically powered ...
- LocoMan: Scientists unveil new quadruped robot with hand-like gripperson April 20, 2024 at 8:33 am
Researchers have developed a LocoMan system that equips quadrupedal robots with dexterous gripping arms for precision interactions.
via Google News and Bing News