Dual arm robot holding crisp
CREDIT: Yijiong Lin
An innovative bimanual robot displays tactile sensitivity close to human-level dexterity using AI to inform its actions.
The new Bi-Touch system, designed by scientists at the University of Bristol and based at the Bristol Robotics Laboratory, allows robots to carry out manual tasks by sensing what to do from a digital helper.
The findings, published in IEEE Robotics and Automation Letters, show how an AI agent interprets its environment through tactile and proprioceptive feedback, and then control the robots’ behaviours, enabling precise sensing, gentle interaction, and effective object manipulation to accomplish robotic tasks.
This development could revolutionise industries such as fruit picking, domestic service, and eventually recreate touch in artificial limbs.
Lead author Yijiong Lin from the Faculty of Engineering, explained: “With our Bi-Touch system, we can easily train AI agents in a virtual world within a couple of hours to achieve bimanual tasks that are tailored towards the touch. And more importantly, we can directly apply these agents from the virtual world to the real world without further training.
“The tactile bimanual agent can solve tasks even under unexpected perturbations and manipulate delicate objects in a gentle way.”
Bimanual manipulation with tactile feedback will be key to human-level robot dexterity. However, this topic is less explored than single-arm settings, partly due to the availability of suitable hardware along with the complexity of designing effective controllers for tasks with relatively large state-action spaces. The team were able to develop a tactile dual-arm robotic system using recent advances in AI and robotic tactile sensing.
The researchers built up a virtual world (simulation) that contained two robot arms equipped with tactile sensors. They then design reward functions and a goal-update mechanism that could encourage the robot agents to learn to achieve the bimanual tasks and developed a real-world tactile dual-arm robot system to which they could directly apply the agent.
The robot learns bimanual skills through Deep Reinforcement Learning (Deep-RL), one of the most advanced techniques in the field of robot learning. It is designed to teach robots to do things by letting them learn from trial and error akin to training a dog with rewards and punishments.
For robotic manipulation, the robot learns to make decisions by attempting various behaviours to achieve designated tasks, for example, lifting up objects without dropping or breaking them. When it succeeds, it gets a reward, and when it fails, it learns what not to do. With time, it figures out the best ways to grab things using these rewards and punishments. The AI agent is visually blind relying only on proprioceptive feedback – a body’s ability to sense movement, action and location and tactile feedback.
They were able to successfully enable to the dual arm robot to successfully safely lift items as fragile as a single Pringle crisp.
Co-author Professor Nathan Lepora added: “Our Bi-Touch system showcases a promising approach with affordable software and hardware for learning bimanual behaviours with touch in simulation, which can be directly applied to the real world. Our developed tactile dual-arm robot simulation allows further research on more different tasks as the code will be open-source, which is ideal for developing other downstream tasks.”
Yijiong concluded: “Our Bi-Touch system allows a tactile dual-arm robot to learn sorely from simulation, and to achieve various manipulation tasks in a gentle way in the real world.
“And now we can easily train AI agents in a virtual world within a couple of hours to achieve bimanual tasks that are tailored towards the touch.”
Original Article: New dual-arm robot achieves bimanual tasks by learning from simulation
More from: University of Bristol
The Latest Updates from Bing News
Go deeper with Bing News on:
Bimanual robot
- Rainbow Robotics begins pre-orders of Bimanual Mobile Manipulator RB-Y1, the world's first research platform for AI experts for $80,000 USD
DAEJEON, South Korea, May 8, 2024 /PRNewswire/ -- Rainbow Robotics(CEO Jungho Lee), a robot platform specialized company, will begin pre-orders for the Bimanual Mobile Manipulator RB-Y1 from May 8.
- Could a robot care for grandma?
As part of the photographer’s project, elderly people were asked to imagine how they would interact with a robot. A resident at Maison Ferrari, a retirement home in Clamart, France, volunteered ...
- Rainbow Robotics participates in ICRA 2024 in Yokohama, Japan: Bimanual Mobile Manipulator 'RB-Y1' debuts overseas
That includes Isaac Manipulator, a collection of foundational models designed to create workflows for robot arms. The offering launched at GTC back in March, with some of the biggest names in ...
- Best robot mops in 2024
The best robot mops deep clean your floors without leaving streaks. These handy cleaners complement the best robot vacuums, coming in once all debris has been removed from the floor. Robot mops ...
- The Best Robot Vacuums for 2024
Robot vacuums with around 2,500Pa can still lift pet hair, but might need more passes to get it all. Dyson, which uses the less common metric of air watts, says its new 360 Vis Nav robot vacuum ...
Go deeper with Bing News on:
Deep Reinforcement Learning
- Collaborative hunting in artificial agents with deep reinforcement learning
Collaborative hunting, characterized by the division of roles among predators, has emerged within a group of artificial agents through deep reinforcement learning.
- 'People don't realize the innovation that happens here' AI at work in Houston's oil and gas industry
A recent report predicts the value of AI in the oil and gas sector will rise to nearly $6 billion by 2028, with Houston being a hub for innovation in the development of AI in the industry.
- Robot dog masters walking, trotting, pronking in a major milestone
Researchers at the Swiss Federal Institute of Technology Lausanne (EPFL) have trained a quadrupedal robot using machine learning, enabling it to adeptly navigate without falling by seamlessly ...
- Robot Masters Terrain with Animal-Like Gait Transitions
Researchers leveraged deep reinforcement learning (DRL) to enable a robot to adaptively switch gaits, mimicking animal movements like trotting and pronking, to traverse complex terrains effectively.
- Deep Reinforcement Learning: Applications and Future Directions
Deep reinforcement learning (DRL) integrates the reinforcement learning's decision-making ability and deep learning's feature representation ability to attain robust end-to-end learning control ...