Many manufacturing jobs require a physical presence to operate machinery. But what if such jobs could be done remotely? This week researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) presented a virtual-reality (VR) system that lets you teleoperate a robot using an Oculus Rift headset.
The system embeds the user in a VR control room with multiple sensor displays, making it feel like they are inside the robot’s head. By using gestures, users can match their movements to the robot’s to complete various tasks.
“A system like this could eventually help humans supervise robots from a distance,” says CSAIL postdoctoral associate Jeffrey Lipton, who was lead author on a related paper about the system. “By teleoperating robots from home, blue-collar workers would be able to tele-commute and benefit from the IT revolution just as white-collars workers do now.”
The researchers even imagine that such a system could help employ increasing numbers of jobless video-gamers by “game-ifying” manufacturing positions.
The team demonstrated their VC control approach with the Baxter humanoid robot from Rethink Robotics, but said that the approach can work on other robot platforms and is also compatible with the HTC Vive headset.
Lipton co-wrote the paper with CSAIL director Daniela Rus and researcher Aidan Fay. They presented the paper this week at the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) in Vancouver.
How it works
There have traditionally been two main approaches to using VR for teleoperation.
In a “direct” model, the user’s vision is directly coupled to the robot’s state. With these systems, a delayed signal could lead to nausea and headaches, and the user’s viewpoint is limited to one perspective.
In the “cyber-physical” model, the user is separate from the robot. The user interacts with a virtual copy of the robot and the environment. This requires much more data, and specialized spaces.
The CSAIL team’s system is halfway between these two methods. It solves the delay problem, since the user is constantly receiving visual feedback from the virtual world. It also solves the the cyber-physical issue of being distinct from the robot: once a user puts on the headset and logs into the system, they will feel as if they are inside Baxter’s head.
The system mimics the “homunculus model of mind” — the idea that there’s a small human inside our brains controlling our actions, viewing the images we see and understanding them for us. While it’s a peculiar idea for humans, for robots it fits: “inside” the robot is a human in a control room, seeing through its eyes and controlling its actions.
Using Oculus’ controllers, users can interact with controls that appear in the virtual space to open and close the hand grippers to pick up, move, and retrieve items. A user can plan movements based on the distance between the arm’s location marker and their hand while looking at the live display of the arm.
To make these movements possible, the human’s space is mapped into the virtual space, and the virtual space is then mapped into the robot space to provide a sense of co-location.
The system is also more flexible compared to previous systems that require many resources. Other systems might extract 2-D information from each camera, build out a full 3-D model of the environment, and then process and redisplay the data.
In contrast, the CSAIL team’s approach bypasses all of that by taking the 2-D images that are displayed to each eye. (The human brain does the rest by automatically inferring the 3-D information.)
To test the system, the team first teleoperated Baxter to do simple tasks like picking up screws or stapling wires. They then had the test users teleoperate the robot to pick up and stack blocks.
Users successfully completed the tasks at a much higher rate compared to the “direct” model. Unsurprisingly, users with gaming experience had much more ease with the system.
Tested against state-of-the-art systems, CSAIL’s system was better at grasping objects 95 percent of the time and 57 percent faster at doing tasks. The team also showed that the system could pilot the robot from hundreds of miles away, testing it on a hotel’s wireless network in Washington, DC to control Baxter at MIT.
“This contribution represents a major milestone in the effort to connect the user with the robot’s space in an intuitive, natural, and effective manner.” says Oussama Khatib, a computer science professor at Stanford University who was not involved in the paper.
The team eventually wants to focus on making the system more scalable, with many users and different types of robots that can be compatible with current automation technologies.
Learn more: Teleoperating robots with virtual reality: Making it easier for factory workers to telecommute
The Latest on: Teleoperating robots
[google_news title=”” keyword=”teleoperating robots” num_posts=”10″ blurb_length=”0″ show_thumb=”left”]
- PCB Stepper Motor Micro Robotson March 21, 2023 at 5:01 pm
These are basically a stepper motor that’s been unrolled and made flat — traces on the PCB act as the coils and tiny magnetic “robots” act as the rotor. If you want to try this concept ...
- Best robot vacuum deals in March 2023on March 21, 2023 at 1:59 pm
A robot vacuum can set you back a few hundred dollars. Fortunately, the best robot vacuum deals can help lower the cost of your robotic assistant. Currently, we're seeing some great sales on some ...
- Robots are replacing security guards. Should we give them guns?on March 19, 2023 at 10:32 am
However, I'm not so sure these bots can be reliable. Let's see how the robot security experiments are turning into reality. These robot guards are being used mainly in office buildings and have ...
- 4 Robots That Look Like Humanson March 16, 2023 at 6:06 pm
Robots that look like humans can walk, talk and even appear to show emotions. Meet the most human-like robots that have come creepily close to crossing the uncanny valley. A humanoid robot named ...
- Robots Are Folding Laundry, But They Suck At Iton March 16, 2023 at 5:00 pm
Robots are used in all sorts of industries on a wide variety of tasks. Typically, it’s because they’re far faster, more accurate, and more capable than we are. Expert humans could not compete ...
- Save up to $500 on robot vacuums and get a start on your spring cleaningon March 16, 2023 at 2:00 am
Find the robot vac that's right for your home. Credit: iRobot Deal pricing and availability subject to change after time of publication. Learn more about how we select deals. Only three things are ...
- Robots Showcase Electrical Engineering Innovationson March 15, 2023 at 5:00 pm
The subject of robotics is normally enough to raise eyebrows in any story or setting, but when it is brought to life with a live robot appearance, the story is brought to life. That was the case at ...
- Robots may improve mental health at workon March 15, 2023 at 4:00 am
Robots can enhance mental health at work, but they must first appear appropriate, according to studies. The study found that the appearance of the robot has a significant impact on how effective ...
- Best robot vacuum deals of 2023on March 15, 2023 at 3:38 am
Robot vacuums are a welcome addition to just about any home -- these automated workhorses clean your floors so you don't have to. They can free up time for you to focus more on the things you'd ...
- Best robot games on PC 2023on March 14, 2023 at 10:34 am
What are the best robot games on PC? Strapping into an overpowered robot filled to the brim with rockets and lasers is an old but popular power fantasy. Classic shows from the ‘80s and ‘90s ...
via Google News and Bing News