Robotics: Gesturing for control

Robotics: Gesturing for control
Credit: DIUSGOVUK

New intelligent algorithms could help robots to quickly recognize and respond to human gestures. Researchers at A*STAR Institute for Infocomm Research in Singapore have created a computer program which recognises human gestures quickly and accurately, and requires very little training.

Many works of science fiction have imagined robots that could interact directly with people to provide entertainment, services or even . Robotics is now at a stage where some of these ideas can be realized, but it remains difficult to make robots easy to operate.

One option is to train robots to recognize and respond to human gestures. In practice, however, this is difficult because a simple gesture such as waving a hand may appear very different between different people. must develop intelligent that can be ‘trained’ to identify general patterns of motion and relate them correctly to individual commands.

Now, Rui Yan and co-workers at the A*STAR Institute for Infocomm Research in Singapore have adapted a cognitive memory model called a localist attractor network (LAN) to develop a new system that recognize gestures quickly and accurately, and requires very little training.

“Since many social robots will be operated by non-expert users, it is essential for them to be equipped with natural interfaces for interaction with humans,” says Yan. “Gestures are an obvious, natural means of human communication. Our LAN gesture recognition system only requires a small amount of training data, and avoids tedious training processes.”

Yan and co-workers tested their software by integrating it with ShapeTape, a special jacket that uses fibre optics and inertial sensors to monitor the bending and twisting of hands and arms. They programmed the ShapeTape to provide data 80 times per second on the three-dimensional orientation of shoulders, elbows and wrists, and applied velocity thresholds to detect when gestures were starting.

In tests, five different users wore the ShapeTape jacket and used it to control a virtual through simple arm motions that represented commands such as forward, backwards, faster or slower. The researchers found that 99.15% of gestures were correctly translated by their system. It is also easy to add new commands, by demonstrating a new control gesture just a few times.

The next step in improving the gesture recognition system is to allow humans to control robots without the need to wear any special devices. Yan and co-workers are tackling this problem by replacing the ShapeTape jacket with motion-sensitive cameras.

“Currently we are building a new system by incorporating our method with a Microsoft Kinect camera,” says Yan. “We will implement the proposed system on an autonomous robot to test its usability in the context of a realistic service task, such as cleaning!”

More information: Yan, R., et al. Gesture recognition based on localist attractor networks with application to robot control. IEEE Computational Intelligence Magazine 7, 64–74 (2012).

Citation: Robotics: Gesturing for control (2012, May 24) retrieved 28 March 2024 from https://phys.org/news/2012-05-robotics-gesturing.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Guiding robot planes with hand gestures

0 shares

Feedback to editors