Researchers Improve Robot Gesture Recognition With Computer Vision

thumbnail
Researchers Improve Robot Gesture Recognition With Computer Vision

The high accuracy is due to the fact that the visual and somatosensory information can interact and complement each other at an early stage before carrying out complex interpretation.

Written By
thumbnail
David Curry
David Curry
Aug 27, 2020

Researchers from Nanyang Technological University have developed an artificial intelligence system able to recognize gestures more accurately, through the use of wearable strain sensors.

The strain sensor far exceeds previous wearable devices, such as fitness wristwatches or gloves, by being much thinner and flexible. NTU scientists are able to receive data on skin senses and vision, which they then input into a neural network for processing.

SEE ALSO: Researchers Augment Robotic Prosthetics With AI and Computer Vision

“Our data fusion architecture has its own unique bioinspired features which include a man-made system resembling the somatosensory-visual fusion hierarchy in the brain. We believe such features make our architecture unique to existing approaches,” said NTU professor Chen Xiaodong.

“Compared to rigid wearable sensors, our innovation uses stretchable strain sensors that comfortably attaches onto the human skin. This allows for high-quality signal acquisition, which is vital to high-precision recognition tasks.”

Amazon and Tesla both suffered from setbacks in deploying robots to perform human tasks. In Tesla’s case, it found that robots were unable to find the right screw in a box and struggled to lift certain objects.

With better computer vision and precision, those issues may be fixable.

NTU’s AI system was highly accurate even in poor conditions, achieving a 96.7 percent recognition accuracy rating in the dark. The team intends to build a VR and AR system using the AI, which will be used as a show-piece for prospective clients.

“The secret behind the high accuracy in our architecture lies in the fact that the visual and somatosensory information can interact and complement each other at an early stage before carrying out complex interpretation. As a result, the system can rationally collect coherent information with less redundant data and less perceptual ambiguity, resulting in better accuracy,” said Dr. Wang Ming of NTU.

thumbnail
David Curry

David is a technology writer with several years experience covering all aspects of IoT, from technology to networks to security.

Recommended for you...

Agentic AI and the Death of SaaS
Domain-Specific LLMs: How to Make AI Useful for Your Business
Hardik Parikh
Mar 11, 2026
Engineering the Agentic Enterprise: Building Smarter, Adaptive, Autonomous Systems
Varun Goswami
Mar 10, 2026
The AI That Actually Scales Is Boring. That’s the Point.
Jared Coyle
Mar 9, 2026

Featured Resources from Cloud Data Insights

Agentic AI and the Death of SaaS
The Business Case for a Unified Semantic Layer
Alex Merced
Mar 12, 2026
Domain-Specific LLMs: How to Make AI Useful for Your Business
Hardik Parikh
Mar 11, 2026
Engineering the Agentic Enterprise: Building Smarter, Adaptive, Autonomous Systems
Varun Goswami
Mar 10, 2026
RT Insights Logo

Analysis and market insights on real-time analytics including Big Data, the IoT, and cognitive computing. Business use cases and technologies are discussed.

Property of TechnologyAdvice. © 2026 TechnologyAdvice. All Rights Reserved

Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.