Kinect-based sign language interpretation

Kinect-based sign language interpretation

What started out as a game has become so much more.

More Vision Articles

How accurate is the Kinect?

Enhanced 3D sensing capabilities in Microsoft Kinect

NASA's JPL researchers operate a robotic arm with Kinect 2 and Oculus Rift

Kinect -based sign language interpretation

Kinect helps researchers steer roaches to explore and map disaster sites

Okay, so by now you are well aware of some of the more unique applications of the Microsoft Kinect. If you aren’t, here are just a few:

The latest innovative use of the Kinect involves researchers from Microsoft Research Asia using the device’s body-tracking abilities to perform sign-language recognition. By tracking the position of the user’s hands, many signs can be detected instantly by using the group’s 3D trajectory matching algorithm. Check out the video here:

The system consists of two modes, the first of which is translation mode, which translates sign language into text or speech. This includes isolated word recognition and sentence recognition. The raising and putting down of hands is defined as the “start” and the “end” of each sign language word to be recognized. The system gives the recognized rank-5 candidate works in the bottom region of the interface, and if the rank-1 word is not correct, the signer can adjust the results by manual interaction.

The second mode is communication mode, in which a non-hearing impaired person can communicate with the signer through an avatar. The impaired person can make an immediate response through signing, and the system translates the answer into text.

Once again, an extremely cool and practical application of the Microsoft Kinect that makes you wonder if the developers of the initial model had any inclination that the product would branch out as it has. Another thing that may make you wonder—I am, anyway—is what courtroom sign language interpreters are thinking of this technology. Probably akin to the way that a farmhand might think of this.


Upcoming Webcasts

An update on USB3 Vision

This presentation,  sponsored by Point Grey, Matrox Imaging, and Ximea GmbH., will cover system design issues related to USB3 Vision, and will integrate real-world examples to facilitate under...

Choosing and testing smart cameras for your application

This webcast will guide viewers in the selection of smart camera hardware for applications such as inspection and verification, covering various functions and common features.

March 25, 2014

Food inspection from farm to table

The webcast will use actual examples--including long-term academic research and industrial applications--to illustrate and to elaborate the future perspective of R&D in this field. Interactive ...
March 11, 2014

Automotive parts inspection evolves with vision

This webcast, sponsored by Cognex Corporation, Matrox Imaging, Omron Electronics, and Point Grey will identify the technologies and setup requirements, including the system specifications nece...
February 18, 2014


Click here to view archived Vision Systems Design articles