Saturday 25 May 2013

Skinput Technology

Skinput is an input technology  that uses bio-acoustic sensing to localize finger taps on the skin. When augmented with a pico-projector, the device can provide a direct manipulation, GUI in the body. The technology was developed by Chris Harrison, Desney Tan, and Dan Morris, at Microsoft Reseach's Computational User Experiences Group. Skinput represents one way to decouple input from electronic devices with the aim of allowing devices to become smaller without simultaneously shrinking the surface area on which input can be performed. While other systems, like Sixth Sensehave attempted this with computer vision, Skinput employs acoustics, which take advantage of the human body's natural sound conductive properties. This allows the body to be annexed as an input surface without the need for the skin to be invasively instrumented with sensors, tracking markers, or other items. Microsofthas not commented on the future of the project, other than it is under active development. It has been reported this may not appear in commercial devices for at least 2 years.

Skinput has been publicly demonstrated as an armband, which sits on the biceps. This prototype contains ten small cantilevered Piezo elements configured to be highly resonant, sensitive to frequencies between 25 and 78 Hz.This configuration acts like a mechanical Fast Fourier transform and provides extreme out-of-band noise suppression, allowing the system to function even while the user is in motion. From the upper arm, the sensors can localize finger taps provided to any part of the arm, all the way down to the finger tips, with accuracies in excess of 90% (as high as 96% for five input locations).Classification is driven by a  support vector machine using a series of time-independent acoustic features that act like a fingerprint. Like speech recognition systems, the Skinput recognition engine must be trained on the "sound" of each input location before use. After training, locations can be bound to interactive functions, such as pause/play song, increase/decrease music volume,speed dial, and menu navigation.
With the addition of a pico-projector to the armband, Skinput allows users to interact with a graphical user interface displayed directly on the skin. This enables several interaction modalities, including button-based hierarchical navigation, list-based sliding navigation (similar to an iPod/SmartPhone/MID), text/number entry (e.g., telephone number keypad), and gaming .
 Despite being a Microsoft Research internal project, Skinput has been demonstrated publicly several times. The first public appearance was at Microsoft's TechFest 2010, where the recognition model was trained live on stage, during the presentation, followed by an interactive walkthrough of a simple mobile application with four modes: music player, email inbox, Tetris and voice.A similar live demo was given at the ACM CHI 2010 conference, where the academic paper received a "Best Paper" award. Attendees were allowed to try the system. Numerous media outlets have covered the technology,with several featuring live demos.

If you’re confused (or scared), here’s how “Skinput” works: with the help of an arm-mounted pico-projector and a Bluetooth connection, the palm and forearm of the user’s body becomes the navigation center of their phone, MP3 player, or other personal technology item. Keyboards/keypads are projected onto the user’s skin which, in turn, respond to touch/tap motions. While touch is a key component to the function of “Skinput,” much of the accuracy lies in its ability to distinguish between specific, inaudible sounds generated by particular motions in the skin and bone of one’s arm.

In addition to the keyboard projection, “Skinput” also responds to various hand gestures, all of which can be programmed per desired function – tap fingers together to answer phone calls, rotate your wrist to scroll up and down, close your fist to exit programs, etc.
 







No comments:

Post a Comment