The SlideShow Gestures-WPF sample shows how you are able to use the Kinect for Windows SDK to control Windows applications through the use of gestures. It uses research from the Microsoft Research Cambridge lab to trigger events when the user performs a gesture.
We explore grip and motion sensing to afford new techniques that leverage how users naturally manipulate tablet and stylus devices during pen-and-touch interaction. We can detect whether the user holds the pen in a writing grip or tucked between his fingers. We can distinguish bare-handed inputs, such as drag and pinch gestures, from touch gestures produced by the hand holding the pen, and we can sense which hand grips the tablet, and determine the screen's relative orientation to the pen.
SecondLight is a new surface computing technology that can project images and detect gestures "in mid-air" above the display, in addition to supporting multitouch interactions on the surface.
Motivated by advances in touch-sensing technologies, Surface Computing and multi-touch input support in Windows 7, we have developed a number of prototype mice that consider different physical form-factors and sensing techniques. These devices allow us to do more than point-and-click: they can detect the position of users’ hands and fingers; recognise and react to gestures, and support novel interaction techniques.
SideSight expands the multi-touch capabilities of small mobile devices beyond the screen. Infrared sensors embedded along each side of device are capable of detecting the presence and position of fingers within the proximity of the device. When the device is rested on a flat surface, such as a table, the user can carry out single and multi-touch gestures using the space around the device. This gives a larger input space than would otherwise be possible, and which may be used in conjunction with,
Stroke Recovery with Kinect is an interactive rehabilitation system that helps stroke patients improve their upper-limb motor functioning in the comfort of their own home. By using the Microsoft Kinect sensor’s gesture recognition technology, the system recognizes and interprets the user’s movements, assesses their rehabilitation progress, and adjusts the level of difficulty for subsequent therapy sessions.
Conversational systems interact with people through language to assist, enable, or entertain. Research at Microsoft spans dialogs that use language exclusively, or in conjunctions with additional modalities like gesture; where language is spoken or in text; and in a variety of settings, such as conversational systems in apps or devices, and situated interactions in the real world.
We contribute a thin, transparent, and low-cost design for electric field sensing, allowing for 3D finger and hand tracking, as well as in-air gestures on mobile devices.
We present a new type of augmented mechanical keyboard, sensing rich and expressive motion gestures performed both on and directly above the device.