Thursday, May 6, 2010

Real-time robust body part tracking for augmented reality interface

Authors:
Jinki Jung
Kyusung Cho
Hyun S. Yang
KAIST - interaction lab in korea

Summary:

Goal: to model the 3D body from the 2D detected body parts. a step towards more intuitive interaction in AR environment. (project NATAL)
- system deletes the background to get one single body blob.
- uses skin texture to detect face and 2 hands.
- uses lower body contour to detect leg and feet.
- users particle tracking to track the head through the frames.

The 3D pose is then estimated from the body parts recognized. An adaptive color detection is used to provide clothing and illumination independent detection of body parts. The 3D location is calculated based on the center point camera and its distance to head , and 2 feet.

Discussion:
There is one more person in MIT media lab with strange name doing the same kind of research. The skin texture detection independent of clothing color is cool.

Comments: Franck, Murat

That one there! Pointing to establish device identity

Authors:
Colin Swindells
Kori M. Inkpen
John C. Dill
Melanie Tory


Summary:
Goal: Facilitate Human - Computer Identification with a pointing device.

As the number of devices increase, users need to go through the cumbersome process of remembering the wireless settings and other parameters to connect to devices in the environment. The scenario can also applied to when the user wants to share documents with other users around in a wireless network. This paper provides a device which works on micro-controller with IR receivers and transmitters that can used to point at some device and connect to it easily.
A user study was conducted with 4 users each for 2 phases. First phases required users to select mobile devices and the second included selection of item from the given list. A pre and post questionnaire data were collected. The results were then compared to the performance of users with graphical list from which the users can select a device of choice.

Discussion:
The user study and the data analysis were pretty thorough. There does not seem to much preference for the device over the graphical list (UI).

Webcam Mouse Using Face and Eye Tracking in Various Illumination Environments

Summary:
Goal: environment lighting independent face detection from the webcam video stream.

The red and blue subspace is used to reduce the illumination noise. Since the skin texture for fac eis dependent on the illumination condition. The illumination condition is identified first and then face texture pattern is modeled using 10 examples for each illumination condition. Motion detection technique is used to eliminate color region similar to face in the background.
Iris gives out luminance and its detected using the sharp change in the Y component.

Discussion:
Statistics on accuracy is not given. The KNN classifier would be drastically slow in case of more examples. Each illumination setting requires a set of examples to be provided. How do we enumerate all the illumination setting ? I do not think the recognition algorithm would scale to settings that has no examples.

Comment: franck, Murat

Wednesday, May 5, 2010

XWand: UI for intelligent spaces

Authors:
Andrew Wilson
Stephen Shafer
Microsoft Research


Summary:

Goal: to build an interaction device which can be used to point and interact with multiple devices around the user. (including voice medium).
- Orientation - combination of magnetometer and accelerometer. Can be affected by the metal in the environment
- Position - uses vision techniques to find the 3D position from 2 - 2D position. Tracks 2 IR LEDs.
An average error of 6' is found in pointing tasks (pointing accuracy of device).

Dynamic bayes network is used to process the events from the wand (gestures / button events) and the speed recognition events. Speech recognition allows multimodal interaction with devices in the user environment and also provides multiple methods to perform one operation. Ignoring speech recognition results based on the pointing context helps improve the speech recognition. "Volume up" is ignored while pointing at the lights.

A user study with 10 male users was performed. Variables - Time taken to complete task, accuracy of pointing and responses to questionnaire. Scenarios - Tracking & No tracking , Audio Feedback & No Audio feedback were tested.
The users did not find the audio feedback very useful while tracking of the wand is enabled.

Discussion:
The wand is familiar to the wand simulated with Wii for Wiizards games and i do not remember another such want pointing device (mounted on mouse). Given the usage scenario, the device is novel.

Comments: franck

Wednesday, April 28, 2010

Online, Interactive Learning of Gestures for Human/Robot Interfaces

Authors:
Christopher Lee
Yangsheng Xu

Summary:
Goal: Provde better interactivity and control over robots (effectiveness of tele-operation).

Cyberglove is used record hand gesture input from the users. Hidden Markov model is used to recognize the gestures.
- Users perform a series of gesture
- HMM train on the gesture if they classify them into one of the existing gestures and perform the related action.
- if HMM is not able to classify,
- asks the user to disambiguate and train.
- new gesture added by the user.

Discussion:

the preprocessing procedure is very interesting. It is applicable to PowerGlove++ project by Drew & co.

Comments: Franck

RealTime HandTracking as a User Input Device

Author:
Robert Y Wang

Summary:
Goal: Easy to use and inexpensive system for user input using hands.
Design: a glove with color patterns . optimal color pattern and pose recognition algorithms have been explained in this paper.

A nearest neighbor approach is used to recognize a hand pose. every pose gives different image. So the image lookup approach is being used. The query image is normalized and down-sampled before nearest neighbor is looked up.
As the database size increased the RMS error to the nearest neighbor image decreased. To increase the retrieval rate the author compressed the image into a 128 bit binary sequence and used hamming distance to compare images.

future work:
The experiments to determine the optimal color pattern. Three dimensions of change - color, spatial frequency of blobs and icons / shapes used.

Discussion:
It is a extremely cheap solution. I would like to read about the studies on the pattern color/ spatial frequency studies to know its significance.

I think the separability / difference between the different poses is a catch. The system works well for poses for which the images are different but i do not know if that would restrict the number of the poses.

Wednesday, April 21, 2010

Liquids, smoke, and soap bubbles: reflections on materials for ephemeral user interfaces

Authors:
Axel Sylvester
Tanja Dring
Albrecht Schmidt

Summary:
Goal: Tangible ephemeral interface design with soap, bubble and smoke.
Setup - a round table - 20 inch in diameter, bubbles can be blown on the surface which stays on it for minutes together. The movement of the bubbles can be tracked with the camera beneath the table. The user moves the bubbles using either moist hand or slightly blowing the bubble. In a first application bubbles are used to influence the brightness and hue of the
surrounding room light. The room illumination gets brighter the bigger a recognized bubble is. Hue is set according to the position of a recognized bubble using the x and y axis to
bring up blue and red tones

Discussion:
Interesting mode of interaction. I could see applications in fun filled games for children.