Category: Research

Leap Motion v2: Much improved!

The leap motion controller lets you track your hands and fingers very much like Microsoft’s Kinect lets you track the whole body. This lets you create in-air gesture interfaces for e.g. controlling a robot hand (remote surgery), play games (shoot guns, fly planes) or musical instruments (pull strings).

However, when the leap motion device was released in 2013 it was quite a disappointment. The sensor plus software often loses track of individual fingers, usually when the hand is rotated (even by a small angle). This greatly limited its practical use in applications. It has never been clear whether this has been a software problem or a limitation of the hardware.

Now leap motion has released a new software which seems to indicate that this was a software problem. Even better: it has been fixed in the new v2 release. Look at this video. By the way, I’ve given it a try myself and the signals seem to be much, much more stable.

For downloading the beta version of the new software, go here: https://developer.leapmotion.com

 

Let’s Get Physical

A current direction in human-computer interaction is concerned with bringing back physical elements like real buttons, cubes, bricks, rulers etc. to the world of UI’s. This has been coined “Tangible User Interfaces” (TUIs) by Ishii from MIT (Ishii, Ullmer 1997). The idea is to exploit the physical properties of the object because they suggest a certain operations (round objects invite you to rotate them, buttons invite pressing them) and there is a sensual experience (haptic feedback) involved that is sometimes painful amiss in current touch interfaces.

Here’s an example that is based on the original ideas of Ulmer and Ishii (metaDESK):

Augmented Urban Model from Katja Knecht on Vimeo.

So in TUIs you control digital information (e.g. a desktop) by manipulating physical objects. However, what about the other way round? I.e. if the digital information changes, how can the physical objects be changed? This brings up a whole new set of potentials but also of technical challenges (Ishii et al. 2012).

The latest incarnation of this idea is inFORM, again from Ishii’s Tangible Media research group at MIT (Follmer et al. 2013):

In the arts the idea of bringing digital back to physical is manifest in “kinetic sculptures”, i.e. sculptures that change over time. Here is one impressive example at Singapore airport:

“Kinetic Rain” Changi Airport Singapore from ART+COM on Vimeo.

Literature

Hiroshi Ishii and Brygg Ullmer (1997) Tangible bits: towards seamless interfaces between people, bits and atoms. In: Proceedings of the ACM SIGCHI Conference on Human factors in computing systems (CHI ’97). ACM, New York, NY, USA, 234-241.

Sean Follmer, Daniel Leithinger, Alex Olwal, Akimitsu Hogge, and Hiroshi Ishii (2013) inFORM: dynamic physical affordances and constraints through shape and object actuation. In Proceedings of the 26th annual ACM symposium on User interface software and technology (UIST ’13). ACM, New York, NY, USA, 417-426.

Hiroshi Ishii, Dávid Lakatos, Leonardo Bonanni, and Jean-Baptiste Labrune. (2012) Radical atoms: beyond tangible bits, toward transformable materials. interactions 19, 1 (January 2012), 38-51.

Interfaces with the Kinect

Walking in 3D

Here, the Kinect is used to navigate through Google Street View:

  • body orientation => rotate sideways
  • body tilt => rotate up/down
  • walking (on the spot) => move forward
  • jump => move forward by some distance

Also see post on developkinect.com

In-Air Interaction with ARCADE

A combination of interaction techniques to create and control virtual 3D objects that are placed into the live video of the presenter.

  • selection/picking: hold hand over object
  • menu: browse with finger position, select with swipe gesture
  • drawing: two finger touching switches to draw mode, go to 3D rotate mode with a key press
  • rotate/scale: rotate with finger up/down/left/right, scale with two fingers
  • delete: wave gesture over object

Also see post on developkinect.com

Gesture Recognition: Kinetic Space 2.0

The toolkit allows to define your own 3D gestures.

Also see post on developkinect.com

Real-Time Facial Animation

Photorealism is slowly reaching a stage where the “uncanny valley” does not apply anymore. Activision just showed an impressive video of a face animated in real-time at GDC 2013 (Games Developers Conference):

One of the collaborators is Paul Debevec (USC – ICT, California), one of the superstars of the computer graphics community, inventor of the “light stage“, a device that allows to recreate a huge amount of light moods for a piece of video! Another project of his was “digital emily” (2008) which is probably the base technology behind the Activision demo. Here’s a talk at TEDx:

Anybody interested in facial animation should have a look at the German Animationsinstitut of the Filmakademie Baden-Württemberg who developed the Facial Animation Toolset based on scientific findings on facial expressions (most notably by researcher Paul Ekman who developed the facial action coding system aka FACS).

References

Alexander O., Rogers M., Lambeth W., Chiang M., Debevec P. Creating a Photoreal Digital Actor : The Digital Emily Project. IEEE European Conference on Visual Media Production (CVMP), November 2009. (also to appear in IEEE Computer Graphics and Applications.)

Paul Debevec, Tim Hawkins, Chris Tchou, Haarm-Pieter Duiker, Westley Sarokin, and Mark Sagar. 2000. Acquiring the reflectance field of a human face. In Proceedings of the 27th annual conference on Computer graphics and interactive techniques (SIGGRAPH 2000). ACM Press/Addison-Wesley Publishing Co., New York, NY, USA, 145-156.

 

Copyright © 2024 Michael Kipp's Blog

Theme by Anders NorenUp ↑