Michael Kipp's Blog

Tag: Interaction (page 1 of 4)

Multitouch 84

Please welcome our latest addition to our Interaction Lab at Augsburg University of Applied Sciences, a 84″ multitouch display with a motorized stand which can be transformed into a table, a tilted table and a wall.

The monitor is part of a research grant called Labor zur Erforschung und Entwicklung von Interaktionsmodellen mit Virtuellen Realitäten (laboratory for research and development on interaction models for virtual realities) by Rose, Müller, Rothaug, Kipp which is funded by Hochschule Augsburg.

We intend to investigate the question of how efficiency and ergonomics of multitouch interaction can be measured to compare various input modalities (mouse, controller, gesture, touch). See the publications below to get an idea of our goals and methods. The new display allows to extend our previous work to large display sizes and multi-party scenarios.

Some technical data:

  • 84″
  • 4K resolution (3840×2160)
  • 50 touch points

Related Publications

Lehmann, Florian (2016) Ergonomie von Multi-Touch Oberflächen, Bachelorarbeit, Studiengang Interaktive Medien, Hochschule Augsburg. | Bachelorarbeit |Präsentation | Poster | Read the blog post about this work

Nguyen, Q., and Kipp, M. (2015) Where to Start? Exploring the efficiency of translation movements on multitouch devices. In: Proceedings of 15th IFIP TC 13 International Conference (INTERACT), Springer, pp. 173-191.

Nguyen, Q., Kipp, M. (2014) Orientation Matters: Efficiency of translation-rotation multitouch tasks. In: Proc. of CHI 2014. Link to Video.

Student projects: Interaction Engineering (2016/17)

The next group of talented students completed their interaction engineering projects.  In this interdisciplinary course (computer science and design), we think up potential future human-computer interaction techniques based on current research publications. This semester there was a record-breaking set of 12 completed projects. Feel free to check them out by visiting the project website



Actuated Tangibles: ChainFORM

After the inFORM project (see my post from 2013) here is another spectacular research outcome from Professor Ishii’s Tangible Media Group at MIT.

The idea of tangible interaction goes back as far as 1997 when Ishii first formulated his idea of bringing back physical items to human-computer interfaces. He invented physical controls that allows you to manipulate digital data more intuitively.

Pushing this idea a step further Ishii wondered how to bring digital information back into the real world using actuated tangibles that can dynamically show the changes of the digital information. One problem is changing the position of physical controls (e.g. by air, vibration or magnetic control), more challenging is to change the shape of physical controls on the fly. Both inFORM and ChainFORM deal with the problem of changing shape dynamically.

Relevant Publications

Ken Nakagaki, Artem Dementyev, Sean Follmer, Joseph A. Paradiso, Hiroshi Ishii. ChainFORM: A Linear Integrated Modular Hardware System for Shape Changing Interfaces. In Proceedings of the 29th Annual ACM Symposium on User Interface Software & Technology (UIST ‘16).

Sean Follmer, Daniel Leithinger, Alex Olwal, Akimitsu Hogge, and Hiroshi Ishii. 2013. inFORM: dynamic physical affordances and constraints through shape and object actuation. In Proceedings of the 26th annual ACM symposium on User interface software and technology (UIST ’13). ACM, New York, NY, USA, 417-426.

Hiroshi Ishii and Brygg Ullmer. 1997. Tangible bits: towards seamless interfaces between people, bits and atoms. In Proceedings of the ACM SIGCHI Conference on Human factors in computing systems (CHI ’97). ACM, New York, NY, USA, 234-241.

Bachelor Thesis: Florian Lehmann – Ergonomics of Multi-Touch Surfaces (2016)

Lehmann, Florian (2016) Ergonomie von Multi-Touch Oberflächen, Bachelorarbeit, Studiengang Interaktive Medien, Hochschule Augsburg. | Bachelorarbeit |Präsentation | Poster



The investigation of multi-touch surfaces on smartphones is a relevant topic in the field of human-computer interaction. The main focus is to analyze and to understand touch input in detail, as well as delivering comprehensible insights for user interface designers and developers.

Continue reading

Microsoft Vision Video 2020

Another “Future vision” video from Microsoft that contains snippets of older vision videos. In this video, they added a gesture-controlled bracelet/smart watch (0:17) and a 3D holographic display (0:27).

See my older post for more videos.


Student projects: Interaction Engineering (2015/16)

This semester we had quite impressive student projects, both in number and in quality. Check them out under http://michaelkipp.de/interaction/projects

Bildschirmfoto 2016-02-16 um 08.11.03

Ontenna: Sound to Vibration Device

Tatsuya Honda developed a device that translates sound to vibrations (and light), making it possible for Deaf people to react to environmental sounds and even to differentiate between different sounds. The device is worn as a hairclip as can be seen in the video below. It is yet a prototype.

Read article on venturebeat.com


Student projects: Interaction Engineering (2015)

Two more teams finished their Interaction Engineering projects in summer 2015. You can find them under http://interaction.michaelkipp.de/projects.

Bildschirmfoto 2015-09-15 um 00.45.58


City Pulse: Circular multi-screen installation with gesture control

A new installation on the 100th floor of the 1 World Trade Center features circular screens with gesture control.

The project was realized by Local Projects.

Student projects: Interaction Engineering (2014/15)

Under http://interaction.michaelkipp.de/projects you will find the student prototypes of the Interaction Engineering course of winter 2014/15, Augsburg University of Applied Sciences.

Unter http://interaction.michaelkipp.de/projects finden Sie die Dokumentation der Abschlussprojekte im Kurs “Interaction Engineering” vom WS 2014/15.



Copyright © 2017 Michael Kipp's Blog

Theme by Anders NorenUp ↑