Volltext-Downloads (blau) und Frontdoor-Views (grau)

Gesture-based human-machine interaction for assistance systems

  • This contribution demonstrates the efficient embedding of a single depth-camera into the automotive environment making mid-air gesture interaction for mobile applications viable in such a scenario. In this setting a new human-machine interface is implemented to give an idea of future improvements in automation processes in industrial applications. Our system is based on a data-driven approach by learning hand poses as well as gestures from a large database in order to apply them on mobile devices. We register any movement in a nearby driver area and crop data efficiently with the means of PCA transforming it into so-called feature vectors which present the input for our multi-layer perceptrons (MLPs). After MLP classification, the interpretation of user input is sent via WiFi to a tablet PC mounted into the car interior visualizing an infotainment system which the user is able to interact with. We demonstrate that by this setup hand gestures as well as hand poses are easily and efficiently interpretable insofar as that they become an intuitive and supplementary means of interaction for automotive HMI in mobile scenarios realizable in real-time.

Export metadata

Additional Services

Share in Twitter Search Google Scholar


Author:Thomas Kopinski, Stefan Geisler, Uwe Handmann
Parent Title (English):2015 IEEE International Conference on Information and Automation
Document Type:Conference Proceeding
Year of Completion:2015
Release Date:2019/04/29
Issue:8-10 Aug
Page Number:510-517
Institutes:Fachbereich 1 - Institut Informatik
DDC class:000 Allgemeines, Informatik, Informationswissenschaft / 004 Informatik
Licence (German):License LogoNo Creative Commons