@inproceedings{KopinskiGeislerHandmann2015, author = {Thomas Kopinski and Stefan Geisler and Uwe Handmann}, title = {Gesture-based human-machine interaction for assistance systems}, series = {2015 IEEE International Conference on Information and Automation}, volume = {2015}, number = {8-10 Aug}, isbn = {978-1-4673-9104-7}, doi = {10.1109/ICInfA.2015.7279341}, year = {2015}, abstract = {This contribution demonstrates the efficient embedding of a single depth-camera into the automotive environment making mid-air gesture interaction for mobile applications viable in such a scenario. In this setting a new human-machine interface is implemented to give an idea of future improvements in automation processes in industrial applications. Our system is based on a data-driven approach by learning hand poses as well as gestures from a large database in order to apply them on mobile devices. We register any movement in a nearby driver area and crop data efficiently with the means of PCA transforming it into so-called feature vectors which present the input for our multi-layer perceptrons (MLPs). After MLP classification, the interpretation of user input is sent via WiFi to a tablet PC mounted into the car interior visualizing an infotainment system which the user is able to interact with. We demonstrate that by this setup hand gestures as well as hand poses are easily and efficiently interpretable insofar as that they become an intuitive and supplementary means of interaction for automotive HMI in mobile scenarios realizable in real-time.}, language = {en} }