@inproceedings{SacharaKopinskiGepperthetal.2017, author = {Fabian Sachara and Thomas Kopinski and Alexander Gepperth and Uwe Handmann}, title = {Free-hand gesture recognition with 3D-CNNs for in-car infotainment control in real-time}, series = {IEEE Intelligent Transportation Systems Conference (ITSC2017)}, publisher = {IEEE}, address = {Yokohama, Japan}, isbn = {978-1-5386-1526-3}, issn = {2153-0017}, doi = {10.1109/ITSC.2017.8317684}, pages = {959 -- 964}, year = {2017}, abstract = {In this contribution we present a novel approach to transform data from time-of-flight (ToF) sensors to be interpretable by Convolutional Neural Networks (CNNs). As ToF data tends to be overly noisy depending on various factors such as illumination, reflection coefficient and distance, the need for a robust algorithmic approach becomes evident. By spanning a three-dimensional grid of fixed size around each point cloud we are able to transform three-dimensional input to become processable by CNNs. This simple and effective neighborhood-preserving methodology demonstrates that CNNs are indeed able to extract the relevant information and learn a set of filters, enabling them to differentiate a complex set of ten different gestures obtained from 20 different individuals and containing 600.000 samples overall. Our 20-fold cross-validation shows the generalization performance of the network, achieving an accuracy of up to 98.5\% on validation sets comprising 20.000 data samples. The real-time applicability of our system is demonstrated via an interactive validation on an infotainment system running with up to 40fps on an iPad in the vehicle interior.}, language = {en} }