Refine
Year of publication
- 2014 (23) (remove)
Document Type
- Conference Proceeding (16)
- Article (4)
- Part of a Book (1)
- Doctoral Thesis (1)
- Preprint (1)
Language
- English (23) (remove)
Has Fulltext
- no (23)
Is part of the Bibliography
- no (23)
Keywords
- Halberzeugnis (1)
- Hochtemperatur (1)
- Inprozesskontrolle (1)
- Rundstahl (1)
- Warmwalzen (1)
We present a study on 3D based hand pose recognition using a new generation of low-cost time-of-flight(ToF) sensors intended for outdoor use in automotive human-machine interaction. As signal quality is impaired compared to Kinect-type sensors, we study several ways to improve performance when a large number of gesture classes is involved. We investigate the performance of different 3D descriptors, as well as the fusion of two ToF sensor streams. By basing a data fusion strategy on the fact that multilayer perceptrons can produce normalized confidences individually for each class, and similarly by designing information-theoretic online measures for assessing confidences of decisions, we show that appropriately chosen fusion strategies can improve overall performance to a very satisfactory level. Real-time capability is retained as the used 3D descriptors, the fusion strategy as well as the online confidence measures are computationally efficient.
Building upon prior results, we present an alternative approach to efficiently classifying a complex set of 3D hand poses obtained from modern Time-Of-Flight-Sensors (TOF). We demonstrate it is possible to achieve satisfactory results in spite of low resolution and high noise (inflicted by the sensors) and a demanding outdoor environment. We set up a large database of pointclouds in order to train multilayer perceptrons as well as support vector machines to classify the various hand poses. Our goal is to fuse data from multiple TOF sensors, which observe the poses from multiple angles. The presented contribution illustrates that real-time capability can be maintained with such a setup as the used 3D descriptors, the fusion strategy as well as the online confidence measures are computationally efficient.
We present a system for 3D hand gesture recognition based on low-cost time-of-flight(ToF) sensors intended for outdoor use in automotive human-machine interaction. As signal quality is impaired compared to Kinect-type sensors, we study several ways to improve performance when a large number of gesture classes is involved. Our system fuses data coming from two ToF sensors which is used to build up a large database and subsequently train a multilayer perceptron (MLP). We demonstrate that we are able to reliably classify a set of ten hand gestures in real-time and describe the setup of the system, the utilised methods as well as possible application scenarios.
PROPRE is a generic and modular neural learning paradigm that autonomously extracts meaningful concepts of multimodal data flows driven by predictability across modalities in an unsupervised, incremental and online way. For that purpose, PROPRE consists of the combination of projection and prediction. Firstly, each data flow is topologically projected with a self-organizing map, largely inspired from the Kohonen model. Secondly, each projection is predicted by each other map activities, by mean of linear regressions. The main originality of PROPRE is the use of a simple and generic predictability measure that compares predicted and real activities for each modal stream. This measure drives the corresponding projection learning to favor the mapping of predictable stimuli across modalities at the system level (i.e. that their predictability measure overcomes some threshold). This predictability measure acts as a self-evaluation module that tends to bias the representations extracted by the system so that to improve their correlations across modalities. We already showed that this modulation mechanism is able to bootstrap representation extraction from previously learned representations with artificial multimodal data related to basic robotic behaviors [1] and improves performance of the system for classification of visual data within a supervised learning context [2]. In this article, we improve the self-evaluation module of PROPRE, by introducing a sliding threshold, and apply it to the unsupervised classification of gestures caught from two time-of-flight (ToF) cameras. In this context, we illustrate that the modulation mechanism is still useful although less efficient than purely supervised learning.
Mobile devices are nowadays used almost ubiquitously by a large number of users. 2013 was the first year in which the number of sold mobile devices (tablet computers and mobile phones) outperformed the number of PCs’ sold. And this trend seems to be continuing in the coming years. Additionally, the scenarios in which these kinds of devices are used, grow almost day by day. Another trend in modern landscapes is the idea of Cloud Computing, that basically allows for a very flexible provision of computational services to customers. Yet, these two trends are not well connected. Of course there exists already quite a large amount of mobile applications (apps) that utilize Cloud Computing based services. The other way round, that mobile devices provide one of the building blocks for the provision of Cloud Computing based services is not well established yet. Therefore, this paper concentrates on an extension of a technology that allows to provide standardized Web Services, as one of the building blocks for Cloud Computing, on mobile devices. The extension hereby consists of a new approach that now also allows to provide asynchronous Web Services on mobile devices, in contrast to synchronous ones. Additionally, this paper also illustrates how the described technology was already used in an app provided by a business partner.
This paper describes the design and development stages of a web-based framework, aiming to support the creation of mobile applications within the context of mobile learning. The suggested approach offers the opportunity to deploy and execute these applications on mobile devices. This web-based solution additionally offers the possibility to visualize the data collected by the mobile applications in a web-browser. Despite previous research efforts carried out in this domain, few of the projects have addressed these processes from a purely web-based perspective. Currently, a prototype of an authoring tool for creating mobile data collection applications is already implemented. In order to integrate and validate this solution in everyday educational settings, we are collaborating with a network of high schools. On the basis of workshops with teachers we will carry out, refinements and requirements for further enhancements will be collected and will be used to guide our coming efforts.
Recommender systems have become an important application domain related to the development of personalized mobile services. Thus, various recommender mechanisms have been developed for filtering and delivering relevant information to mobile users. This paper presents a rich context model to provide the relevant content of news to the current context of mobile users. The proposed rich context model allows not only providing relevant news with respect to the user’s current context but, at the same time, also determines a convenient representation format of news suitable for mobile devices.
In recent years, teachers have started to conduct pedagogical activities to promote different kinds of learning interactions supported by rich media. The deployment of such activities is rapidly increasing, as teachers and students own technological means that allow supporting them along such interactions. These activities can be carried out in traditional classroom settings while using regular computers. Additionally, they can also be conducted from anywhere at any time while using smartphones and tablets. In this paper, we describe a pedagogical activity requiring students to author and later peer- assess learning interactions
incorporated to videos in YouTube. We describe EDU.Tube, an environment that enables them to create, share and consume such rich media learning activities across a variety of devices. We then detail a plan for the implementation of an activity that took place in 3 different classes dealing with diverse materials addressing computer science related topics. Finally, we also
provide an evaluation presenting students' insights and feedbacks resulting from the experienced activity. We discuss and analyze these outcomes in order to elaborate on them as concerns that could be applied for the further deployment of the EDU.Tube environment.