Refine
Year of publication
Document Type
- Conference Proceeding (63)
- Article (25)
- Report (11)
- Part of a Book (8)
- Contribution to a Periodical (1)
- Doctoral Thesis (1)
- Lecture (1)
- Other (1)
- Research Data (1)
Is part of the Bibliography
- no (112)
Keywords
- Architektur (1)
- Augmented Reality (1)
- Computer Vision (1)
- DamokleS (1)
- Fahrerassistenzsystem (1)
- INTELLIGENT VEHICLES (1)
- Künstliche Intelligenz (1)
- Psychoacoustics (1)
Schlussbericht ; Laufzeit des Vorhabens/Berichtszeitraum: 01.10.2005-30.11.2009
Auch als elektronische Ressource vorh.
Förderkennzeichen BMBF 01IME01K [richtig] - 01IMEO1K [falsch]. - Verbund-Nr. 01042153. - Engl. Berichtsbl. u.d.T.: Face recognition on mobile robot plattforms. - Literaturverz. Bl. 32
Unterschiede zwischen dem gedruckten Dokument und der elektronischen Ressource können nicht ausgeschlossen werden
DamokleS 4.0
(2019)
Dieser interne Bericht beschreibt die Zielsetzung, Durchführung und Auswertung des Projektes Damokles 4.0. Das Projekt zielt darauf ab, neue, digitale Technologien in die Schwerindustrie einzuführen um Produktionsprozesse zu modernisieren. Unter Einsatz neuer Technologien, insbesondere mobiler Geräte, soll ein cyberphyiskalisches System (CPS) eine kontextbasierte und künstlich intelligente Unterstützung der Mitarbeiter in den Bereichen der Schwerindustrie ermöglichen. Hierzu werden typische Anwendungsfälle und die damit verbundenen Szenarien zur Unterstützung der Mitarbeiter auf Basis von neuen, flexiblen, adaptiven und mobilen Technologien, wie Augmented Reality und künstlicher Intelligenz, modelliert. Um den Prototypen einer AR-Anwendung und einer kamerabasierte Personenverfolgung zu entwickeln, hat die Hochschule Ruhr West im kleinen Technikum am Campus Bottrop eine entsprechende industrielle Umgebung simuliert. Die Projektergebnisse zeigen die Anwendbarkeit der vorgeschlagenen Softwareansätze und die Ergebnisse einer Untersuchung der psychologischen Einflüsse auf die Mitarbeiter.
In this review, we describe current Machine Learning approaches to hand gesture recognition with depth data from time-of-flight sensors. In particular, we summarise the achievements on a line of research at the Computational Neuroscience laboratory at the Ruhr West University of Applied Sciences. Relating our results to the work of others in this field, we confirm that Convolutional Neural Networks and Long Short-Term Memory yield most reliable results. We investigated several sensor data fusion techniques in a deep learning framework and performed user studies to evaluate our system in practice. During our course of research, we gathered and published our data in a novel benchmark dataset (REHAP), containing over a million unique three-dimensional hand posture samples.
We present a pipeline for recognizing dynamic freehand gestures on mobile devices based on extracting depth information coming from a single Time-of-Flight sensor. Hand gestures are recorded with a mobile 3D sensor, transformed frame by frame into an appropriate 3D descriptor and fed into a deep LSTM network for recognition purposes. LSTM being a recurrent neural model, it is uniquely suited for classifying explicitly time-dependent data such as hand gestures. For training and testing purposes, we create a small database of four hand gesture classes, each comprising 40 × 150 3D frames. We conduct experiments concerning execution speed on a mobile device, generalization capability as a function of network topology, and classification ability ‘ahead of time’, i.e., when the gesture is not yet completed. Recognition rates are high (>95%) and maintainable in real-time as a single classification step requires less than 1 ms computation time, introducing freehand gestures for mobile systems.
We present a system for 3D hand gesture recognition based on low-cost time-of-flight(ToF) sensors intended for outdoor use in automotive human-machine interaction. As signal quality is impaired compared to Kinect-type sensors, we study several ways to improve performance when a large number of gesture classes is involved. Our system fuses data coming from two ToF sensors which is used to build up a large database and subsequently train a multilayer perceptron (MLP). We demonstrate that we are able to reliably classify a set of ten hand gestures in real-time and describe the setup of the system, the utilised methods as well as possible application scenarios.
In this contribution we present a novel approach to transform data from time-of-flight (ToF) sensors to be interpretable by Convolutional Neural Networks (CNNs). As ToF data tends to be overly noisy depending on various factors such as illumination, reflection coefficient and distance, the need for a robust algorithmic approach becomes evident. By spanning a three-dimensional grid of fixed size around each point cloud we are able to transform three-dimensional input to become processable by CNNs. This simple and effective neighborhood-preserving methodology demonstrates that CNNs are indeed able to extract the relevant information and learn a set of filters, enabling them to differentiate a complex set of ten different gestures obtained from 20 different individuals and containing 600.000 samples overall. Our 20-fold cross-validation shows the generalization performance of the network, achieving an accuracy of up to 98.5% on validation sets comprising 20.000 data samples. The real-time applicability of our system is demonstrated via an interactive validation on an infotainment system running with up to 40fps on an iPad in the vehicle interior.
Coming out of the labs, the first robots are currently appearing on the consumer market. Initially they target rather simple application scenarios ranging from entertainment to home convenience. However, one can expect, that they will capture more complex areas soon. These robots will have a higher and higher level and a broad range of functional competence, and will collaborate and interactively communicate with their human users. All this requires considerable cognitive abilities on the robot’s side and appropriate man-machine interaction technologies. Apart from further development of individual functions and technologies it is crucial to build and evaluate fully integrated systems. This paper describes our approach to construct a robotic assistance system. We present experience with an integrated technology demonstration and the exposure of the integrated system to the public.
In diesem Artikel wird eine flexible Architektur vorgestellt, mit deren Hilfe eine modulare Lösung von Fahrerassistenzaufgaben in Kraftfahrzeugen gezeigt werden kann. Es wird eine Objektbezogene Analyse von Sensordaten, eine Verhaltensbasierte Szeneninterpretation und eine Verhaltensplanung vorgestellt. Eine globale Wissensbasis, auf der jedes einzelne Modul arbeitet, beinhaltet die Beschreibung physikalischer Zusammenhänge, Verhaltensregeln für den Straßenverkehr, sowie Objekt- und Szenenwissen. Externes Wissen (z.B. GPS - Global Positioning System) kann ebenfalls in die Wissensbasis eingebunden werden. Als Anwendungsbeispiel der Verhaltensplanung wird ein intelligenter Tempomat vorgestellt.
Pedestrian movement analysis at airports - videobased analysis across multiple camera systems
(2013)
Analyse dynamischer Szenen
(1999)
In diesem Artikel wird die Analyse dynamischer Szenen im Rahmen einer flexiblen Architektur zur Lösung von Fahrerassistenzaufgaben in Kraftfahrzeugen vorgestellt. Die Lösung unterschiedlicher Aufgaben mit verwandten Ansätzen bedingt einen hohen Grad an Modularität und Flexibilität. Nur so können die gestellten Aufgaben mit den vorhandenen Algorithmen optimal gelöst werden. In der vorgestellten Architektur wird eine objektbezogene Analyse von Sensordaten, eine verhaltensbasierte Szeneninterpretation und eine Verhaltensplanung durchgeführt. Eine globale Wissensbasis, auf der jedes einzelne Modul arbeitet, beinhaltet die Beschreibung physikalischer Zusammenhänge, Verhaltensregeln für den Straßenverkehr, sowie Objekt- und Szenenwissen.
Externes Wissen (z.B. GPS – Global Positioning System) kann ebenfalls in die Wissensbasis eingebunden werden. Als Anwendungsbeispiel der Verhaltensplanung ist ein intelligenter Tempomat realisiert.
Touch versus mid-air gesture interfaces in road scenarios-measuring driver performance degradation
(2016)
We present a study aimed at comparing the degradation of the driver's performance during touch gesture vs mid-air gesture use for infotainment system control. To this end, 17 participants were asked to perform the Lane Change Test. This requires each participant to steer a vehicle in a simulated driving environment while interacting with an infotainment system via touch and mid-air gestures. The decrease in performance is measured as the deviation from an optimal baseline. This study concludes comparable deviations from the baseline for the secondary task of infotainment interaction for both interaction variants. This is significant as all participants are experienced in touch interaction, however have had no experience at all with mid-air gesture interaction, favoring mid-air gestures for the long-term scenario.
Applying step heating thermography to wind turbine rotor blades as a non-destructive testing method
(2017)
The Desire project aimed at the development and implementation of a mobile service robotic research platform (technology platform) able to handle real world scenarios regarding service robotic tasks. Different modules for different tasks plus an interaction infrastructure were integrated on this platform. An example of a real world scenario task is the support of a handicapped person to clean up a kitchen in home environments.
One of the main challenges to be solved in this field is the interaction with people. To start an interaction process between a robot and a person, the most important information is the knowledge about the interacting partner’s identity and whether the interacting partner is present or not. This means, the robot must be able to detect and be finally able to identify persons. Accurate identification of specific individuals has to be done by analyzing the individual features of each person. A typical feature set that allows for a distinct identification of a specific person is often extracted from the facial image acquired by a camera. This feature-set is stored in a database to allow the identification of different persons independent from place and time by comparing given feature-sets. Thus, a face recognition module was integrated into the technology platform which includes face detection and identification algorithms.
In this paper, we describe a method to model human clothes for a later recognition by the use of RGB- and SWIR-cameras. A basic model is estimated during people detection and tracking. This model will be refined if the recognition is triggered. For the refining, several saliency maps are used to extract individual features. These individual features are located separately for any human body parts. The body parts are estimated by the use of a silhouette extraction combined with a skeleton estimation. In this way, the model describes the human clothes in a compact manner which allows the use of a simple and fast comparison method for people recognition. Such models can be used in security and service applications.
Im vorliegenden Beitrag wird ein hochsprachenprogrammierbares System zur schritthaltenden Vollbild-Interpretation natürlich beleuchteter Szenenfolgen im Videotakt vorgestellt. Im einzelnen werden folgende Teilmodule und Subsysteme beschrieben: eine hochdynamische, pixellokal autoadaptive CMOS-Kamera mit ca. 120 dB Helligkeitsdynamik (20Bits/Pixel) ein hochsprachenprogrammierbarer Systolic Array Prozessor (für die pixelbezogenen Verarbeitungsmodule) im PCI-Kartenformat, samt optimierendem Compiler, Simulator und Emulator Systemprozeßgerüste unter Linux auf den für die Echtzeit-Anwendungen eingesetzten Hostrechnern (z.B. DEC/Alpha oder Intel/ Pentium)eine prototypische Anwendung zur bildverarbeitungsbasierten Eigenbewegungsbeobachtung (Translationsrichtung, Eotationsraten)eine prototypische, automotive Anwendung zur schritthalt enden Detektion und Kartierung des Straßen- und Spurverlaufs unter partieller monokularer 3D-Rekonstruktion, sowie prototypische Anwendungen zur Klassifikation verkehrsrelevanter Hindernisse (Verkehrsteilnehmer)
Aktiv im Alter
(2016)
Die Prognosen für den demografischen Wandel sind eindeutig: In den kommenden Jahren wird es immer mehr Menschen über 65 Jahre geben. Damit verbunden sind große Herausforderungen für die Gesellschaft und ihre Sozialsysteme, aber auch für viele Angehörige, die ihre Verwandten im Alter pflegen. Doch nicht alle älteren Menschen leben im Kreise ihrer Familie oder können sich Fremdbetreuung durch Pflegedienste leisten. Häufig übernehmen Nachbarn oder Freunde aus der Umgebung diese Aufgabe. Für diese Menschen wird das Wohnquartier zum zentralen Gesundheitsstandort.
Im besten Fall können sie dort ihren Alltag noch lange selbstständig bewältigen und ihre sozialen Kontakte aufrechterhalten. Das soll bald eine App unterstützen. Sie ist Teil eines Trainingsprogramms, das die Hochschule für Gesundheit (hsg) im Verbund mit der Hochschule Ruhr West erarbeitet. Der Name des Projekts ist Programm: „Quartier agil – Aktiv vor Ort“. Mit Übungen zum kognitiven und körperlichen Training, Angeboten für Gruppenaktivitäten, Kommunikationsforen und Funktionen zur Selbstkontrolle wollen die Forscherinnen und Forscher
ältere Menschen fit halten.
Currently in home environments, robot assisting systems with emotion understanding ability are generally achieved in two several manners. The first is the implementing of such systems in such a way that they offer general services for all considered persons without considering privacy, special needs of their interaction partners. The second way is the targetting of such systems for merely one person. In this work we present a robot assisting system, which has both the abilities of assisting several persons at the same time and sustaining their privacy and security issues. The robot can interact with it's interaction partner emotionally by analyzing the emotions of her expressed either visually, facial expression, or auditive, speech prosody. The role of this system is the providing of person-specific support in home environment. In order to identify its interaction partner the system uses diverse biometric traits. According to the recognized ID the system, first, adopts towards the needs of recognized person. Second the system loads the corresponding emotional profile of the detected interaction partner in order to practice a person-specific emotional human-robot interaction, which has an advantage over the person independent interaction.
Positive Computing umfasst Design, Realisierung und Bewertung von Anwendungssystemen und deren Einflüsse mit dem Ziel, Lebensqualität und Wohlbefinden von Menschen zu verbessern und sie bei der Entfaltung ihrer Potenziale zu unterstützen. Das Institut Positive Computing (IPCo) an der Hochschule Ruhr West soll dieses neue Paradigma in einem inter- und transdisziplinären Ansatz erschließen, untersuchen und umsetzen. Das Paradigma ist anwendbar auf nahezu alle Bereiche des privaten und beruflichen Lebens. Die Forschung des IPCo fokussiert zunächst jedoch auf die positive Nutzung von Informations- und Kommunikationstechnologien (IKT) für generationenübergreifende Herausforderungen. Hierzu sollen technologische Lösungen unter kontinuierlicher Einbeziehung menschlicher Bedürfnisse und sozialer Fragestellungen erarbeitet
werden.
Systeme zur automatisierten Bildanalyse sind vielfältig einsetzbar und gewinnen aufgrund technologischer Weiterentwicklungen und gesellschaftlicher Akzeptanz zunehmend an Bedeutung. Schwerpunkt im Bereich der "Technischen Bildverarbeitung dynamischer Szenen" ist die Entwicklung von Methoden, die bei der Interpretation von Bildern aus verschiedenen Sensordaten Verwendung finden. Dies sind neben den herkömmlichen Kamerabildern im wesentlichen Röntgen- und Radarbilder. Unter geeigneter Berücksichtigung der durch die jeweiligen Anwendungen vorgegebenen Randbedingungen werden daraus entsprechende Verfahren abgeleitet. Derzeitige Projekte beschäftigen sich mit der Analyse von Straßenverkehrsszenen, der Detektion von Sprengstoffzündern bei der Durchleuchtung von Fluggepäck, sowie mit der Bestimmung von Art und Ausdehnung von Ölverschmutzungen bei der Meeresüberwachung.
Systems for automated image analysis are useful for a variety of tasks and their importance is still increasing due to technological advances and an increase of social acceptance. The main focus of "Technical Image Processing of Dynamic Scenes" lies
with the development of methods for the interpretation of images derived from various sensors. Apart from conventional visual images, this involves mainly X-ray and radar images. Taking into account the requirements of the various applications, suitable methods are derived. Current projects are dealing with the analysis of traffic scenes, detection of detonators when X-raying luggage and determination of type and expansion of oil pollution in maritime surveillance.
Technologie die beflügelt
(2016)
We present a study on 3D based hand pose recognition using a new generation of low-cost time-of-flight(ToF) sensors intended for outdoor use in automotive human-machine interaction. As signal quality is impaired compared to Kinect-type sensors, we study several ways to improve performance when a large number of gesture classes is involved. We investigate the performance of different 3D descriptors, as well as the fusion of two ToF sensor streams. By basing a data fusion strategy on the fact that multilayer perceptrons can produce normalized confidences individually for each class, and similarly by designing information-theoretic online measures for assessing confidences of decisions, we show that appropriately chosen fusion strategies can improve overall performance to a very satisfactory level. Real-time capability is retained as the used 3D descriptors, the fusion strategy as well as the online confidence measures are computationally efficient.
Increasing economic viability and safety through structural health monitoring of wind turbines
(2017)
Serious accidents with property damage or even human casualties, result from structural flaws in wind turbine rotor blades. Common maintenance practices result in long downtimes and do not lead to the required results. Therefore, the Ruhr West University of Applied Sciences and the iQbis Consulting GmbH, currently research a new structural health monitoring method for wind turbine rotor blades. The goal of this project is to build a sensor system that can detect structural weaknesses inside of rotor blades without the need of downtime for industrial climbers. This technology has the potential to prevent accidents, save lives, extend the useful life of wind turbines and optimize the production of green energy.
In this article we present a system for coupling different base algorithms and sensors for segmentation. Three different solutions for image segmentation by fusion are described, compared and results are shown. The fusion of base algorithms with colorinformation and a sensor fusion process of an optical and a radar sensor including a feedback over time is realized. A feature-in decision-out fusion process is solved. For the fusion process a multi layer perceptron (MLP) with one hidden layer is used as a coupling net. The activity of the output neuron represents the membership of each pixel to an initial segment.
Derzeitige Projekte am Institut für Neuroinformatik in Bochum beschäftigen sich mit der Analyse von Straßenverkehrsszenen mittels Computer Vision [12]. Dies impliziert, wegen der durch die natürliche Umwelt aufgestellten Randbedingungen, hohe Anforderungen an die zu entwickelnden Algorithmen. Im speziellen wird versucht, Verkehrsteilnehmer aus Videobildern zu extrahieren und die so gewonnenen Objekthypothesen weiter zu attributieren (z.B. Objektklasse, Abstand, Geschwindigkeit, Gefahrenpotential hinsichtlich der beabsichtigten Eigentrajektorie etc.), um im Hinblick auf den Einsatz in Assistenzsystemen in Fahrzeugen eine möglichst genaue Beschreibung der Umwelt zu erreichen. Nicht nur die große Vielfalt der unterschiedlichen Umweltszenarien, sondern auch das hohe Maß an Sicherheit, das die gestellte Aufgabe erfordert, bedingen ein breitbandiges und flexibles Gesamtsystem [6]. Ein Lösungsvorschlag wird im folgenden behandelt.
We present a publicly available benchmark database for the problem of hand posture recognition from noisy depth data and fused RGB-D data obtained from low-cost time-of-flight (ToF) sensors. The database is the most extensive database of this kind containing over a million data samples (point clouds) recorded from 35 different individuals for ten different static hand postures. This captures a great amount of variance, due to person-related factors, but also scaling, translation and rotation are explicitly represented. Benchmark results achieved with a standard classification algorithm are computed by cross-validation both over samples and persons, the latter implying training on all persons but one and testing on the remaining one. An important result using this database is that cross-validation performance over samples (which is the standard procedure in machine learning) is systematically higher than cross-validation performance over persons, which is to our mind the true application-relevant measure of generalization performance.
Given the success of convolutional neural networks (CNNs) during recent years in numerous object recognition tasks, it seems logical to further extend their applicability to the treatment of three-dimensional data such as point clouds provided by depth sensors. To this end, we present an approach exploiting the CNN’s ability of automated feature generation and combine it with a novel 3D feature computation technique, preserving local information contained in the data. Experiments are conducted on a large data set of 600.000 samples of hand postures obtained via ToF (time-of-flight) sensors from 20 different persons, after an extensive parameter search in order to optimize network structure. Generalization performance, measured by a leave-one-person-out scheme, exceeds that of any other method presented for this specific task, bringing the error for some persons down to 1.5 %.
Fahrerassistenzsysteme werden eingesetzt, um dem Fahrer
eines Kraftfahrzeugs Handlungsabläufe abzunehmen. Diese Handlungsabläufe
werden definiert durch eine Aufgabenstellung, die vom Fahrer an das Fahrerassi-
stenzsystem übergeben oder systembedingt gelöst wird. Bei komplexen Fahreras-
sistenzsystemen ist an eine autonome Navigation im Straßenverkehr gedacht. Es
wird ein neues Verfahren vorgestellt, welches eine Bewegungssteuerung eines
autonomen Fahrzeugs durchführen kann. Es werden der Lenkwinkel und die Ge-
schwindigkeit beeinflußt. Für diese Aufgabe wird ein dynamischer Ansatz aus
dem Bereich der neuronalen Felder gewählt. Relevante Attribute für den Fahrt-
verlauf auf unterschiedlichem Abstraktionsniveau können dabei einfach (additiv)
verarbeitet werden.
Checking wind turbines for damage is a common problem for operators of wind parks, as regular inspections are legally required in many countries and prevention is economically viable. While some of the common forms of damage are easily visible on the surface, structural problems can remain invisible for years before they eventually result in catastrophic failure of a rotor blade. Common forms of testing fibre composite parts like ultrasonic testing or X-ray tests are impractical due to the large dimensions of wind turbine components and their limited accessibility for any short-range methods. Active thermographic inspection of wind turbines is a promising approach to testing for structural flaws beneath the surface of rotor blades. As part of an ongoing research project, a setup for testing the general viability of this method was built and used to compare different thermographic cameras. A sample cut from a discarded rotor blade was modified to emulate structural damage. The results are promising for the development of a cost effective on-site testing system.
Systems for automated image analysis are useful for a variety of tasks. Their importance is still growing due to technological advances and increased social acceptance. Especially driver assistance systems have reached a high level of sophistication. Fully or partly autonomously guided vehicles, particularly for road traffic, require highly reliable algorithms due to the conditions imposed by natural environments. At the Institut fur Neuroinformatik, methods for analyzing driving relevant scenes by computer vision are developed in cooperation with several partners from the automobile industry. We present a system extracting important information from an image taken by a CCD camera installed at the rear-view mirror in a car. The approach is divided into a sequential and a parallel phase of sensor and information processing. Three main tasks, namely initial segmentation (object detection), object tracking and object classification are realized by integration in the sequential phase and by fusion in the parallel phase. The main advantage of this approach is integrative coupling of different algorithms providing partly redundant information. q 2000 Elsevier Science B.V. All rights reserved.
In this paper, we describe an efficient method for a fast people re-identification based on models of human clothes. An initial model is estimated during people detection and tracking, which will be refined during the re-identification. This stepwise extraction, combination and comparing of features speeds up the whole re-identification. For the refining, several saliency maps are used to extract individual features. These individual features are located separately for any human body part. The body parts are located with an optimized GPU-based HOG detector. Furthermore, we introduce a meanshift-based fusion concept which utilizes multiple detectors in order to increase the detection reliability.
Autonomous driving is one of the future visions in which many vehicle manufacturers are working with high pressure.
Nowadays, it is already supported partially by high-class vehicles. A completely autonomous journey is indeed the goal, but in cars for
the public road traffic still not available. Automatic lane keeping assistants, speed regulators as well as shield and obstacle detections
are parts or precursors on the way to completely autonomous driving.
The American vehicle manufacturer Tesla is not only known for its electric drive, but also for the fact that high-pressure work is carried out on the autonomous drive. Tesla is thus the only vehicle manufacturer to use its users as so-called beta testers for its assistance systems. The progress and the function of the currently available Model S in the field of assistance systems and autonomic driving is documented and described in this paper. It is shown how good or bad the test vehicle manages scenarios in normal road traffic situations
with the assistance systems, e.g. lane keeping assistant, speed control, lane change and distance assistant, and which scenarios can
not be managed by the vehicle itself.
Ziel des Verbundprojektes APFel (Projektlaufzeit: 01.01.2010 ‐ 31.03.2014)war eine zeitlich vorwärts‐ und rückwärtsgerichtete Lokalisation von Personen innerhalb eines Kameranetzwerkes aus sich nicht überlappenden Kameras in Hyperechtzeit zu ermöglichen. Einsatzbereiche dieses Szenarios sind kritische Infrastrukturen wie Flughäfen und Flugplätze. Zunächst fokussierte das Projekt APFel auf die Lokalisation einer einzelnen Zielperson. Weiterführend wurden die entwickelten Verfahren auf die Analyse von Gruppen erweitert, um Personen als Teil einer Gruppe lokalisieren zu können.