@InProceedings{Manthey2018_400,
author = {Robert Manthey and Hussein Hussein and René Erler and Richard Siegel and Danny Kowerko},
booktitle = {Studientexte zur Sprachkommunikation: Elektronische Sprachsignalverarbeitung 2018},
title = {Realisation of an Audio & Video Laboratory for Precise Object Localization and Tracking},
year = {2018},
editor = {André Berton and Udo Haiber and Wolfgang Minker},
month = mar,
pages = {135--142},
publisher = {TUDpress, Dresden},
abstract = {This paper presents the realisation of an audiovisual laboratory for detec- tion, localisation, classification and tracking of objects in indoor environments us- ing visual as well as audio information. The laboratory is property of the endowed junior professorship Media Computing at the Chemnitz University of Technology. Visual information is retrieved by 10 optical embedded smart stereo sensors (In- tenta S2000). The visual sensors can be used to trigger an identification process of objects and persons as well as a tracking operation to follow them in the field of view. For audio signal processing, a total of 64 microphones and 16 loudspeakers are used. Three microphone arrays are constructed using three different types of microphones. The camera sensors and microphones can be mounted in different positions, directions and heights. The loudspeakers can be set up freely within the tracking area using standard monitor stands. Within a separate air-conditioned server room, all audio signals are pre-amplified and AD/DA converted using in- dustry standard rack-mounted audio hardware. A server cluster and workstations with high-end Nvidia P6000 graphics cards provide the raw processing power for all processing tasks within the scope of use cases for the laboratory. Software which is available for users in the laboratory comprises commercial products, such as Steinberg Cubase 8.5, as well as self-developed solutions, such as an audio & video localisation and annotation tool. This will be complemented by documenting acoustic sources (e.g. loudspeakers) and detectors (microphones) into a Blender- animated 3D laboratory using the video sensor data. Interested individuals will also have the opportunity to make use of the laboratory's extra peripheral equip- ment, e.g. a Yamaha DGX-660 Keyboard, further types of mobile wireless micro- phones, HTC Wive VR glasses and wireless Internet of Things (IoT) smart home sensor probes for motion capturing. The active tracking area comprises a total of 82 sensors.},
isbn = {978-3-959081-28-3},
issn = {0940-6832},
keywords = {Poster},
url = {https://www.essv.de/pdf/2018_135_142.pdf},
}