Cooperative and asynchronous stereo vision for dynamic vision sensors

Dynamic vision sensors (DVSs) encode visual input as a stream of events generated upon relative light intensity changes in the scene. These sensors have the advantage of allowing simultaneously high temporal resolution (better than 10 µs) and wide dynamic range (>120 dB) at sparse data representa...

Full description

Saved in:
Bibliographic Details
Published inMeasurement science & technology Vol. 25; no. 5; pp. 55108 - 55115
Main Authors Piatkowska, E, Belbachir, A N, Gelautz, M
Format Journal Article
LanguageEnglish
Published IOP Publishing 01.05.2014
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Dynamic vision sensors (DVSs) encode visual input as a stream of events generated upon relative light intensity changes in the scene. These sensors have the advantage of allowing simultaneously high temporal resolution (better than 10 µs) and wide dynamic range (>120 dB) at sparse data representation, which is not possible with clocked vision sensors. In this paper, we focus on the task of stereo reconstruction. The spatiotemporal and asynchronous aspects of data provided by the sensor impose a different stereo reconstruction approach from the one applied for synchronous frame-based cameras. We propose to model the event-driven stereo matching by a cooperative network (Marr and Poggio 1976 Science 194 283-7). The history of the recent activity in the scene is stored in the network, which serves as spatiotemporal context used in disparity calculation for each incoming event. The network constantly evolves in time, as events are generated. In our work, not only the spatiotemporal aspect of the data is preserved but also the matching is performed asynchronously. The results of the experiments prove that the proposed approach is well adapted for DVS data and can be successfully used for disparity calculation.
ISSN:0957-0233
1361-6501
DOI:10.1088/0957-0233/25/5/055108