LIPSFUS: A neuromorphic dataset for audio-visual sensory fusion of lip reading

This paper presents a sensory fusion neuromorphic dataset collected with precise temporal synchronization using a set of Address-Event-Representation sensors and tools. The target application is the lip reading of several keywords for different machine learning applications, such as digits, robotic...

Full description

Saved in:
Bibliographic Details
Main Authors Rios-Navarro, Antonio, Piñero-Fuentes, Enrique, Canas-Moreno, Salvador, Javed, Aqib, Harkin, Jin, Linares-Barranco, Alejandro
Format Journal Article
LanguageEnglish
Published 28.03.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:This paper presents a sensory fusion neuromorphic dataset collected with precise temporal synchronization using a set of Address-Event-Representation sensors and tools. The target application is the lip reading of several keywords for different machine learning applications, such as digits, robotic commands, and auxiliary rich phonetic short words. The dataset is enlarged with a spiking version of an audio-visual lip reading dataset collected with frame-based cameras. LIPSFUS is publicly available and it has been validated with a deep learning architecture for audio and visual classification. It is intended for sensory fusion architectures based on both artificial and spiking neural network algorithms.
DOI:10.48550/arxiv.2304.01080