A comprehensive multimodal dataset for contactless lip reading and acoustic analysis

Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech recognition. Our dataset paper aims to facilitate the enhancement and restoration of speech information from diverse data sources for speakers. In this paper, w...

Full description

Saved in:
Bibliographic Details
Published inScientific data Vol. 10; no. 1; pp. 895 - 17
Main Authors Ge, Yao, Tang, Chong, Li, Haobo, Chen, Zikang, Wang, Jingyan, Li, Wenda, Cooper, Jonathan, Chetty, Kevin, Faccio, Daniele, Imran, Muhammad, Abbasi, Qammer H.
Format Journal Article
LanguageEnglish
Published London Nature Publishing Group UK 13.12.2023
Nature Publishing Group
Nature Portfolio
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech recognition. Our dataset paper aims to facilitate the enhancement and restoration of speech information from diverse data sources for speakers. In this paper, we introduce a novel multimodal dataset based on Radio Frequency, visual, text, audio, laser and lip landmark information, also called RVTALL. Specifically, the dataset consists of 7.5  GHz Channel Impulse Response (CIR) data from ultra-wideband (UWB) radars, 77  GHz frequency modulated continuous wave (FMCW) data from millimeter wave (mmWave) radar, visual and audio information, lip landmarks and laser data, offering a unique multimodal approach to speech recognition research. Meanwhile, a depth camera is adopted to record the landmarks of the subject’s lip and voice. Approximately 400 minutes of annotated speech profiles are provided, which are collected from 20 participants speaking 5 vowels, 15 words, and 16 sentences. The dataset has been validated and has potential for the investigation of lip reading and multimodal speech recognition.
AbstractList Abstract Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech recognition. Our dataset paper aims to facilitate the enhancement and restoration of speech information from diverse data sources for speakers. In this paper, we introduce a novel multimodal dataset based on Radio Frequency, visual, text, audio, laser and lip landmark information, also called RVTALL. Specifically, the dataset consists of 7.5 GHz Channel Impulse Response (CIR) data from ultra-wideband (UWB) radars, 77 GHz frequency modulated continuous wave (FMCW) data from millimeter wave (mmWave) radar, visual and audio information, lip landmarks and laser data, offering a unique multimodal approach to speech recognition research. Meanwhile, a depth camera is adopted to record the landmarks of the subject’s lip and voice. Approximately 400 minutes of annotated speech profiles are provided, which are collected from 20 participants speaking 5 vowels, 15 words, and 16 sentences. The dataset has been validated and has potential for the investigation of lip reading and multimodal speech recognition.
Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech recognition. Our dataset paper aims to facilitate the enhancement and restoration of speech information from diverse data sources for speakers. In this paper, we introduce a novel multimodal dataset based on Radio Frequency, visual, text, audio, laser and lip landmark information, also called RVTALL. Specifically, the dataset consists of 7.5 GHz Channel Impulse Response (CIR) data from ultra-wideband (UWB) radars, 77 GHz frequency modulated continuous wave (FMCW) data from millimeter wave (mmWave) radar, visual and audio information, lip landmarks and laser data, offering a unique multimodal approach to speech recognition research. Meanwhile, a depth camera is adopted to record the landmarks of the subject's lip and voice. Approximately 400 minutes of annotated speech profiles are provided, which are collected from 20 participants speaking 5 vowels, 15 words, and 16 sentences. The dataset has been validated and has potential for the investigation of lip reading and multimodal speech recognition.
Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech recognition. Our dataset paper aims to facilitate the enhancement and restoration of speech information from diverse data sources for speakers. In this paper, we introduce a novel multimodal dataset based on Radio Frequency, visual, text, audio, laser and lip landmark information, also called RVTALL. Specifically, the dataset consists of 7.5  GHz Channel Impulse Response (CIR) data from ultra-wideband (UWB) radars, 77  GHz frequency modulated continuous wave (FMCW) data from millimeter wave (mmWave) radar, visual and audio information, lip landmarks and laser data, offering a unique multimodal approach to speech recognition research. Meanwhile, a depth camera is adopted to record the landmarks of the subject’s lip and voice. Approximately 400 minutes of annotated speech profiles are provided, which are collected from 20 participants speaking 5 vowels, 15 words, and 16 sentences. The dataset has been validated and has potential for the investigation of lip reading and multimodal speech recognition.
Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech recognition. Our dataset paper aims to facilitate the enhancement and restoration of speech information from diverse data sources for speakers. In this paper, we introduce a novel multimodal dataset based on Radio Frequency, visual, text, audio, laser and lip landmark information, also called RVTALL. Specifically, the dataset consists of 7.5 GHz Channel Impulse Response (CIR) data from ultra-wideband (UWB) radars, 77 GHz frequency modulated continuous wave (FMCW) data from millimeter wave (mmWave) radar, visual and audio information, lip landmarks and laser data, offering a unique multimodal approach to speech recognition research. Meanwhile, a depth camera is adopted to record the landmarks of the subject's lip and voice. Approximately 400 minutes of annotated speech profiles are provided, which are collected from 20 participants speaking 5 vowels, 15 words, and 16 sentences. The dataset has been validated and has potential for the investigation of lip reading and multimodal speech recognition.Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech recognition. Our dataset paper aims to facilitate the enhancement and restoration of speech information from diverse data sources for speakers. In this paper, we introduce a novel multimodal dataset based on Radio Frequency, visual, text, audio, laser and lip landmark information, also called RVTALL. Specifically, the dataset consists of 7.5 GHz Channel Impulse Response (CIR) data from ultra-wideband (UWB) radars, 77 GHz frequency modulated continuous wave (FMCW) data from millimeter wave (mmWave) radar, visual and audio information, lip landmarks and laser data, offering a unique multimodal approach to speech recognition research. Meanwhile, a depth camera is adopted to record the landmarks of the subject's lip and voice. Approximately 400 minutes of annotated speech profiles are provided, which are collected from 20 participants speaking 5 vowels, 15 words, and 16 sentences. The dataset has been validated and has potential for the investigation of lip reading and multimodal speech recognition.
ArticleNumber 895
Author Chen, Zikang
Cooper, Jonathan
Chetty, Kevin
Faccio, Daniele
Imran, Muhammad
Tang, Chong
Abbasi, Qammer H.
Ge, Yao
Li, Haobo
Wang, Jingyan
Li, Wenda
Author_xml – sequence: 1
  givenname: Yao
  orcidid: 0000-0001-6314-3891
  surname: Ge
  fullname: Ge, Yao
  organization: James Watt School of Engineering, University of Glasgow
– sequence: 2
  givenname: Chong
  surname: Tang
  fullname: Tang, Chong
  organization: James Watt School of Engineering, University of Glasgow, Department of Security and Crime Science, University College London
– sequence: 3
  givenname: Haobo
  surname: Li
  fullname: Li, Haobo
  organization: School of Physics & Astronomy, University of Glasgow
– sequence: 4
  givenname: Zikang
  surname: Chen
  fullname: Chen, Zikang
  organization: James Watt School of Engineering, University of Glasgow
– sequence: 5
  givenname: Jingyan
  surname: Wang
  fullname: Wang, Jingyan
  organization: James Watt School of Engineering, University of Glasgow
– sequence: 6
  givenname: Wenda
  orcidid: 0000-0001-6617-9136
  surname: Li
  fullname: Li, Wenda
  organization: School of Science and Engineering, University of Dundee
– sequence: 7
  givenname: Jonathan
  surname: Cooper
  fullname: Cooper, Jonathan
  organization: James Watt School of Engineering, University of Glasgow
– sequence: 8
  givenname: Kevin
  surname: Chetty
  fullname: Chetty, Kevin
  organization: Department of Security and Crime Science, University College London
– sequence: 9
  givenname: Daniele
  orcidid: 0000-0001-8397-334X
  surname: Faccio
  fullname: Faccio, Daniele
  organization: School of Physics & Astronomy, University of Glasgow
– sequence: 10
  givenname: Muhammad
  orcidid: 0000-0003-4743-9136
  surname: Imran
  fullname: Imran, Muhammad
  organization: James Watt School of Engineering, University of Glasgow
– sequence: 11
  givenname: Qammer H.
  surname: Abbasi
  fullname: Abbasi, Qammer H.
  email: qammer.abbasi@glasgow.ac.uk
  organization: James Watt School of Engineering, University of Glasgow
BackLink https://www.ncbi.nlm.nih.gov/pubmed/38092796$$D View this record in MEDLINE/PubMed
BookMark eNp9Uk1vFSEUJabG1to_4MJM4sbNKFxgGFamafxo0sRNXRMGmFdeGHjCTJv-e3lvam1duLgBLuccTi7nNTqKKTqE3hL8kWDafyqMcClaDLSWkLS9e4FOAHNoGevo0ZP9MTorZYsxJpRhLvArdEx7LCupO0HX541J0y67GxeLv3XNtITZT8nq0Fg96-LmZky5guKszRxcKU3wuyY7bX3cNDraRpu0lNmbetDhvvjyBr0cdSju7GE9RT-_frm--N5e_fh2eXF-1RomYW5J37ne9MJaKxkQzEdrRiaosQADFlQy0tW-Bg5CV78DZ9CBGGHEo-NC0FN0uerapLdql_2k871K2qtDI-WN0rkaC045aztN5AACKBsAayk4AYqrNun7gVetz6vWbhkmZ42Lc9bhmejzm-hv1CbdKoIFkdD1VeHDg0JOvxZXZjX5YlwIOro6IAUSg-RMAKnQ9_9At2nJdXoHFNlXvxd899TSo5c_v1cBsAJMTqVkNz5CCFb7lKg1JaqmRB1Sou4qia6kUsFx4_Lft__D-g26074j
Cites_doi 10.1109/JIOT.2022.3217968
10.1038/s41467-022-32231-1
10.1007/s10579-020-09503-7
10.1121/1.2229005
10.3390/app11219870
10.1109/TASLP.2023.3250846
10.3390/s23010330
10.1364/BOE.451416
10.1007/s10579-008-9076-6
10.1016/j.neuroimage.2019.116283
10.1109/ACCESS.2020.3026579
10.6084/m9.figshare.c.6942846.v1
10.3390/s22020649
10.1109/JSEN.2019.2946095
10.1109/TMTT.2014.2320464
10.1109/JSEN.2018.2877043
10.1109/COMST.2022.3145856
10.3390/s21103465
10.1121/1.397959
10.1371/journal.pone.0196391
10.1016/j.procs.2021.12.222
10.1145/3277883.3277884
10.1145/3539490.3539599
10.1016/j.specom.2007.02.006
10.1109/TNSRE.2023.3266299
10.1145/3447993.3448626
10.1109/SLT48900.2021.9383619
10.1145/3397321
10.1145/3485730.3485945
10.1109/CVPR46437.2021.00500
ContentType Journal Article
Copyright The Author(s) 2023
2023. The Author(s).
The Author(s) 2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Copyright_xml – notice: The Author(s) 2023
– notice: 2023. The Author(s).
– notice: The Author(s) 2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
DBID C6C
AAYXX
CITATION
NPM
3V.
7X7
7XB
88E
8FE
8FH
8FI
8FJ
8FK
ABUWG
AFKRA
AZQEC
BBNVY
BENPR
BHPHI
CCPQU
DWQXO
FYUFA
GHDGH
GNUQQ
HCIFZ
K9.
LK8
M0S
M1P
M7P
PHGZM
PHGZT
PIMPY
PJZUB
PKEHL
PPXIY
PQEST
PQGLB
PQQKQ
PQUKI
PRINS
7X8
5PM
DOA
DOI 10.1038/s41597-023-02793-w
DatabaseName Springer Nature OA Free Journals
CrossRef
PubMed
ProQuest Central (Corporate)
ProQuest Health & Medical Collection (NC LIVE)
ProQuest Central (purchase pre-March 2016)
Medical Database (Alumni Edition)
ProQuest SciTech Collection
ProQuest Natural Science Collection
Hospital Premium Collection
Hospital Premium Collection (Alumni Edition)
ProQuest Central (Alumni) (purchase pre-March 2016)
ProQuest Central (Alumni)
ProQuest Central UK/Ireland
ProQuest Central Essentials
Biological Science Collection
ProQuest Central
Natural Science Collection
ProQuest One Community College
ProQuest Central Korea
Health Research Premium Collection
Health Research Premium Collection (Alumni)
ProQuest Central Student
SciTech Premium Collection (via ProQuest)
ProQuest Health & Medical Complete (Alumni)
ProQuest Biological Science Collection
ProQuest Health & Medical Collection
Medical Database
Biological Science Database
ProQuest Central Premium
ProQuest One Academic
Publicly Available Content Database
ProQuest Health & Medical Research Collection
ProQuest One Academic Middle East (New)
ProQuest One Health & Nursing
ProQuest One Academic Eastern Edition (DO NOT USE)
ProQuest One Applied & Life Sciences
ProQuest One Academic
ProQuest One Academic UKI Edition
ProQuest Central China
MEDLINE - Academic
PubMed Central (Full Participant titles)
DOAJ Directory of Open Access Journals
DatabaseTitle CrossRef
PubMed
Publicly Available Content Database
ProQuest Central Student
ProQuest One Academic Middle East (New)
ProQuest Central Essentials
ProQuest Health & Medical Complete (Alumni)
ProQuest Central (Alumni Edition)
SciTech Premium Collection
ProQuest One Community College
ProQuest One Health & Nursing
ProQuest Natural Science Collection
ProQuest Central China
ProQuest Central
ProQuest One Applied & Life Sciences
ProQuest Health & Medical Research Collection
Health Research Premium Collection
Health and Medicine Complete (Alumni Edition)
Natural Science Collection
ProQuest Central Korea
Health & Medical Research Collection
Biological Science Collection
ProQuest Central (New)
ProQuest Medical Library (Alumni)
ProQuest Biological Science Collection
ProQuest One Academic Eastern Edition
ProQuest Hospital Collection
Health Research Premium Collection (Alumni)
Biological Science Database
ProQuest SciTech Collection
ProQuest Hospital Collection (Alumni)
ProQuest Health & Medical Complete
ProQuest Medical Library
ProQuest One Academic UKI Edition
ProQuest One Academic
ProQuest One Academic (New)
ProQuest Central (Alumni)
MEDLINE - Academic
DatabaseTitleList
PubMed
Publicly Available Content Database

CrossRef
MEDLINE - Academic

Database_xml – sequence: 1
  dbid: C6C
  name: Springer Nature OA Free Journals
  url: http://www.springeropen.com/
  sourceTypes: Publisher
– sequence: 2
  dbid: DOA
  name: DOAJ Directory of Open Access Journals
  url: https://www.doaj.org/
  sourceTypes: Open Website
– sequence: 3
  dbid: NPM
  name: PubMed
  url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed
  sourceTypes: Index Database
– sequence: 4
  dbid: BENPR
  name: ProQuest Central
  url: https://www.proquest.com/central
  sourceTypes: Aggregation Database
DeliveryMethod fulltext_linktorsrc
Discipline Sciences (General)
EISSN 2052-4463
EndPage 17
ExternalDocumentID oai_doaj_org_article_edd6a19b27234b20a9751230252188b5
PMC10719268
38092796
10_1038_s41597_023_02793_w
Genre Journal Article
GrantInformation_xml – fundername: RCUK | Engineering and Physical Sciences Research Council (EPSRC)
  grantid: EP/T021063/1; EP/W003228/1; EP/T021020/1
  funderid: https://doi.org/10.13039/501100000266
– fundername: Royal Society of Edinburgh (RSE)
  grantid: SAPHIRE
  funderid: https://doi.org/10.13039/501100000332
– fundername: RCUK | Engineering and Physical Sciences Research Council (EPSRC)
  grantid: EP/T021063/1
– fundername: RCUK | Engineering and Physical Sciences Research Council (EPSRC)
  grantid: EP/T021020/1
– fundername: RCUK | Engineering and Physical Sciences Research Council (EPSRC)
  grantid: EP/W003228/1
– fundername: Royal Society of Edinburgh (RSE)
  grantid: SAPHIRE
GroupedDBID 0R~
3V.
53G
5VS
7X7
88E
8FE
8FH
8FI
8FJ
AAJSJ
ABUWG
ACGFS
ACSFO
ACSMW
ADBBV
ADRAZ
AFKRA
AGHDO
AJTQC
ALIPV
ALMA_UNASSIGNED_HOLDINGS
AOIJS
BBNVY
BCNDV
BENPR
BHPHI
BPHCQ
BVXVI
C6C
CCPQU
DIK
EBLON
EBS
EJD
FYUFA
GROUPED_DOAJ
HCIFZ
HMCUK
HYE
KQ8
LK8
M1P
M48
M7P
M~E
NAO
OK1
PGMZT
PIMPY
PQQKQ
PROAC
PSQYO
RNT
RNTTT
RPM
SNYQT
UKHRP
AASML
AAYXX
CITATION
PHGZM
PHGZT
NPM
7XB
8FK
AARCD
AZQEC
DWQXO
GNUQQ
K9.
PJZUB
PKEHL
PPXIY
PQEST
PQGLB
PQUKI
PRINS
7X8
5PM
PUEGO
ID FETCH-LOGICAL-c492t-186e8c87ddd942105fdcf473cd22b0739416421a2527a927b542627f2f0fe5773
IEDL.DBID 7X7
ISSN 2052-4463
IngestDate Wed Aug 27 01:07:33 EDT 2025
Thu Aug 21 18:38:00 EDT 2025
Fri Jul 11 03:10:29 EDT 2025
Wed Aug 13 04:20:35 EDT 2025
Thu Apr 03 07:06:18 EDT 2025
Tue Jul 01 00:39:01 EDT 2025
Fri Feb 21 02:39:03 EST 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed true
IsScholarly true
Issue 1
Language English
License 2023. The Author(s).
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c492t-186e8c87ddd942105fdcf473cd22b0739416421a2527a927b542627f2f0fe5773
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ObjectType-Article-2
ObjectType-Undefined-1
ObjectType-Feature-3
content type line 23
ORCID 0000-0001-8397-334X
0000-0003-4743-9136
0000-0001-6314-3891
0000-0001-6617-9136
OpenAccessLink https://www.proquest.com/docview/2901290188?pq-origsite=%requestingapplication%
PMID 38092796
PQID 2901290188
PQPubID 2041912
PageCount 17
ParticipantIDs doaj_primary_oai_doaj_org_article_edd6a19b27234b20a9751230252188b5
pubmedcentral_primary_oai_pubmedcentral_nih_gov_10719268
proquest_miscellaneous_2902954721
proquest_journals_2901290188
pubmed_primary_38092796
crossref_primary_10_1038_s41597_023_02793_w
springer_journals_10_1038_s41597_023_02793_w
ProviderPackageCode CITATION
AAYXX
PublicationCentury 2000
PublicationDate 2023-12-13
PublicationDateYYYYMMDD 2023-12-13
PublicationDate_xml – month: 12
  year: 2023
  text: 2023-12-13
  day: 13
PublicationDecade 2020
PublicationPlace London
PublicationPlace_xml – name: London
– name: England
PublicationTitle Scientific data
PublicationTitleAbbrev Sci Data
PublicationTitleAlternate Sci Data
PublicationYear 2023
Publisher Nature Publishing Group UK
Nature Publishing Group
Nature Portfolio
Publisher_xml – name: Nature Publishing Group UK
– name: Nature Publishing Group
– name: Nature Portfolio
References Futrell (CR30) 2021; 55
CR14
Ferreira, Silva, Curado, Teixeira (CR8) 2022; 22
CR13
CR12
CR11
Livingstone, Russo (CR21) 2018; 13
Gonzalez-Lopez, Gomez-Alanis, Martín Doñas, Pérez-Córdoba, Gomez (CR3) 2020; 8
Cooke, Barker, Cunningham, Shao (CR27) 2006; 120
Titze (CR16) 1989; 85
Ge (CR17) 2023
Abdrakhmanova (CR22) 2021; 21
Wang, Zhang, Wu, Kong, Zhang (CR25) 2021; 11
Cester (CR10) 2022; 13
CR2
Hao, Zhu, Wang, Yang, Zeng (CR20) 2023; 23
Ozturk, Wu, Wang, Liu (CR6) 2023; 10
Bednar, Lalor (CR4) 2020; 205
CR5
Wang, Muñoz-Ferreras, Gu, Li, Gómez-García (CR19) 2014; 62
Busso (CR28) 2008; 42
CR29
CR9
CR26
Cai, Zheng, Luo (CR1) 2022; 24
CR23
Hameed (CR7) 2022; 13
Li, Shrestha, Heidari, Le Kernec, Fioranelli (CR15) 2020; 20
Kim, Lee (CR18) 2019; 19
Ozturk, Wu, Wang, Wu, Liu (CR24) 2023; 31
JA Gonzalez-Lopez (2793_CR3) 2020; 8
M Abdrakhmanova (2793_CR22) 2021; 21
A Bednar (2793_CR4) 2020; 205
S Kim (2793_CR18) 2019; 19
D Ferreira (2793_CR8) 2022; 22
Y Wang (2793_CR25) 2021; 11
2793_CR11
L Cester (2793_CR10) 2022; 13
Y Ge (2793_CR17) 2023
H Hameed (2793_CR7) 2022; 13
2793_CR14
2793_CR13
2793_CR12
G Wang (2793_CR19) 2014; 62
C Cai (2793_CR1) 2022; 24
C Busso (2793_CR28) 2008; 42
X Hao (2793_CR20) 2023; 23
M Cooke (2793_CR27) 2006; 120
MZ Ozturk (2793_CR24) 2023; 31
MZ Ozturk (2793_CR6) 2023; 10
S Livingstone (2793_CR21) 2018; 13
2793_CR26
IR Titze (2793_CR16) 1989; 85
2793_CR2
2793_CR23
2793_CR29
R Futrell (2793_CR30) 2021; 55
2793_CR5
H Li (2793_CR15) 2020; 20
2793_CR9
References_xml – volume: 10
  start-page: 4431
  year: 2023
  end-page: 4448
  ident: CR6
  article-title: RadioMic: Sound sensing via radio signals
  publication-title: IEEE Internet of Things Journal
  doi: 10.1109/JIOT.2022.3217968
– volume: 13
  start-page: 1
  year: 2022
  end-page: 9
  ident: CR7
  article-title: Pushing the limits of remote RF sensing by reading lips under the face mask
  publication-title: Nature communications
  doi: 10.1038/s41467-022-32231-1
– ident: CR14
– volume: 55
  start-page: 63
  year: 2021
  end-page: 77
  ident: CR30
  article-title: The natural stories corpus: a reading-time corpus of english texts containing rare syntactic constructions
  publication-title: Language Resources and Evaluation
  doi: 10.1007/s10579-020-09503-7
– ident: CR2
– ident: CR12
– volume: 120
  start-page: 2421
  year: 2006
  end-page: 2424
  ident: CR27
  article-title: An audio-visual corpus for speech perception and automatic speech recognition
  publication-title: The Journal of the Acoustical Society of America
  doi: 10.1121/1.2229005
– volume: 11
  start-page: 9870
  year: 2021
  ident: CR25
  article-title: Speckle noise detection and removal for laser speech measurement systems
  publication-title: Applied Sciences
  doi: 10.3390/app11219870
– ident: CR29
– volume: 31
  start-page: 1333
  year: 2023
  end-page: 1347
  ident: CR24
  article-title: Radioses: mmWave-based audioradio speech enhancement and separation system
  publication-title: IEEE/ACM Transactions on Audio, Speech, and Language Processing
  doi: 10.1109/TASLP.2023.3250846
– volume: 23
  start-page: 330
  year: 2023
  ident: CR20
  article-title: A speech enhancement algorithm for speech reconstruction based on laser speckle images
  publication-title: Sensors
  doi: 10.3390/s23010330
– ident: CR23
– volume: 13
  start-page: 3743
  year: 2022
  end-page: 3750
  ident: CR10
  article-title: Remote laser-speckle sensing of heart sounds for health assessment and biometric identification
  publication-title: Biomedical Optics Express
  doi: 10.1364/BOE.451416
– volume: 42
  start-page: 335
  year: 2008
  end-page: 359
  ident: CR28
  article-title: Iemocap: Interactive emotional dyadic motion capture database
  publication-title: Language resources and evaluation
  doi: 10.1007/s10579-008-9076-6
– volume: 205
  start-page: 116283
  year: 2020
  ident: CR4
  article-title: Where is the cocktail party? Decoding locations of attended and unattended moving sound sources using EEG
  publication-title: NeuroImage
  doi: 10.1016/j.neuroimage.2019.116283
– volume: 8
  start-page: 177995
  year: 2020
  end-page: 178021
  ident: CR3
  article-title: Silent speech interfaces for speech restoration: A Review
  publication-title: IEEE Access
  doi: 10.1109/ACCESS.2020.3026579
– year: 2023
  ident: CR17
  publication-title: figshare
  doi: 10.6084/m9.figshare.c.6942846.v1
– volume: 22
  start-page: 649
  year: 2022
  ident: CR8
  article-title: Exploring silent speech interfaces based on frequency-modulated continuous-wave radar
  publication-title: Sensors
  doi: 10.3390/s22020649
– volume: 20
  start-page: 1191
  year: 2020
  end-page: 1201
  ident: CR15
  article-title: Bi-LSTM network for multimodal continuous human activity recognition and fall detection
  publication-title: IEEE Sensors Journal
  doi: 10.1109/JSEN.2019.2946095
– volume: 62
  start-page: 1387
  year: 2014
  end-page: 1399
  ident: CR19
  article-title: Application of linear-frequency-modulated continuous-wave (lfmcw) radars for tracking of vital signs
  publication-title: IEEE Transactions on Microwave Theory and Techniques
  doi: 10.1109/TMTT.2014.2320464
– volume: 19
  start-page: 2205
  year: 2019
  end-page: 2216
  ident: CR18
  article-title: Low-Complexity joint extrapolation-MUSIC-based 2-D parameter estimator for vital FMCW radar
  publication-title: IEEE Sensors Journal
  doi: 10.1109/JSEN.2018.2877043
– ident: CR13
– volume: 24
  start-page: 432
  year: 2022
  end-page: 454
  ident: CR1
  article-title: Ubiquitous acoustic sensing on commodity IoT devices: A Survey
  publication-title: IEEE Communications Surveys Tutorials
  doi: 10.1109/COMST.2022.3145856
– ident: CR11
– volume: 21
  start-page: 3465
  year: 2021
  ident: CR22
  article-title: Speakingfaces: A large-scale multimodal dataset of voice commands with visual and thermal video streams
  publication-title: Sensors
  doi: 10.3390/s21103465
– ident: CR9
– ident: CR5
– volume: 85
  start-page: 1699
  year: 1989
  end-page: 1707
  ident: CR16
  article-title: Physiologic and acoustic differences between male and female voices
  publication-title: The Journal of the Acoustical Society of America
  doi: 10.1121/1.397959
– ident: CR26
– volume: 13
  start-page: e0196391
  year: 2018
  ident: CR21
  article-title: Ryerson audiovisual database of emotional speeches and songs (ravdess): a dynamic, multimodal set of north american english face and voice expressions
  publication-title: Plos One
  doi: 10.1371/journal.pone.0196391
– ident: 2793_CR12
  doi: 10.1016/j.procs.2021.12.222
– volume: 13
  start-page: 1
  year: 2022
  ident: 2793_CR7
  publication-title: Nature communications
  doi: 10.1038/s41467-022-32231-1
– volume: 55
  start-page: 63
  year: 2021
  ident: 2793_CR30
  publication-title: Language Resources and Evaluation
  doi: 10.1007/s10579-020-09503-7
– ident: 2793_CR13
  doi: 10.1145/3277883.3277884
– volume: 21
  start-page: 3465
  year: 2021
  ident: 2793_CR22
  publication-title: Sensors
  doi: 10.3390/s21103465
– volume: 19
  start-page: 2205
  year: 2019
  ident: 2793_CR18
  publication-title: IEEE Sensors Journal
  doi: 10.1109/JSEN.2018.2877043
– ident: 2793_CR9
  doi: 10.1145/3539490.3539599
– volume: 8
  start-page: 177995
  year: 2020
  ident: 2793_CR3
  publication-title: IEEE Access
  doi: 10.1109/ACCESS.2020.3026579
– volume: 85
  start-page: 1699
  year: 1989
  ident: 2793_CR16
  publication-title: The Journal of the Acoustical Society of America
  doi: 10.1121/1.397959
– volume: 42
  start-page: 335
  year: 2008
  ident: 2793_CR28
  publication-title: Language resources and evaluation
  doi: 10.1007/s10579-008-9076-6
– ident: 2793_CR2
  doi: 10.1016/j.specom.2007.02.006
– ident: 2793_CR26
  doi: 10.1109/TNSRE.2023.3266299
– volume: 13
  start-page: 3743
  year: 2022
  ident: 2793_CR10
  publication-title: Biomedical Optics Express
  doi: 10.1364/BOE.451416
– volume: 20
  start-page: 1191
  year: 2020
  ident: 2793_CR15
  publication-title: IEEE Sensors Journal
  doi: 10.1109/JSEN.2019.2946095
– ident: 2793_CR23
  doi: 10.1145/3447993.3448626
– volume: 11
  start-page: 9870
  year: 2021
  ident: 2793_CR25
  publication-title: Applied Sciences
  doi: 10.3390/app11219870
– ident: 2793_CR29
  doi: 10.1109/SLT48900.2021.9383619
– volume: 22
  start-page: 649
  year: 2022
  ident: 2793_CR8
  publication-title: Sensors
  doi: 10.3390/s22020649
– ident: 2793_CR14
  doi: 10.1145/3397321
– volume: 24
  start-page: 432
  year: 2022
  ident: 2793_CR1
  publication-title: IEEE Communications Surveys Tutorials
  doi: 10.1109/COMST.2022.3145856
– volume: 205
  start-page: 116283
  year: 2020
  ident: 2793_CR4
  publication-title: NeuroImage
  doi: 10.1016/j.neuroimage.2019.116283
– volume: 31
  start-page: 1333
  year: 2023
  ident: 2793_CR24
  publication-title: IEEE/ACM Transactions on Audio, Speech, and Language Processing
  doi: 10.1109/TASLP.2023.3250846
– ident: 2793_CR5
  doi: 10.1145/3485730.3485945
– year: 2023
  ident: 2793_CR17
  doi: 10.6084/m9.figshare.c.6942846.v1
– volume: 120
  start-page: 2421
  year: 2006
  ident: 2793_CR27
  publication-title: The Journal of the Acoustical Society of America
  doi: 10.1121/1.2229005
– volume: 23
  start-page: 330
  year: 2023
  ident: 2793_CR20
  publication-title: Sensors
  doi: 10.3390/s23010330
– ident: 2793_CR11
  doi: 10.1109/CVPR46437.2021.00500
– volume: 62
  start-page: 1387
  year: 2014
  ident: 2793_CR19
  publication-title: IEEE Transactions on Microwave Theory and Techniques
  doi: 10.1109/TMTT.2014.2320464
– volume: 10
  start-page: 4431
  year: 2023
  ident: 2793_CR6
  publication-title: IEEE Internet of Things Journal
  doi: 10.1109/JIOT.2022.3217968
– volume: 13
  start-page: e0196391
  year: 2018
  ident: 2793_CR21
  publication-title: Plos One
  doi: 10.1371/journal.pone.0196391
SSID ssj0001340570
Score 2.2920346
Snippet Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech recognition. Our...
Abstract Small-scale motion detection using non-invasive remote sensing techniques has recently garnered significant interest in the field of speech...
SourceID doaj
pubmedcentral
proquest
pubmed
crossref
springer
SourceType Open Website
Open Access Repository
Aggregation Database
Index Database
Publisher
StartPage 895
SubjectTerms 639/166/985
639/166/987
639/766/930/1032
Data Descriptor
Datasets
Humanities and Social Sciences
Lip
Lipreading
Motion detection
multidisciplinary
Remote sensing
Science
Science (multidisciplinary)
Speech
Speech recognition
Voice recognition
SummonAdditionalLinks – databaseName: DOAJ Directory of Open Access Journals
  dbid: DOA
  link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1NT9wwEB1VnLhUfLWEL7lSD60gIrGT2D5CVYSQ2hNI3Cw7dgQShNVmK_5-Z-zswlIQlx5yyJdivbHznjOTZ4CvoWkpf4PKrZBxCTOdW1faXOma-F5yGyij--t3c35VXVzX18-W-qKasGQPnIA7Dt43ttQObxOV44XVEjlKIFUjOSkX3UuR855NpuLXFUFCpBj_kimEOh6Qqch4lMe0pRb54xITRcP-11Tmv8WSLzKmkYjO1uDjqCDZSWr5OnwI_Qasj2N0YN9GI-nvm3B5wqhifBpuUpU6i9WD9w8eb6fK0CHMGGpWRuXqtsVOMwzs7nbCpqmwntneM3xhxvW-cCe5l2zB1dnPyx_n-biKQt5Wms_yUjVBtUp673WFE7y6821XSdF6zh3l6VCS4XGLgEqruXQ1mdTLjndFF2opxSdY6R_6sA0MX4xOIeUpJysktVp3KHCcrUPj6X_bOoPDOaJmkswyTExyC2US_gbxNxF_85jBKYG-uJKMruMBDL8Zw2_eC38Ge_OQmXH0DYZyw7QplcGXxWkcN5QMsX1A4OgaSnHiBDiDzynCi5YIVSAQuslALcV-qanLZ_rbm-jNjbNp1MwNPvho3k2e2vU2Fjv_A4tdWOXUv0uel2IPVmbTP2EfJdPMHcTR8RfJyQ3q
  priority: 102
  providerName: Directory of Open Access Journals
– databaseName: Scholars Portal Journals: Open Access
  dbid: M48
  link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1Nb9QwEB2V9sIFUcpHoCBX6oEKAomdxPYBoYJaVZXaU1fqzbJjh1Yq2ZIsKvx7Zpxk0cLCIYd8Kc7zOPOcN54B2A9VTfoNMrdMxhJmOrUut6nSJfl7yW0gRffsvDqZFaeX5eUGTOWORgD7tVM7qic1627e_fj28yMO-A_DknH1vkcnRDlFeVQktUjv7sEWeiZJFQ3ORrof_7kIoifZuHZm_a0r_imm8V_HPf8OofxDR43u6fghPBh5JTscDGEbNkL7CLbHkduz12N66YMduDhkFEfehashdp3FmMKvc4-3U7xoHxYMmSyjIHZboyn1Pbu5vmXdEG7PbOsZfkZjFTDcGXKaPIbZ8dHF55N0rK2Q1oXmizRXVVC1kt57XeC0r2x83RRS1J5zR-odEjU8bjn2l9VcupJS18uGN1kTSinFE9hs5214Bgw_l06hI1ROFujqSt0g7XG2DJWnVbhlAm8mRM3tkELDROlbKDPgbxB_E_E3dwl8ItCXV1L663hg3n0x42gywfvK5tqhLYnC8cxqicRFIH9DxqIcPnJ36jIzmZQhxZg2pRLYW57G0UQSiW0DAkfXkPCJ0-IEng49vGyJUBkCoasE1ErfrzR19Ux7fRUzduMcG5l0hQ9-O5nJ73b9G4vn_3-NF3Cfk-XmPM3FLmwuuu_hJVKkhXsV7f4X0q4KYw
  priority: 102
  providerName: Scholars Portal
– databaseName: Springer Nature HAS Fully OA
  dbid: AAJSJ
  link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV1NT9wwEB3xceGCChSa8iEjcWhVIrJ2EtvHbVWEVmovgMTNsmMHkNos2izi7zPjJIsW6KGHHBI7ijVje54zz88AJ6GsKH-DyC2T8QgznVo3sqnSBcV7yW2gjO6v3-XFdT65KW5WgA97YSJpP0paxml6YIedtRhoSDeUx6yjFunTKqyTVDv27fXxeHI5efmzIgiEZP0OmUyod15eikJRrP89hPmWKPkqWxqD0PkH2OzRIxt37d2CldBsw1Y_Plv2pReR_roDV2NGbPFZuOsY6iwyB_9OPb5OrNA2zBniVUZUdVthh2lb9uf-gc06Uj2zjWc4WcazvvCmUy75CNfnP69-XKT9CQpplWs-T0eqDKpS0nuvc1zcFbWv6lyKynPuKEeHcAyfW45esZpLV5BAvax5ndWhkFLswlozbcInYDgpOoXhTjmZY0ArdI3gxtkilJ722hYJfBssah46oQwTE9xCmc7-Bu1vov3NUwLfyeiLmiRyHR9MZ7emd7oJ3pd2pB32GJE7nlktEZ4IRGmIS5TDTx4MLjP9yGsN5YXpUiqB40UxjhlKhNgmoOGoDqU3cfGbwF7n4UVLhMrQELpMQC35fqmpyyXN_V3U5caVNOLlEj98OnSTl3b92xaf_6_6PmzQkfdEqRmJA1ibzx7DIQKjuTvqR8Izc9gGpQ
  priority: 102
  providerName: Springer Nature
Title A comprehensive multimodal dataset for contactless lip reading and acoustic analysis
URI https://link.springer.com/article/10.1038/s41597-023-02793-w
https://www.ncbi.nlm.nih.gov/pubmed/38092796
https://www.proquest.com/docview/2901290188
https://www.proquest.com/docview/2902954721
https://pubmed.ncbi.nlm.nih.gov/PMC10719268
https://doaj.org/article/edd6a19b27234b20a9751230252188b5
Volume 10
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV3fb9MwED6N7YUXxBg_AqMyEg8giJbGSWw_oa7aNFXahGCT-mbZsbNNgrQ0Rfv3uXPcVoWNh6aqkyrO3dn32d_lDuC9r2ribxC5ZSKUMFOpsUOTSlWSvxe58cTonl9UZ1fFZFpO44ZbF8MqV3NimKjdrKY98iPi--gj5Zf5r5SqRhG7GktoPII9Sl1GVi2mYrPHwgmOZPFdmYzLow79FaUfzQN5qXh6t-WPQtr--7DmvyGTf_GmwR2dPoUnEUeyUa_4fdjx7TPYjyO1Yx9iOumPB3A5YhQ3vvA3faw6CzGEP2cO_07xoZ1fMkSujILWTY2m03Xsx-2cLfrwemZax3DaDFW_8Eefw-Q5XJ2eXI7P0lhLIa0LlS_Toay8rKVwzqkCl3ll4-qmELx2eW6JrUNghu0mR_0YlQtbUqp60eRN1vhSCP4CdttZ618Bw-nRSnR80ooCXVupGoQ51pS-cvTWbZnAp5VE9bxPmaED1c2l7uWvUf46yF_fJXBMQl9fSemuQ8Nsca3j6NHeucoMlUXb4YXNM6MEAhWOeA0RirR4y8OVynQcg53eWEwC79ancfQQJWJaj4Kja4joxGVwAi97Da97wmWGglBVAnJL91td3T7T3t6EDN24pkbkXOGNP6_MZNOvh2Xx-v-P8QYeU7F7CqYZ8kPYXS5--7cIiZZ2EOx-AHuj0eT7BL-PTy6-fsPWcTUehG0GPJ4X8g8hbwzC
linkProvider ProQuest
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV1Lb9NAEB6VcoALorzqUmCRQAKBVWf92N0DQuVRpfRxSqXcll3vmlYCJ42DIv4Uv5GZtZ0ovG495BDbidez8_jW3-wMwDNflMTfIHJLRGhhpmJjByaWKqd4L7jxxOienBbDs-zTOB9vwM9-LwylVfY-MThqNynpHfke8X30kfLt9DKmrlHErvYtNFq1OPI_Frhka94cfsD5fc75wcfR-2HcdRWIy0zxeTyQhZelFM45leGCJ69cWWUiLR3nlngrhCh43HAcqVFc2JyKtouKV0nlcyFS_N9rcB0Db0KLPTEWq3c6KcGfpNubk6Ryr8H4SOVOeSBLVRov1uJfaBPwN2z7Z4rmbzxtCH8Ht-FWh1vZfqtoW7Dh6zuw1XmGhr3oyle_vAujfUZ56jN_3ubGs5Cz-G3i8OeUj9r4OUOkzChJ3pSoqk3Dvl5M2axN52emdgzddOgyhl_amin34OxKpHwfNutJ7beBoTu2EgOttCLDUJqrCmGVNbkvHO3yzSN41UtUT9sSHTpQ66nUrfw1yl8H-etFBO9I6Msrqbx2ODCZfdGdtWrvXGEGyqKuppnliVECgVGK-BARkbR4y91-ynRn841eaWgET5en0VqJgjG1R8HRNUSs4rI7ggftDC9HksoEBaGKCOTa3K8Ndf1MfXEeKoKjKiJSL_DGr3s1WY3r37LY-f9jPIEbw9HJsT4-PD16CDc5afGAx4N0Fzbns-_-EcKxuX0cbIDB56s2ul-SDUHH
linkToPdf http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV3db9MwED-NTkK8IMZnxgAjgQSCaInzYecBoY2t2hhUE9qkvRk7dtikkZamqOJf46_jzklala-3PfShSdo45_Pd7_I73wE8c3lJ_A0it0j4FmZFqE2sQ1lk5O8F144Y3Y-j_OA0fX-Wna3Bz34vDKVV9jbRG2o7Lukd-TbxffTBgK3q0iKO94ZvJ99C6iBFTGvfTqNVkSP3Y47hW_PmcA_n-jnnw_2Tdwdh12EgLNOCz8JY5k6WUlhrixSDn6yyZZWKpLScG-KwEK7gcc1x1LrgwmRUwF1UvIoqlwmR4P9eg3VBUdEA1nf3R8eflm94EgJDUbdTJ0rkdoPekoqfck-dFkk4X_GGvmnA35Dunwmbv7G23hkOb8HNDsWynVbtNmDN1bdho7MTDXvRFbN-eQdOdhhlrU_deZspz3wG49exxZ9TdmrjZgxxM6OUeV2i4jYNu7yYsGmb3M90bRkabd9zDL-0FVTuwumVyPkeDOpx7R4AQ-NsJLpdaUSKjjUrKgRZRmcut7TnNwvgVS9RNWkLdihPtCdStfJXKH_l5a_mAeyS0BdXUrFtf2A8_aK6tauctbmOC4Oam6SGR7oQCJMSRIuIj6TBW271U6Y6C9Copb4G8HRxGtcuETK6dig4uoZoVgzCA7jfzvBiJImMUBBFHoBcmfuVoa6eqS_OfX1wjOgRt-d449e9mizH9W9ZbP7_MZ7AdVxw6sPh6Ogh3OCkxDEP42QLBrPpd_cIsdnMPO4WAYPPV73ufgFQO0di
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=A+comprehensive+multimodal+dataset+for+contactless+lip+reading+and+acoustic+analysis&rft.jtitle=Scientific+data&rft.au=Ge%2C+Yao&rft.au=Tang%2C+Chong&rft.au=Li%2C+Haobo&rft.au=Chen%2C+Zikang&rft.date=2023-12-13&rft.pub=Nature+Publishing+Group&rft.eissn=2052-4463&rft.volume=10&rft.issue=1&rft.spage=895&rft_id=info:doi/10.1038%2Fs41597-023-02793-w&rft.externalDBID=HAS_PDF_LINK
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2052-4463&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2052-4463&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2052-4463&client=summon