An integrated framework for flood disaster information extraction and analysis leveraging social media data: A case study of the Shouguang flood in China

Flood disasters cause significant casualties and economic losses annually worldwide. During disasters, accurate and timely information is crucial for disaster management. However, remote sensing cannot balance temporal and spatial resolution, and the coverage of specialized equipment is limited, mak...

Full description

Saved in:
Bibliographic Details
Published inThe Science of the total environment Vol. 949; p. 174948
Main Authors Hou, Huawei, Shen, Li, Jia, Jianan, Xu, Zhu
Format Journal Article
LanguageEnglish
Published Netherlands Elsevier B.V 01.11.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Flood disasters cause significant casualties and economic losses annually worldwide. During disasters, accurate and timely information is crucial for disaster management. However, remote sensing cannot balance temporal and spatial resolution, and the coverage of specialized equipment is limited, making continuous monitoring challenging. Real-time disaster-related information shared by social media users offers new possibilities for monitoring. We propose a framework for extracting and analyzing flood information from social media, validated through the 2018 Shouguang flood in China. This framework innovatively combines deep learning techniques and regular expression matching techniques to automatically extract key flood-related information from Weibo textual data, such as problems, floodings, needs, rescues, and measures, achieving an accuracy of 83 %, surpassing traditional models like the Biterm Topic Model (BTM). In the spatiotemporal analysis of the disaster, our research identifies critical time points during the disaster through quantitative analysis of the information and explores the spatial distribution of calls for help using Kernel Density Estimation (KDE), followed by identifying the core affected areas using the Hierarchical Density-Based Spatial Clustering of Applications with Noise (HDBSCAN) algorithm. For semantic analysis, we adopt the Latent Dirichlet Allocation (LDA) algorithm to perform topic modeling on Weibo texts from different regions, identifying the types of disasters affecting each township. Additionally, through correlation analysis, we investigate the relationship between disaster rescue requests and response measures to evaluate the adequacy of flood response measures in each township. The research results demonstrate that this analytical framework can accurately extract disaster information, precisely identify critical time points in flood disasters, locate core affected areas, uncover primary regional issues, and further validate the sufficiency of response measures, therefore enhancing the efficiency in collecting disaster information and analytical capabilities. [Display omitted] •Flood disaster information extraction from text combining deep learning and regular expressions, achieving 83 % accuracy•Multidimensional flood analysis framework for comprehensive flood dynamics analysis•LDA-based semantic topic modeling for regional flood issues and needs identification•Correlation analysis of rescue requests and measures for sufficiency of response measures
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:0048-9697
1879-1026
1879-1026
DOI:10.1016/j.scitotenv.2024.174948