Next Article in Journal
Applications of Artificial Intelligence for Heat Stress Management in Ruminant Livestock
Previous Article in Journal
Research on On-Line Monitoring of Grinding Wheel Wear Based on Multi-Sensor Fusion
Previous Article in Special Issue
A Precise and Stable Space-Based Time System for Navigation in Smart Cities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multimodal Social Sensing for the Spatio-Temporal Evolution and Assessment of Nature Disasters

by
Chen Yu
1,2 and
Zhiguo Wang
1,*
1
Center for Public Security Technology, University of Electronic Science and Technology of China, Chengdu 610054, China
2
Institute of Public Security, Kashi Institute of Electronics and Information Industry, Kashi 844000, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(18), 5889; https://doi.org/10.3390/s24185889
Submission received: 30 July 2024 / Revised: 8 September 2024 / Accepted: 10 September 2024 / Published: 11 September 2024
(This article belongs to the Special Issue Advances in Wireless Sensor Networks for Smart City)

Abstract

:
Social sensing, using humans as sensors to collect disaster data, has emerged as a timely, cost-effective, and reliable data source. However, research has focused on the textual data. With advances in information technology, multimodal data such as images and videos are now shared on media platforms, aiding in-depth analysis of social sensing systems. This study proposed an analytical framework to extract disaster-related spatiotemporal information from multimodal social media data. Using a pre-trained multimodal neural network and a location entity recognition model, the framework integrates disaster semantics with spatiotemporal information, enhancing situational awareness. A case study of the April 2024 heavy rain event in Guangdong, China, using Weibo data, demonstrates that multimodal content correlates more strongly with rainfall patterns than textual data alone, offering a dynamic perception of disasters. These findings confirm the utility of multimodal social media data and offer a foundation for future research. The proposed framework offers valuable applications for emergency response, disaster relief, risk assessment, and witness discovery, and presents a viable approach for safety risk monitoring and early warning systems.

1. Introduction

Natural disasters are sudden, large-scale public emergencies that cause significant economic damage and human casualties annually, such as floods, earthquakes, and typhoons. In 2021, China experienced 42 heavy rainfall events, affecting 59.01 million people, causing 590 deaths and leading to 245.89 billion yuan in economic losses [1]. It is therefore crucial to quickly and accurately identify potential disaster risks, assess the extent of any damage, and make prompt and effective emergency decisions for minimizing losses and maintaining social stability [2,3]. The collection of disaster information serves as the foundation for the decision-making process [4,5]. Traditionally, disaster information is collected via physical sensors and on-site investigations, but these methods struggle with efficiency, especially during widespread disasters, causing delays in emergency response [6,7]. It is therefore essential that new information sources be developed to address these issues.
In recent years, social media has emerged as a key platform for disaster-related information sharing due to its real-time nature and broad reach [8,9,10,11,12]. Affected individuals express opinions, seek help, or organize mutual aid through these platforms [13,14,15]. Social sensors can build a low-cost, wide-coverage network for efficient information dissemination [16]. Studies on social media usage during disasters, such as the Northern Ireland floods, confirm the volume of mutual aid information shared [17]. The rich personal experiences expressed through social media provide a deeper understanding of disaster impacts, driving the development of disaster semantics and spatiotemporal analysis.
Disaster semantic mining extracts information from both textual and visual social media content using advanced algorithms [13,18,19,20,21]. While text data are highly dense and efficient for dissemination, visual data, through object detection, offers a more intuitive understanding of disaster scenes [22,23,24,25]. Multimodal data help capture the full scope of disaster impacts, making it valuable for emergency management. The number of disaster-related posts also serves as a useful indicator of public attitudes toward disaster events, with post fluctuations reflecting disaster phases and impacts [26,27,28]. However, geolocation data are often limited, presenting challenges for spatial analysis [29]. This highlights the importance of integrating multimodal data with spatiotemporal information in disaster research.
While social media data have advanced disaster informatics, there is room to improve data analysis efficiency, as most existing methods rely solely on textual data [30,31,32,33,34]. Spatiotemporal analysis of multimodal content is crucial for intelligent emergency response. This study introduces a framework that combines text and visual content to extract disaster categories and assess damage severity. The framework uses multimodal neural networks and incorporates location data through named entity recognition (NER), addressing geolocation deficiencies.
A case study on the April 2024 Guangzhou rainstorm demonstrates the framework’s effectiveness in analyzing Weibo data. Key contributions include:
  • A disaster analysis framework is proposed for the extraction of disaster-related spatiotemporal information from multimodal social media data. The framework was demonstrated to be an efficient method for analyzing Weibo data pertaining to the Guangdong rainstorm.
  • The LLM is used for geolocation pre-processing, whereby supplementary location information is obtained to refine the modelling of the spatiotemporal distribution of disaster-related posts.
  • Statistical analysis reveals a strong correlation between the number of disaster-related multimodal posts and precipitation levels. The extracted disaster semantics exhibit a distinctive event-related spatiotemporal distribution pattern. These findings demonstrate that social sensors are able to reflect the varying degrees of impact that disasters have on different regions.
The paper is structured as follows: Section 2 reviews related literature; Section 3 outlines the framework and materials; Section 4 presents statistical analysis results; Section 5 discusses contributions and limitations; and Section 6 concludes with the study’s findings.

2. Related Works

With the advancement of information technology and the rise of social media, empirical evidence shows that individuals use these platforms to share and exchange information during crises. Social sensing technology is becoming crucial for gauging humanitarian needs, understanding public sentiment, and monitoring disaster progression. Research in this area is categorized into three main areas: (1) temporal distribution of disaster posts [35], which analyzes the variation in posts and topics throughout disaster phases; (2) spatial distribution of disaster posts, examining the relationship between post locations and disaster impact; and (3) semantic analysis, assessing the public’s reflection on disasters through social media content. While social media timestamps are easily accessible, extracting spatial and semantic information remains challenging.

2.1. Spatial Analysis of Disaster Social Media

Geospatial information from social media is vital for spatial analysis and guiding emergency rescue operations [36]. Most studies treat location data as address entities identified via NER models, but only 1–7% of posts have cleaned geolocation tags [37,38,39]. Some studies employ machine learning and natural language processing (NLP) tools for location extraction from text, though they often struggle with unstructured data. Deep learning models show improved performance in mapping words to entities for address extraction [33,40,41]. For example, Yan et al. [29] developed a bidirectional encoder representation from transformers–bidirectional long short-term memory-conditional random fields (BERT-BiLSTM-CRF) model to classify address words. Despite the accuracy of deep learning models on established datasets, real-world social media data’s noise hampers their performance.
Large language models (LLMs) have exhibited considerable ability to generalize across a range of text-based tasks. Their capacity to draw upon a wealth of prior knowledge has also enabled them to effectively identify addresses. Hu et al. [42] used minimal labelled data and geographical knowledge in chat-generative pre-trained transformer (ChatGPT) to extract locations via dialogue. Ambiguity in geographic names complicates location extraction, often requiring additional data for confirmation.
Table 1 compares studies using social media data for spatiotemporal disaster analysis, highlighting the limitations of relying solely on text and check-in data. The unsupervised methods used for disaster category extraction are data-dependent and lack flexibility [26,27,28,43,44,45,46]. Furthermore, studies that employed solely check-in data failed to consider a considerable volume of social media content devoid of check-in tags [26,27,43,44,45,46]. It is crucial to note that the disaster categories extracted through unsupervised methods [26,27,43,44,45,46] are data-dependent and cannot be designed to extract the desired topics. We implemented improvements to address the underutilization of these data.

2.2. Semantic Analysis of Disaster Social Media

Social media data are often noisy, complicating the extraction of disaster-related semantics. Modern platforms allow multimodal data (text, image, video), expanding disaster-related insights. Related studies fall into text-based, image-based, and multimodal categories. Text-based studies often focus on public sentiment, using statistical, machine learning, and deep learning methods. Statistical methods focus on keywords [47]. For example, Qian et al. [48] studied the evolution of flood events by analyzing the frequency of occurrence of 17 disaster-related words, such as “warning”, “forecast”, “thunderstorm”, “downpour”, and “flood”. While statistical methods are relatively straightforward to implement, they do not fully leverage the information present in the text. In contrast, numerous studies use machine learning models to extract themes from text. Commonly used machine learning methods include supervised methods such as support vector machine, Naïve Bayes, random forest, logistic regression, and unsupervised methods such as latent Dirichlet allocation (LDA) and K-means [4,26,44,49]. Additionally, deep learning methods are used to encode text sequences, including long short-term memory (LSTM) and bidirectional encoder representations from transformers (BERT), which can incorporate subjectively defined categories for the extraction of specified information [50]. For example, Zhang et al. [51] classified disaster-related elements into distinct categories and employed the NER task to train the BERT model for the extraction of disaster information. Wu et al. [52] used a neural network [53] to investigate the spatial and temporal distribution of climate change-induced affective orientations in microblogging data. Their findings indicate that the majority of Chinese people have a positive attitude towards climate change.
Images provide valuable insights into disaster impact, with visual neural networks learning disaster-related features [22,54,55,56,57,58]. For example, Nia et al. [25] estimated financial damage by analyzing images of damaged buildings, while Li et al. [59] used class activation maps for damage localization. Despite the variability of image data, visual networks like the visual geometry group network (VGG) and residual network (ResNet) have been applied successfully to disaster imagery [60,61].
Combining text and image data in multimodal datasets offers more comprehensive insights than unimodal data alone. Multimodal datasets, though expensive, form the basis for advanced analysis. However, the limited size of these datasets presents challenges for model training. Multimodal models, such as those by Abavisani et al. [62] and Liang et al. [63], integrate features from both text and images and outperform unimodal approaches in tasks like humanitarian classification and damage estimation.
While multimodal data offer richer information, their integration with real-world events is still limited. We propose a new framework for disaster monitoring through social sensing, integrating multimodal content analysis with spatiotemporal analysis.

3. Materials and Methods

The workflow of this study is comprised of five steps, as illustrated in Figure 1. The initial step involved the collection of Weibo data based on pre-defined keywords, with the objective of establishing a comprehensive disaster social media database. Secondly, deep neural network models are trained for the extraction of disaster-related information and the recognition of location entities on the publicly available dataset. Thirdly, location information is extracted from Weibo data by LLM and NER models and subsequently encoded into spatial coordinates. Fourthly, the correlation between the number of Weibo posts and precipitation is quantified through quantitative analysis. Finally, the multi-level disaster semantics are extracted using the classification models, and the social perception of the disaster situation is investigated in conjunction with spatiotemporal information.

3.1. Study Area

Guangdong province represents the most significant economic centre in the southern region of China, encompassing an area of about 179,700 square kilometres and a population of approximately 127 million. Given its susceptibility to flooding and heavy precipitation, the social sensor in this region is particularly sensitive to rainfall patterns. It can thus reasonably be assumed that the correlation between the multimodal data and rainfall in this region will be statistically significant. In April 2024, sustained heavy rainfall caused extensive damage across numerous regions within Guangdong province. According to data from the Guangdong Meteorological Department, the April average precipitation exceeded the historical records. The heavy rainfall resulted in a series of secondary disasters, including severe urban flooding, river flooding, landslides, house collapses, and road disruptions. These events resulted in a considerable number of casualties and property losses. This particular precipitation event was selected for analysis in this research due to its widespread impact and prolonged impact.

3.2. Data Collection

Weibo is a Chinese social media platform with characteristics similar to those of Twitter. It allows users to upload, comment on, and retweet text, images, and video content. The platform has been the subject of numerous studies, which have identified it as a valuable resource. To investigate the impacts of this rainfall event, a corpus of Weibo posts was assembled for analysis. A Python-based web crawler was used to collect data associated with specific keywords via the advanced search functionality of Weibo. In particular, keywords related to rainfall and flooding were selected, including “rainstorm”, “heavy rain”, “heavy rainfall”, “ponding”, “waterlogging”, “inundation”, and “flood”. These keywords were then combined with the names of all prefecture-level cities in Guangdong in order to obtain a series of keyword combinations for retrieval. The data collection period was set from 00:00 on 16 April 2024 to 24:00 on 1 May 2024. The dataset comprised multiple tags and the contents of Weibo posts were downloaded, including post ID, user ID, user name, timestamp, and text content. In the case of posts containing visual content, the images were downloaded in their entirety. Ultimately, a total of 37,010 posts were collected for further processing.

3.3. Data Clean

The raw posts contain a considerable number of irregular terms and repetitive content, which impairs the efficiency of content analysis. To enhance the quality of the data, four steps were performed. The text content was initially normalised, whereby uniform resource locators (URLs), special characters, emoticons, and meaningless characters were removed. Secondly, posts with a text length of less than 5 words were considered incomplete and removed. Thirdly, in the case of posts containing duplicate content, one was randomly retained, while the other redundancies were deleted. Fourthly, accounts that solely reposted news and did not contribute original content were removed. Following the aforementioned data-cleaning steps, a total of 20,977 posts were retained providing the necessary materials for the subsequent spatial and temporal distribution study.

3.4. Extracting Location Information

The mapping of location information enables the integration of posts into a geospatial space, thereby establishing a foundation for spatiotemporal analyses. The Weibo platform employs two distinct methods for identifying user location: IP-based location tags and check-in tags. The IP location is limited to the provincial level, whereas the check-in tag can be used to identify the POI at the local level. However, it is noteworthy that Weibo posts containing check-in information are relatively rare, comprising approximately 3% of the total dataset. To obtain a greater number of spatial locations, numerous approaches [28,51,64] use NER models to extract location information directly from the text. The capability of NER models to extract location elements depends on the quantity and quality of the training data. Consequently, NER models are unable to make precise predictions when the textual content of Weibo does not align with the expression of the training data. To address the deficiency of location data, a three-stage methodology was used. Firstly, all potential locations within the text were extracted through the use of LLM. Subsequently, further parsing was performed using the NER model. Finally, map services were used to conduct further verification and geocoding of the locations. The LLM was trained on a considerably larger scale than the NER model, thereby enabling more effective noise reduction in the Weibo text. The processed data exhibit enhanced granularity and is better suited to leveraging the capabilities of the NER model.

3.4.1. LLM-Based Geolocation Pre-Processing

LLMs are a class of generative artificial intelligence models that are trained on vast quantities of data with the objective of developing the capacity to comprehend and generate textual content. The Llama 3 model, released by Facebook, is an advanced open-source model that has demonstrated excellent performance on a range of textual tasks. The Llama3 model, which was fine-tuned on Chinese corpora, was employed for the extracting of potential address information from the text. In particular, the model has 8.03 billion parameters with the pre-trained weights from the Ollama library (https://ollama.com/wangshenzhi/llama3-8b-chinese-chat-ollama-q4, accessed on 1 May 2024). The incorporation of extensive prior knowledge in the form of large-scale pre-training data enables the model to effectively mitigate the noise present in non-standard Weibo text.
Inference is performed by Llama 3 in the form of dialogue, necessitating both texts and prompts as input. The prompts are the text that directs the LLM in the generation of specific content and are directly related to the model output. In order to elicit the desired information, a number of prompts were used, including “This sentence contains location expressions, extract the relevant words” and “Return the locations in this text without annotation”. Figure 2a illustrates the visualization workflow of Llama 3. The absence of a check-in tag in each social media text is indicative of the need to combine these texts with prompts as model inputs for the extraction of location information. It would be optimal for the model to provide a response that includes all of the address elements presented in the input text. Subsequently, all responses were aggregated and de-duplicated. Furthermore, a random selection of Weibo texts from the non-check-in data was used to evaluate the performance of LLM in extracting relevant information. On the test set of 200 samples, the pre-trained LLM achieved a precision rate of 0.99, a recall rate of 0.95, and an F1-score of 0.97. However, the responses invariably included terms that extended beyond mere addresses, as a consequence of the inherent normality of the LLM outputs. Moreover, some datasets include multiple locations with varying geographic levels, which presents challenges in accurately locating them through geocoding services. Therefore, the NER model was used to further normalize these locations.

3.4.2. Location Entity Extraction

The BERT-BiLSTM-CRF model is a frequently used NER model in social media analysis, which is employed for the extraction of geographic entities [28,29]. The model consists of three base models: BERT, BiLSTM, and CRF. BERT is a sequence model based on a self-attention mechanism that encodes words into tokens. The initial parameters for BERT are derived from pre-trained weights that have been trained on Chinese corpora [65], thereby enhancing the efficiency of model training. BiLSTM [66] is a recurrent neural network that is frequently utilized for establishing the contextual relationships between tokens through forward and backward computations. CRF [67] is an undirected graphical model, which can achieve efficient label prediction based on the mutual constraints between tokens. The Chinese address element recognition dataset from the 2021 China Conference on knowledge graph and semantic computing (CCKS) NER challenge [68] was used for the model training. This dataset was divided into 21 geographic levels, such as ‘POI’, ‘City’, ‘District’, ‘Town’, and ‘Road’. The model was trained to extract address nouns by predicting the positions of these labels. The dataset contained 10,826 public address entries and was partitioned into a training set and a test set comprising 8854 and 1972 samples, respectively. The hyper-parameters of the model were determined through a 5-fold cross-validation procedure on the training set. The learning rate was set to 3 × 10−5 for BERT and 3 × 10−3 for the remaining components. The AdamW optimizer was employed to train the model for 10 epochs. The performance of the model was evaluated using three metrics on the test set: precision, recall, and F1-score, which yielded values of 0.93, 0.92, and 0.92, respectively. The results demonstrated that the model is capable of effectively extracting address information from high-quality data.
The model was employed for the examination of LLM outputs, whereby location information was extracted from the non-check-in data. Figure 2b illustrates the visualization workflow of the BERT-BiLSTM-CRF model. The social media texts were encoded in the form of character sequences by the three models, BERT, BiLSTM and CRF in a sequential manner, thus obtaining annotated sequences. The annotations describe the geographic level (City, Road, Other, and so on) of each character and its position in the word (B-Begin, I-Intermediate, E-End). The aforementioned annotations allow the aggregation of all address terms within the text. Table 2 shows the function of the LLM and NER models through the presentation of several examples. It is evident that the NER model is not an effective means of processing Weibo text. One common issue with the NRR model is the occurrence of inaccurate segmentation, whereby the city “Wuhan” is output alongside the adjective “sunny”. Furthermore, the model occasionally identifies abstract address-related terms as actual addresses, such as “city of laborers”, “six villages”, and “Guangdong road” as observed in these samples. While these phrases are indeed descriptions of addresses, the NER model lacks the capacity to discern the underlying semantics. Consequently, LLM was used to eliminate superfluous terms, thereby enhancing the performance of the NER model. Following the incorporation of LLM, the NER model demonstrated enhanced predictive accuracy.
Table 3 demonstrates the predictive accuracy of the NER model when applied to 200 manually labelled data. The optimal recall is only achieved when all place names are extracted without any irrelevant words. The criterion is more rigorous than that employed in the preceding LLM section, which prioritizes the validation and standardization of the location data. Following the combination of LLM, a notable enhancement in the performance of the NER model was observed, with an increase in recall of 34%. The results demonstrate that the noise in the texts is effectively removed following LLM processing, which facilitates the extraction of locations. In conclusion, the combined model achieves an F1-score of 0.9186, indicating that it has a robust capacity to extract geographical names and is well suited to the objectives of the study.

3.4.3. Geocoding

The locations require additional processing to be transformed into actual geographic locations. The application programming interface (API) of the Gaode map (https://lbs.amap.com/ (accessed on 30 June 2024)) was used to ascertain the latitude and longitude coordinates corresponding to all the place names. Addresses that were not located within Guangdong province were excluded from the subsequent analysis. Additionally, address elements with duplicate names were determined based on the information pertaining to the district, city, and province. Ultimately, a total of 5338 posts with location data were obtained, and the latitude and longitude of these addresses were mapped to QGIS [69] for spatial analysis.

3.5. Extracting Disaster Information

The semantics of disasters in social media data are highly heterogeneous, and the mixing of information on different topics reduces their application value [32]. For example, the objective of identifying victims is primarily focused on human-related data, while the objective of disaster reconstruction is centred on damage-related information [70,71]. For this reason, multiple classification tasks were established in a hierarchical structure to facilitate the efficient extraction of disaster-related information. Specifically, three subtasks from the open-source dataset CrisisMMD [72] were designed as strategies for the hierarchical extraction of disaster-related information. Firstly, the collected posts were classified into two categories: “Informative” and “Not informative”. The classification was based on the presence or absence of disaster-related information. Secondly, the collected posts were further classified into a number of fine-grained humanitarian categories based on semantic features. Thirdly, the damage severities were evaluated and classified. The detailed definitions of these categories are in accordance with those set forth in the CrisisMMD dataset. Some categories with approximate meanings have been merged, as shown in Table 4.

3.5.1. Model Training

A two-stream multimodal framework was used for the purpose of conducting a multimodal analysis. In this framework, BERT [65] and DenseNet [73] were used as the text and image encoder, respectively, with the objective of extracting textual and visual features. Figure 2c illustrates the visualization workflow of the multimodal model. The image and text in a social media image–text pair are encoded separately by the encoder of the corresponding modality. The two obtained embeddings are then concatenated to form an overall embedding of the image–text pair. Finally, a linear classifier is employed to make a prediction regarding the disaster category, based on the overall information aforementioned. Since CrisisMMD is in English, the text was translated into Chinese using the Baidu translation API (https://fanyi-api.baidu.com/ (accessed on 30 June 2024)). Text augmentation was achieved through the implementation of the easy data augmentation [74] technique. In order to augment the dataset with regard to the image modality, a series of data augmentation techniques were employed, including random scaling, random cropping, and random flipping. The learning rate was set to 2 × 10−5, the optimizer was AdamW, and the number of epochs was 10. To address the issue of class imbalance in the training data and to optimize network weights, focal loss [75] was employed. To evaluate the performance of the model, a series of performance metrics were employed, including precision, recall, F1-score, receiver operating characteristic curve (ROC), and area under the ROC curve (AUC). The ROC curve is a visual representation of the classification ability of a model, demonstrating the relationship between the true positive rate (TPR) and the false positive rate (FPR) at different thresholds.
TPR = TP / TP + FN FPR = 1 TN / FP + TN
where TP, FP, TN and FN represent the numbers of true-positive, false-positive, true-negative, and false-negative samples, respectively. An ideal classifier should be situated as closely as possible to the upper left corner of the ROC curve, with a false positive rate of 0 and a true positive rate of 1. AUC represents the area under the ROC curve enclosed with the horizontal axes. A value close to 1 indicates the classifier exhibits excellent performance. Furthermore, a unimodal BERT model was trained on the text modality for comparative purposes. A summary of all model and train dataset information is presented in Table 5.
As illustrated in Table 6, the multimodal model exhibits superior performance compared to the unimodal model, achieving F1-scores of 0.9013, 0.8419, and 0.7140 on the three tasks, respectively.
These results demonstrate that the multimodal approach leads to more accurate and comprehensive results. The ROC curves and AUC values on the test set are shown in Figure 3. In particular, the multimodal model achieves macro-average AUCs of 0.95 (Figure 3b), 0.97 (Figure 3d), and 0.76 (Figure 3f) on the three tasks, which are superior to the values of 0.89 (Figure 3a), 0.95 (Figure 3c), and 0.63 (Figure 3e) achieved by the unimodal model, respectively. In summary, the multimodal model demonstrates sufficient accuracy in disaster social media analysis. The superiority of the multimodal model across all metrics demonstrates that the incorporation of the image modality offers valuable supplementary information, thus corroborating the hypothesis that multimodal approaches are more effective in this context. The extracted disaster semantics serve as the foundation for spatiotemporal semantic analyses, which are conducted in conjunction with time and location data.

3.5.2. Prediction

The trained multimodal model is used to predict disaster categories from Weibo posts. Prior to this, it is necessary to undertake preprocessing in order to address the discrepancies between the Weibo data and the training dataset. As illustrated in Figure 4a, the length of Weibo texts is considerably longer than that of the training dataset. The maximum length of the training dataset is approximately 60, yet there are numerous Weibo texts that exceed this length. To mitigate this discrepancy, the Weibo texts were segmented into shorter sentences with punctuation marks with a maximum length of 60. In addition, some Weibo posts contain multiple images, while the multimodal model is unable to process. The distribution of the number of images in Weibo posts is shown in Figure 4b, which depicts 7303 (7303 = 3638 + 958 + 676 + 454 + 98 + 442 + 48 + 51 + 938) multimodal posts and 13,674 unimodal posts. The number of Weibo posts with multiple images (3665 = 7303 − 3638) is approximately equivalent to that of posts with one image (3638), and thus cannot be disregarded. In order to utilize all images, each image in the sample was paired with each text segmentation. Subsequently, a max pooling operation was employed to fuse the aforementioned paired features for the purpose of prediction. Similarly, 200 multimodal Weibo posts were randomly selected and manually classified to serve as the test set. The results for precision, recall, F1-score, and AUC are listed in Table 7. The F1 scores for the three tasks are 0.8713, 0.8796, and 0.7808, with AUC values of 0.9339, 0.9219 and 0.7102, respectively. These results suggest that the classification outcomes accurately represent the multimodal content.
Table 8 presents the results of the predictive analysis on a number of samples, with the objective of illustrating the various disaster categories. The first sample mentions the phrase “heavy rain”, but its content is not related to the specified rainfall event. Consequently, it is classified as “Not informative” in the informativeness task. Conversely, the second sample mentions that the rainfall occurs on the highway, therefore it is designated as “Informative”. The image in the third sample depicts a flooded street, thus the predicted label for this sample in the humanitarian categorization task is “Infrastructure, utility, or vehicle damage”, while the label in the damage assessment task is “Severe damage”. The image and accompanying text in the fourth sample pertain to the act of donating and thus are classified as “Rescue, volunteering, or donation effort”. The image in the fifth sample lacks pertinent content; however, the accompanying text describes the casualty status, thus the label in the second task is “Affected individuals”. The sixth sample is related to weather forecasting, and thus its information type is designated as “Other relevant information”. In general, the multimodal models trained with different tasks can effectively extract disaster-related information from social media posts.

3.6. Weibo Activity

The demographic characteristics of Guangdong province, including population size and age distribution contribute to the inconsistency in the number of Weibo users across the region. This spatial discrepancy leads to a greater number of Weibo posts in areas with a higher concentration of Weibo users, which in turn affects the effectiveness of Weibo as a disaster indicator. Therefore, the relative number of Weibo (RWeibo) is employed as a metric for gauging social media activity, which is defined as follows:
R W e i b o = N W e i b o / P o p u l a t i o n i , j × W e i g h t i , j
where NWeibo is the total number of Weibo posts in a prefecture-level city, i belongs to the age group {0–10, 10–20, 20–30, 30–40, 40–50, 50–60, 60–More}, j belongs to the gender group {male, female}, Populationi,j is the population of a specific age i and gender j [76], and Weighti,j is the ratio of users corresponding to age i with gender j [77]. The relative number of Weibo users is stratified by age and gender across a range of geographic regions, allowing for the suppression of spurious spatial characteristics that are dominated by the majority population.

4. Results

4.1. Spatiotemporal Characteristics of Disaster-Related Weibo

4.1.1. Temporal Distribution of Weibo

To examine the perception of heavy rainfall by social sensors, the number of daily Weibo posts from 16 April to 1 May was quantified. To quantify the severity of the disaster, daily rainfall data were collected and are presented in Figure 5a. From 16 April to 19th, the study area experienced minimal rainfall, resulting in a relatively low number of Weibo posts (Figure 5b). The precipitation on 20 April was twice as high as that of the previous day, and the number of Weibo posts also doubled, indicating a notable public response to the rainfall. From 20 April to 28 April, continuous rainfall caused severe damage, resulting in a sustained period of high Weibo activity with daily posts consistently exceeding 1000. Then, a further round of precipitation from 21 April to 1 May was accompanied by a rise in the number of Weibo posts. The peak in the number of Weibo posts occurs earlier than the peak in precipitation, which may indicate a waning public interest in the ongoing rainfall topic. The fluctuations in the number of Weibo posts largely reflected the patterns of precipitation, with the exception of 19 April and 25 April. To quantify the consistency of this relationship, a correlation test was performed using the Pearson correlation coefficient (PCC). As evidenced in Table 9, the correlation between the number of Weibo and precipitation is 0.6367, with a p-value less than 0.01, indicating a strong correlation. This result demonstrates the reliability of Weibo data as a research object.
The daily number of unimodal (Figure 5c) and multimodal posts were further counted (Figure 5d) and the correlations with precipitation were presented in Table 9. A correlation coefficient of 0.6125 was observed between the number of unimodal Weibo posts and precipitation, with a p-value less than 0.05. In contrast, the result for the multimodal data is 0.6649 with a p-value less than 0.01. These findings suggest that the multimodal data exhibit a stronger correlation with precipitation and is more effective in reflecting rainfall patterns compared to unimodal data.

4.1.2. Spatial Distribution of Weibo

A total of 1482 addresses within Guangdong province were obtained from the 1797 Weibo posts containing check-in information. Of these, 519 were identified as non-duplicates. A total of 19,180 non-check-in Weibo posts were examined, resulting in the extraction of 3856 addresses, 294 of which were unique. The non-check-in data provide a substantial number of locations, rendering it a valuable data source for spatial analysis. In the dataset comprising check-in posts within Guangdong province, the number of multimodal posts (932) exceeds the number of unimodal posts (550). This implies that users frequently capture images when checking in, which enables the observation of the disaster scene. In the non-check-in dataset, the number of unimodal posts (2448) exceeds the number of multimodal posts (1408) due to the higher quantity of unimodal data. In total, there are 2340 multimodal locations and 2998 unimodal locations, which demonstrates that multimodal data are also a significant subject for spatial analysis.
The effects of heavy rainfall vary across different regions, leading to disparate responses from social media users. Consequently, data regarding the precipitation and the number of Weibo posts at the prefecture level were gathered. Figure 6a illustrates the total precipitation across Guangdong province from 16 April to 1st May, demonstrating that the majority of rainfall occurred in the central region, with significantly less in the eastern and western areas. Figure 6b presents a count of the number of Weibo posts in each region. It is notable that the distribution of Weibo activity exhibits a pattern that is similar to that observed in the precipitation distribution. Additionally, the distributions of check-in and non-check-in data for both unimodal and multimodal sources are further analyzed. Figure 6c,e present that the check-in data are concentrated in the central and southern regions. The multimodal data provide additional insight into the Meizhou and Chaozhou areas, which is less evident in the unimodal data. The non-check-in data presented in Figure 6d,f provide supplementary information in Jiangmen, Heyuan, and Shanwei. The spatial distribution of rainfall-related Weibo posts exhibits a clustering pattern that is related to precipitation, rather than a random distribution that lacks meaningful coherence.
The region with the highest number of Weibo posts is Shenzhen, yet it is not the region with the highest precipitation. This can be attributed to the number of users on the platform. The age distribution of Weibo users in 2020 indicates that the majority of users are individuals under the age of 30, representing nearly 80% of the total users. Shenzhen has the largest population of young people in Guangdong province, which results in a higher number of Weibo users compared to other regions. To address the issue of concentrated disaster information in densely populated areas, the relative number of Weibo (Equation (2)) was used as an indicator of Weibo activity. The global bivariate Moran’s I can be used to quantify the spatial dependence between two variables. In this study, it is used to measure the correlation between disaster-related Weibo activities and rainfall in Guangdong Province. A Moran value exceeding zero indicates a positive spatial correlation. Conversely, a negative spatial correlation is indicated by a Moran value less than zero. The bivariate Moran’s I in GeoDa [78] was used to quantify the spatial distribution correlation, with the results shown in Table 10. The Moran’s I for the quantity of Weibo data and precipitation is 0.365, with a high level of significance (p < 0.01, Z = 3.1154). The results indicate a significant positive spatial correlation between the quantity of Weibo data and precipitation. Furthermore, the non-check-in data exhibit a correlation of 0.336 with precipitation, which is identical to that observed in the check-in data. This suggests that the proposed location extraction strategy yields accurate location data and can provide spatial analyses comparable to those derived from check-in locations. Furthermore, the total, non-check-in, and check-in coefficients of multimodal data are 0.397, 0.340, and 0.346, respectively, which are higher than those of unimodal data (0.339, 0.327, and 0.310). This suggests that multimodal data correlate more strongly with precipitation than unimodal data and are more effective at detecting disasters in space.

4.2. Spatiotemporal Characteristics of Disaster Categories

4.2.1. Proportion of Different Disaster Categories

The proportions of disaster-related categories were obtained through the application of the classification models to the dataset. With regard to Task 1 (Figure 7a), the largest proportion is “Not informative”, with informative samples accounting for less than 30%. Figure 7b illustrates the proportion of informative posts within the humanitarian categories. A total of 70% or more of the samples contain humanitarian content, with the largest number of samples falling within the “Other relevant information” category and the smallest number of samples falling within the “Affected individuals” category. This proportion is associated with the characteristics of heavy rainfall. The impacts of persistent precipitation accrue gradually, resulting in the continued accumulation of weather-related data within the “Other relevant information” category. Furthermore, individuals receive warning notifications and proactively avoid flood-prone areas, resulting in a reduction in observations of affected individuals. Figure 7c depicts the damage estimation for informative posts, with approximately half of them classified as severe damage. This suggests that there is a public awareness of the potential for disasters and that social sensors can be used as a means of monitoring the emergence of new risks.

4.2.2. Temporal Statistical Analysis of Disaster Categories

The daily number of Weibo posts pertaining to various humanitarian and damage categories is presented in Figure 8a,b. For the purpose of qualitative analysis, only those categories deemed relevant are retained. It is evident that the activity of each category is closely correlated with the progression of heavy rainfall. In the period preceding the disaster, up until 19 April, the predominant categories were “Other relevant information” and “Mild, little, or no damage”. As the precipitation levels increased and persisted from 20 April to 23 April, the number of “Infrastructure, utility, or vehicle damage”, “Rescue, volunteering, or donation effort”, “Affected individuals” in Figure 8a and “Severe damage” in Figure 8b exhibited a notable increase. This illustrates that the public is attempting to disseminate disaster-related information and seek assistance through social media platforms. The proportion of the “Affected individuals” category reached its peak on 22 April, coinciding with the first official report of fatalities [79]. Concurrently, the “Rescue, volunteering, or donation effort” category exhibited an increase from 22 April to 23 April, reflecting the implementation of timely emergency relief actions in response to the rainfall that occurred on 21 April. The precipitation ceased for a brief period on 24 April, resulting in a notable reduction in the number of Weibo posts. Subsequently, another period of sustained precipitation commenced on 25 April. The highest precipitation levels were recorded on 26 April, yet the number of damaged posts remained relatively low in comparison to previous observations. This can be attributed to the fact that the public took the initiative to implement preventive measures based on their experience of a previous rainfall event, thereby, reducing their sensitivity to disaster-related damage. The results demonstrate a strong correlation between the temporal shifts in disaster-related categories and the progression of the disaster. In the aftermath of the disaster, social observations regarding damage and injuries began to emerge.
The intraday distribution characteristics of public perceptions are further examined in Figure 9a,b. To gain insight into the distribution of public opinion throughout the day, all data collected during the study period were counted on an hourly basis. Overall, public activity exhibits a discernible temporal pattern, with activity intensity declining from the morning (6:00–12:00), afternoon (12:00–18:00), to the evening (18:00–24:00), and reaching minimal levels during the early morning hours (0:00–6:00). With regard to the humanitarian categories (Figure 9a), the “Rescue, volunteering, or donation effort” category is concentrated during the daytime, while the “Affected individuals” category is mainly distributed in the evening. This uneven distribution may be attributed to the data source. The “Rescue, volunteering, or donation effort” category posts are from relief organizations, which typically operate during daylight hours and disseminate accordingly. The “Affected individuals” category posts are predominantly sourced from authoritative media, which gather data during the daytime and disseminate the findings in the evening. Regarding the distribution of the “damage” category (Figure 9b), the peak occurs in the morning due to the fact that reports from the previous night are often included in morning news items. Consequently, the number of posts during the morning is approximately twice that of other times of the day. In conclusion, the distribution of daily and hourly messages can assist crisis managers in enhancing the efficacy of their preparation and response strategies. Moreover, the proposed method uses individual image–text pairs as inputs to extract disaster-related information, eliminating the need for a specified number of sample sets. Therefore, the reduced Weibo activity during the early morning hours does not affect the performance of the method.

4.2.3. Spatial Statistical Analysis of Disaster Categories

Figure 10 illustrates the proportion of various categories across different regions. The different colors within the pie chart represent distinct categories, and the size of the circle signifies the quantity of Weibo posts. As illustrated in Figure 10a, the “Other relevant information” category is the most discussed topic across all regions with the exception of two cities. Some cities have posts of the “Infrastructure, utility, or vehicle damage” and “Rescue, volunteering, or donation effort” categories, which are primarily concentrated in the central region. The “Affected individuals” category is more concentrated in the central region, particularly in Qingyuan in the central-northern region. As illustrated in Figure 10b, the majority of cities are predominantly represented by the “Mild, little, or no damage” category, while the “Severe damage” category is concentrated in the central region. Notably, three adjacent cities in the central-northern area, namely Qingyuan, Shaoguan, and Zhaoqing, collectively account for nearly 50% of the “Severe damage” category. These findings support the assertion that central-northern cities are the most severely impacted, which is consistent with the official reports.

4.2.4. Transition of Disaster Categories

The study also examines the multiple semantic meanings of samples in order to illustrate the ways in which information varies among the three types of categories. A Sankey diagram is presented in Figure 11 for the visualization of the paths of information flow. The left and right endpoints of each strip indicate the multiple semantics of the corresponding sample subset, and the strip width represents the number of samples in the subset [45]. More than half of the samples in the “Informative” category are transitioned into four specific humanitarian categories, while a relatively small proportion are transitioned into the “Not informative” category. Less than half of “Other relevant information” posts are transitioned to the damage category. In contrast, more than half of the “Infrastructure, utility, or vehicle damage”, “Rescue, volunteering, or donation effort”, and “Affected individuals” posts are transitioned to that category. This indicates that the impacts and damages resulting from rainfall can be described using predefined damage categories. The examination of semantic transitions pertaining to disaster-related topics enables decision-makers to respond effectively to a range of potential hazards.

5. Discussion

The examination of actual occurrences within cyberspace can provide invaluable insights for the advancement of disaster research. This study employs a quantitative approach to examine the temporal and spatial correlations between rainfall patterns and Weibo activity in Guangdong province. The findings support the effectiveness of social media can serve as an effective social sensor during this extensive precipitation event. Furthermore, the number of multimodal posts exhibited a slightly higher temporal and spatial correlation with rainfall than the unimodal data, indicating that multimodal data are a valuable research object in disaster informatics. Table 1 presents a comparative analysis of the performance of relevant studies, employing the mean F1-score on the three topic extraction tasks as the evaluation metric. It should be noted that unsupervised models require subjective understanding to specify topic categories and therefore cannot be used to make objective comparisons. It can be seen that the multimodal models outperformed the unimodal model [28,43], which suggests that multimodal applications may offer a promising avenue for future research. The proposed model employed a more advanced visual backbone, which contributed to its superior performance compared to the existing literature [29]. This result highlights the importance of developing advanced models to enhance the efficiency of information extraction.
The analyses of public reactions indicate that in the early phase of the event, public attention was aligned with the onset of the heavy rainfall. The peak in Weibo activity coincides with a peak in precipitation on 21 April, exhibiting minimal temporal lag. However, as the event continues, there is a discernible shift in public attention, which is reflected in a decline in the number of Weibo posts during the subsequent period of rainfall. Furthermore, the shift in various categories demonstrates that messages about rescue efforts and casualties are delayed, whereas posts concerning facility damages and other matters are concurrent with the onset of rainfall. This is due to the fact that the preparation of rescue operations and the gathering of casualty data necessitate a greater investment of time, while other messages are more straightforward to confirm and post. Furthermore, the discrepancy illustrates the real-time character of social media in disseminating disaster-related information. The spatial statistical analysis indicates a notable level of Weibo activity in the affected areas. This suggests that social media can be used to estimate the impact range of disaster events.
Many studies have demonstrated the capability of social sensors in extracting topics, identifying affected areas, and tracking the spatiotemporal evolution of disasters. However, most of these studies [27,28,43,44] focus on mining content from social media to explore the potential of social sensors, with few utilizing statistical methods to characterize the spatiotemporal correlation between rainfall and social media posts. For example, Wang et al. [26] collected rainfall data and Weibo posts during the Zhengzhou rainstorm, analyzing the spatiotemporal evolution of related topics in disaster scenarios. They found that help-seeking messages spiked and then gradually decreased after the disaster, consistent with our findings. However, the lack of quantitative analysis reduces the reliability of using precipitation as a basis for dividing event stages. Wu et al. [45] examined topic transition patterns on social media during heavy rainfall in Hefei. We further characterized the transitions between different thematic categories. Although they presented spatiotemporal distributions for each topic, the absence of statistical indicators makes it difficult to uncover underlying patterns. Yan et al. [29] demonstrated the spatiotemporal distribution of Weibo flood points and rainfall during the Anhui Province rainstorm, confirming the effectiveness of social sensors in revealing affected locations. However, they only analyzed data from two days, limiting insights into the entire event. In contrast, this study quantifies the spatial and temporal correlations between precipitation and social media activity using statistical methods, yielding more compelling results. Although this study makes a notable contribution to the field, it is essential to recognize the limitations of the research. Firstly, the quality of the training data limits the performance of multimodal models. The lack of Chinese multimodal datasets is a significant impediment to the analysis of Chinese social media. Additionally, the majority of commonly used datasets are single-label, yet real-world posts may encompass multiple pieces of information, thereby necessitating the availability of multi-labelled datasets. It is therefore imperative to develop high-quality datasets. Secondly, the current feature fusion methods are not sufficiently interpretable with regard to disaster elements in text and images. This results in the inability to develop targeted emergency response strategies for disaster entities. Accordingly, subsequent research will concentrate on enhancing the interpretability of multimodal models.

6. Conclusions

This study examines the effectiveness of multimodal social sensors in enhancing situation awareness in disaster scenarios, with a particular focus on a widespread and sustained rainfall event. A novel framework is proposed for the extraction of multimodal disaster semantics from the Chinese Weibo platform for the purpose of tracking social reflections. Firstly, this study demonstrates that the combination of location extraction methods with large language models can effectively mine address information from text, thereby providing a feasible path for obtaining the spatial distribution from non-check-in data. Secondly, the classification outcomes provide evidence that multimodal models are effective in discerning disaster-related data within authentic social media data, thereby enhancing disaster awareness. Thirdly, the results of the quantitative analysis demonstrate a significant correlation between the quantity of multimodal data and the precipitation level. Furthermore, the correlation between multimodal data and precipitation is stronger than that of unimodal data, indicating that multimodal data are a more reliable source of information regarding disasters. The disaster information extracted by the proposed method from social media platforms can provide support for a number of important emergency response activities, such as emergency rescue, victim discovery, public opinion analysis, post-disaster reconstruction, and other projects during emergencies. In conclusion, these findings substantiate the utility of multimodal social media data in disaster informatics and offer a foundation for future research on multimodal data.
From a practical perspective, future research will focus more on two key areas. Firstly, the development of more high-quality datasets, especially those that are multi-labeled, multi-scenario, and contain more modalities, will enhance the efficiency of models operating with disparate event types. Secondly, an investigation into the interpretability of disaster-related content on social media could provide more direct evidence for the enhancement of emergency response operations.

Author Contributions

Conceptualization, C.Y. and Z.W.; Formal analysis, C.Y.; Investigation, C.Y.; Writing—original draft, C.Y.; Writing—review & editing, Z.W.; Supervision, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Dataset available on request from the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. The Basic Situations of Natural Disasters of China in 2021. Available online: http://www.mem.gov.cn/xw/yjglbgzdt/202201/t20220123_407204.shtml (accessed on 1 May 2024).
  2. Yuan, F.; Liu, R.; Mao, L.; Li, M. Internet of people enabled framework for evaluating performance loss and resilience of urban critical infrastructures. Safety Sci. 2021, 134, 105079. [Google Scholar] [CrossRef]
  3. Dou, M.; Wang, Y.; Gu, Y.; Dong, S.; Qiao, M.; Deng, Y. Disaster damage assessment based on fine-grained topics in social media. Comput. Geosci. 2021, 156, 104893. [Google Scholar] [CrossRef]
  4. Li, L.; Bensi, M.; Cui, Q.; Baecher, G.B.; Huang, Y. Social media crowdsourcing for rapid damage assessment following a sudden-onset natural hazard event. Int. J. Inf. Manag. 2021, 60, 102378. [Google Scholar] [CrossRef]
  5. Poblete, B.; Guzmán, J.; Maldonado, J.; Tobar, F. Robust detection of extreme events using twitter: Worldwide earthquake monitoring. IEEE Trans. Multimed. 2018, 20, 2551–2561. [Google Scholar] [CrossRef]
  6. Imran, M.; Castillo, C.; Diaz, F.; Vieweg, S. Processing social media messages in mass emergency: A survey. ACM Comput. Surv. 2015, 47, 67. [Google Scholar] [CrossRef]
  7. Liu, S.B.; Palen, L.; Sutton, J.; Hughes, A.L.; Vieweg, S. In search of the bigger picture: The emergent role of on-line photo sharing in times of disaster. In Proceedings of the 5th International Conference on Information Systems for Crisis Response and Management, Washington, DC, USA, 4–7 May 2008; pp. 4–7. [Google Scholar]
  8. Alam, F.; Qazi, U.; Imran, M.; Ofli, F. Humaid: Human-annotated disaster incidents data from twitter with deep learning benchmarks. In Proceedings of the 15 th International AAAI Conference on Web and Social Media, Virtual, 7–10 June 2021; pp. 933–942. [Google Scholar]
  9. Pohl, D.; Bouchachia, A.; Hellwagner, H. Online indexing and clustering of social media data for emergency management. Neurocomputing 2016, 172, 168–179. [Google Scholar] [CrossRef]
  10. Yao, K.; Yang, S.; Tang, J. Rapid assessment of seismic intensity based on Sina Weibo—A case study of the changning earthquake in Sichuan Province, China. Int. J. Disaster Risk Reduct. 2021, 58, 102217. [Google Scholar] [CrossRef]
  11. Jain, P.; Ross, R.; Schoen Phelan, B. Estimating distributed representation performance in disaster-related social media classification. In Proceedings of the 11th IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Vancouver, BC, Canada, 27–30 August 2019; pp. 723–727. [Google Scholar]
  12. Mouzannar, H.; Rizk, Y.; Awad, M. Damage identification in social media posts using multimodal deep learning. In Proceedings of the 15th International Conference on Information Systems for Crisis Response and Management, Rochester, NY, USA, 20–23 May 2018; pp. 529–543. [Google Scholar]
  13. Costa, T.A.G.; Meneguette, R.I.; Ueyama, J. Providing a greater precision of situational awareness of urban floods through multimodal fusion. Expert Syst. Appl. 2022, 188, 115923. [Google Scholar] [CrossRef]
  14. Pandey, R.; Bannan, B.; Purohit, H. CitizenHelper-training: AI-infused System for Multimodal Analytics to assist Training Exercise Debriefs at Emergency Services. In Proceedings of the 17th International Conference on Information Systems for Crisis Response and Management, Blacksburg, VA, USA, 23 May 2020; pp. 42–53. [Google Scholar]
  15. Zhang, C.; Fan, C.; Yao, W.; Hu, X.; Mostafavi, A. Social media for intelligent public information and warning in disasters: An interdisciplinary review. Int. J. Inf. Manag. 2019, 49, 190–207. [Google Scholar] [CrossRef]
  16. Chatfield, A.T.; Brajawidagda, U. Crowdsourcing Hazardous Weather Reports from Citizens via Twittersphere under the Short Warning Lead Times of EF5 Intensity Tornado Conditions. In Proceedings of the 2014 47th Hawaii International Conference on System Sciences, Waikoloa, HI, USA, 6–9 January 2014; pp. 2231–2241. [Google Scholar]
  17. Stephenson, J.; Vaganay, M.; Coon, D.; Cameron, R.; Hewitt, N. The role of Facebook and Twitter as organisational communication platforms in relation to flood events in Northern Ireland. J. Flood Risk Manag. 2018, 11, 339–350. [Google Scholar] [CrossRef]
  18. Cardoso, A.B.; Martins, B.; Estima, J. A Novel Deep Learning Approach Using Contextual Embeddings for Toponym Resolution. ISPRS Int. J. Geo-Inf. 2022, 11, 28. [Google Scholar] [CrossRef]
  19. Karimzadeh, M.; Pezanowski, S.; MacEachren, A.M.; Wallgrün, J.O. GeoTxt: A scalable geoparsing system for unstructured text geolocation. Trans. GIS 2019, 23, 118–136. [Google Scholar] [CrossRef]
  20. Madichetty, S.; Sridevi, M. A novel method for identifying the damage assessment tweets during disaster. Future Gener. Comput. Syst. 2021, 116, 440–454. [Google Scholar] [CrossRef]
  21. Zhang, D.Y.; Huang, Y.; Zhang, Y.; Wang, D. Crowd-Assisted Disaster Scene Assessment with Human-AI Interactive Attention. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 2717–2724. [Google Scholar]
  22. Dinani, S.T.; Caragea, D. Disaster image classification using capsule networks. In Proceedings of the 2021 International Joint Conference on Neural Networks, Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar]
  23. Weber, E.; Papadopoulos, D.P.; Lapedriza, A.; Ofli, F.; Imran, M.; Torralba, A. Incidents1M: A Large-Scale Dataset of Images with Natural Disasters, Damage, and Incidents. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 4768–4781. [Google Scholar] [CrossRef]
  24. Pally, R.J.; Samadi, S. Application of image processing and convolutional neural networks for flood image classification and semantic segmentation. Environ. Modell. Softw. 2022, 148, 105285. [Google Scholar] [CrossRef]
  25. Nia, K.R.; Mori, G. Building damage assessment using deep learning and ground-level image data. In Proceedings of the 14th Conference on Computer and Robot Vision, Edmonton, AB, Canada, 16–19 May 2017; pp. 95–102. [Google Scholar]
  26. Wang, W.; Zhu, X.; Lu, P.; Zhao, Y.; Chen, Y.; Zhang, S. Spatio-temporal evolution of public opinion on urban flooding: Case study of the 7.20 Henan extreme flood event. Int. J. Disaster Risk Reduct. 2024, 100, 104175. [Google Scholar] [CrossRef]
  27. Li, H.; Han, Y.; Wang, X.; Li, Z. Risk perception and resilience assessment of flood disasters based on social media big data. Int. J. Disaster Risk Reduct. 2024, 101, 104249. [Google Scholar] [CrossRef]
  28. Peng, J.; Zhang, J. Spatiotemporal assessment of urban flooding hazard using social media: A case study of Zhengzhou ‘7·20’. Environ. Modell. Softw. 2024, 176, 106021. [Google Scholar] [CrossRef]
  29. Yan, Z.; Guo, X.; Zhao, Z.; Tang, L. Achieving fine-grained urban flood perception and spatio-temporal evolution analysis based on social media. Sust. Cities Soc. 2024, 101, 105077. [Google Scholar] [CrossRef]
  30. Alam, F.; Imran, M.; Ofli, F. Image4Act: Online social media image processing for disaster response. In Proceedings of the 9th IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Sydney, NSW, Australia, 31 July–3 August 2017; pp. 601–604. [Google Scholar]
  31. Arachie, C.; Gaur, M.; Anzaroot, S.; Groves, W.; Zhang, K.; Jaimes, A. Unsupervised detection of sub-events in large scale disasters. Proc. AAAI Conf. Artif. Intell. 2020, 34, 354–361. [Google Scholar] [CrossRef]
  32. Alam, F.; Ofli, F.; Sajjad, H.; Imran, M. CrisisBench: Benchmarking crisis-related social media datasets for humanitarian information processing. In Proceedings of the 15th International Conference on Web and Social Media, Virtual, 7–10 June 2021. [Google Scholar]
  33. Li, S.; Lin, Y.; Huang, H. Relief Supply-Demand Estimation Based on Social Media in Typhoon Disasters Using Deep Learning and a Spatial Information Diffusion Model. ISPRS Int. J. Geo-Inf. 2024, 13, 29. [Google Scholar] [CrossRef]
  34. Zou, L.; Liao, D.; Lam, N.S.N.; Meyer, M.A.; Gharaibeh, N.G.; Cai, H.; Zhou, B.; Li, D. Social media for emergency rescue: An analysis of rescue requests on Twitter during Hurricane Harvey. Int. J. Disaster Risk Reduct. 2023, 85, 103513. [Google Scholar] [CrossRef]
  35. Eyre, R.; De Luca, F.; Simini, F. Social media usage reveals recovery of small businesses after natural hazard events. Nat. Commun. 2020, 11, 1629. [Google Scholar] [CrossRef] [PubMed]
  36. Xiang, Z.; Qin, H.; He, B.; Han, G.; Chen, M. Heat vulnerability caused by physical and social conditions in a mountainous megacity of Chongqing, China. Sust. Cities Soc. 2022, 80, 103792. [Google Scholar] [CrossRef]
  37. Malik, M.; Lamba, H.; Nakos, C.; Pfeffer, J. Population bias in geotagged tweets. In Proceedings of the 9th International Conference on Web and Social Media, Oxford, UK, 26–29 May 2015; pp. 18–27. [Google Scholar]
  38. Bryan-Smith, L.; Godsall, J.; George, F.; Egode, K.; Dethlefs, N.; Parsons, D. Real-time social media sentiment analysis for rapid impact assessment of floods. Comput. Geosci. 2023, 178, 105405. [Google Scholar] [CrossRef]
  39. Huang, B.; Carley, K.M. A large-scale empirical study of geotagging behavior on twitter. In Proceedings of the 11th IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Vancouver, BC, Canada, 27–30 August 2019; pp. 365–373. [Google Scholar]
  40. Wang, J.; Hu, Y.; Joseph, K. NeuroTPR: A neuro-net toponym recognition model for extracting locations from social media messages. Trans. GIS 2020, 24, 719–735. [Google Scholar] [CrossRef]
  41. Suwaileh, R.; Elsayed, T.; Imran, M.; Sajjad, H. When a disaster happens, we are ready: Location mention recognition from crisis tweets. Int. J. Disaster Risk Reduct. 2022, 78, 103107. [Google Scholar] [CrossRef]
  42. Hu, Y.; Mai, G.; Cundy, C.; Choi, K.; Lao, N.; Liu, W.; Lakhanpal, G.; Zhou, R.Z.; Joseph, K. Geo-knowledge-guided GPT models improve the extraction of location descriptions from disaster-related social media messages. Int. J. Geogr. Inf. Sci. 2023, 37, 2289–2318. [Google Scholar] [CrossRef]
  43. Li, S.; Wang, Y.; Huang, H.; Huang, L.; Chen, Y. Study on typhoon disaster assessment by mining data from social media based on artificial neural network. Nat. Hazards 2023, 116, 2069–2089. [Google Scholar] [CrossRef]
  44. Qu, Z.; Wang, J.; Zhang, M. Mining and analysis of public sentiment during disaster events: The extreme rainstorm disaster in megacities of China in 2021. Heliyon 2023, 9, e18272. [Google Scholar] [CrossRef]
  45. Wu, W.; Li, J.; He, Z.; Ye, X.; Zhang, J.; Cao, X.; Qu, H. Tracking spatio-temporal variation of geo-tagged topics with social media in China: A case study of 2016 hefei rainstorm. Int. J. Disaster Risk Reduct. 2020, 50, 101737. [Google Scholar] [CrossRef]
  46. Karimiziarani, M.; Jafarzadegan, K.; Abbaszadeh, P.; Shao, W.; Moradkhani, H. Hazard risk awareness and disaster management: Extracting the information content of twitter data. Sustain. Cities Soc. 2022, 77, 103577. [Google Scholar] [CrossRef]
  47. Kankanamge, N.; Yigitcanlar, T.; Goonetilleke, A.; Kamruzzaman, M. Determining disaster severity through social media analysis: Testing the methodology with South East Queensland Flood tweets. Int. J. Disaster Risk Reduct. 2020, 42, 101360. [Google Scholar] [CrossRef]
  48. Qian, J.; Du, Y.; Liang, F.; Yi, J.; Wang, N.; Tu, W.; Huang, S.; Pei, T.; Ma, T. Quantifying Urban Linguistic Diversity Related to Rainfall and Flood across China with Social Media Data. ISPRS Int. J. Geo-Inf. 2024, 13, 92. [Google Scholar] [CrossRef]
  49. Ghafarian, S.H.; Yazdi, H.S. Identifying crisis-related informative tweets using learning on distributions. Inf. Process. Manag. 2020, 57, 102145. [Google Scholar] [CrossRef]
  50. Zhi, G.; Meng, B.; Wang, J.; Chen, S.; Tian, B.; Ji, H.; Yang, T.; Wang, B.; Liu, J. Spatial Analysis of Urban Residential Sensitivity to Heatwave Events: Case Studies in Five Megacities in China. Remote Sens. 2021, 13, 4086. [Google Scholar] [CrossRef]
  51. Zhang, Y.; Chen, Z.; Zheng, X.; Chen, N.; Wang, Y. Extracting the location of flooding events in urban systems and analyzing the semantic risk using social sensing data. J. Hydrol. 2021, 603, 127053. [Google Scholar] [CrossRef]
  52. Wu, M.; Long, R.; Chen, F.; Chen, H.; Bai, Y.; Cheng, K.; Huang, H. Spatio-temporal difference analysis in climate change topics and sentiment orientation: Based on LDA and BiLSTM model. Resour. Conserv. Recycl. 2023, 188, 106697. [Google Scholar] [CrossRef]
  53. Wang, S.; Li, X.; Liao, G.; Liu, J.; Liao, C.; Liu, M.; Liao, J.; Liu, L. A spatio-temporal graph neural network for fall prediction with inertial sensors. Knowl.-Based Syst. 2024, 293, 111709. [Google Scholar] [CrossRef]
  54. Khattar, A.; Quadri, S.M.K. Generalization of convolutional network to domain adaptation network for classification of disaster images on twitter. Multimed. Tools Appl. 2022, 81, 30437–30464. [Google Scholar] [CrossRef]
  55. Alam, F.; Ofli, F.; Imran, M.; Alam, T.; Qazi, U. Deep learning benchmarks and datasets for social media image classification for disaster response. In Proceedings of the 12th IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, The Hague, The Netherlands, 7–10 December 2020; pp. 151–158. [Google Scholar]
  56. Alam, F.; Alam, T.; Hasan, M.A.; Hasnat, A.; Imran, M.; Ofli, F. MEDIC: A multi-task learning dataset for disaster image classification. Neural Comput. Appl. 2023, 35, 2609–2632. [Google Scholar] [CrossRef]
  57. Alam, F.; Ofli, F.; Imran, M. Processing social media images by combining human and machine computing during crises. Int. J. Hum.-Comput. Interact. 2018, 34, 311–327. [Google Scholar] [CrossRef]
  58. Li, X.; Caragea, C.; Caragea, D.; Imran, M.; Ofli, F. Identifying disaster damage images using a domain adaptation approach. In Proceedings of the 16th International Conference on Information Systems for Crisis Response and Management, Valencia, Spain, 19–22 May 2019; pp. 633–645. [Google Scholar]
  59. Li, X.; Caragea, D.; Zhang, H.; Imran, M. Localizing and quantifying damage in social media images. In Proceedings of the 10th IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Barcelona, Spain, 28–31 August 2018; pp. 194–201. [Google Scholar]
  60. Nguyen, D.T.; Ofli, F.; Imran, M.; Mitra, P. Damage assessment from social media imagery data during disasters. In Proceedings of the 9th IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Sydney, NSW, Australia, 31 July–3 August 2017; pp. 569–576. [Google Scholar]
  61. Pennington, C.V.L.; Bossu, R.; Ofli, F.; Imran, M.; Qazi, U.; Roch, J.; Banks, V.J. A near-real-time global landslide incident reporting tool demonstrator using social media and artificial intelligence. Int. J. Disaster Risk Reduct. 2022, 77, 103089. [Google Scholar] [CrossRef]
  62. Abavisani, M.; Wu, L.; Hu, S.; Tetreault, J.; Jaimes, A. Multimodal categorization of crisis events in social media. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020; pp. 14679–14689. [Google Scholar]
  63. Liang, T.; Lin, G.; Wan, M.; Li, T.; Ma, G.; Lv, F. Expanding large pre-trained unimodal models with multimodal information injection for image-text multimodal classification. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 15492–15501. [Google Scholar]
  64. Wu, Y.; Chen, Y.; Zhang, R.; Cui, Z.; Liu, X.; Zhang, J.; Wang, M.; Wu, Y. A Spatial Information Extraction Method Based on Multi-Modal Social Media Data: A Case Study on Urban Inundation. ISPRS Int. J. Geo-Inf. 2023, 12, 368. [Google Scholar] [CrossRef]
  65. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  66. Zhang, S.; Zheng, D.; Hu, X.; Yang, M. Bidirectional long short-term memory networks for relation classification. In Proceedings of the 29th Pacific Asia Conference on Language, Information and Computation, Shanghai, China, 30 October–1 Novermber 2015; pp. 73–78. [Google Scholar]
  67. Huang, Z.; Xu, W.; Yu, K. Bidirectional LSTM-CRF models for sequence tagging. arXiv 2015, arXiv:1508.01991. [Google Scholar]
  68. Qin, B.; Jin, Z.; Wang, H.; Pan, J.; Liu, Y.; An, B. Knowledge Graph and Semantic Computing: Knowledge Graph Empowers New Infrastructure Construction; Springer Nature: Singapore, 2021. [Google Scholar]
  69. QGIS Geographic Information System. Available online: https://www.qgis.org/ (accessed on 1 May 2024).
  70. Tan, L.; Schultz, D.M. Damage classification and recovery analysis of the Chongqing, China, floods of August 2020 based on social-media data. J. Clean Prod. 2021, 313, 127882. [Google Scholar] [CrossRef]
  71. Paradkar, A.S.; Zhang, C.; Yuan, F.; Mostafavi, A. Examining the consistency between geo-coordinates and content-mentioned locations in tweets for disaster situational awareness: A Hurricane Harvey study. Int. J. Disaster Risk Reduct. 2022, 73, 102878. [Google Scholar] [CrossRef]
  72. Alam, F.; Ofli, F.; Imran, M. CrisisMMD: Multimodal twitter datasets from natural disasters. In Proceedings of the 12th International AAAI Conference on Web and Social Media, Palo Alto, CA, USA, 25–28 June 2018; pp. 456–473. [Google Scholar]
  73. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
  74. Wei, J.; Zou, K. EDA: Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China, 3–7 November 2019; pp. 6382–6388. [Google Scholar]
  75. Lin, T.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef]
  76. Guangdong Statistical Yearbook 2020. Available online: http://stats.gd.gov.cn/gdtjnj/content/post_3098041.html (accessed on 1 May 2024).
  77. 2020 Weibo User Development Report. Available online: https://data.weibo.com/report/reportDetail?id=456 (accessed on 1 May 2024).
  78. Anselin, L.; Syabri, I.; Kho, Y. GeoDa: An introduction to spatial data analysis. In Handbook of Applied Spatial Analysis: Software Tools, Methods and Applications; Springer: Berlin/Heidelberg, Germany, 2009; pp. 73–89. [Google Scholar]
  79. Continuous Heavy Rainfall Has Caused 4 Deaths and 10 People Are Still Missing in Guangdong. Available online: http://www.xinhuanet.com/20240422/495244acd6d04c06977288c0e7a2927c/c.html (accessed on 1 May 2024).
Figure 1. The framework of the proposed social sensing model for spatio-temporal analysis and assessment of disasters for this study.
Figure 1. The framework of the proposed social sensing model for spatio-temporal analysis and assessment of disasters for this study.
Sensors 24 05889 g001
Figure 2. Workflow diagrams for (a) Llama3, (b) BERT-BiLSTM-CRF, and (c) the multimodal models.
Figure 2. Workflow diagrams for (a) Llama3, (b) BERT-BiLSTM-CRF, and (c) the multimodal models.
Sensors 24 05889 g002
Figure 3. The ROC curves and AUC values of (a) unimodal model on task 1, (b) multimodal model on task 1, (c) unimodal model on task 2, (d) multimodal model on task 2, (e) unimodal model on task 3, and (f) multimodal model on task 3.
Figure 3. The ROC curves and AUC values of (a) unimodal model on task 1, (b) multimodal model on task 1, (c) unimodal model on task 2, (d) multimodal model on task 2, (e) unimodal model on task 3, and (f) multimodal model on task 3.
Sensors 24 05889 g003
Figure 4. Distributions of (a) text length and (b) number of images for the collected Weibo data.
Figure 4. Distributions of (a) text length and (b) number of images for the collected Weibo data.
Sensors 24 05889 g004
Figure 5. Daily distributions of (a) precipitation, (b) the total number of Weibo posts, (c) the number of unimodal Weibo posts, and (d) the number of multimodal Weibo posts.
Figure 5. Daily distributions of (a) precipitation, (b) the total number of Weibo posts, (c) the number of unimodal Weibo posts, and (d) the number of multimodal Weibo posts.
Sensors 24 05889 g005
Figure 6. Spatial distributions of (a) precipitation, (b) the total number of Weibo posts, (c) the number of multimodal check-in posts, (d) the number of unimodal check-in posts, (e) the number of multimodal non-check-in posts, and (f) the number of unimodal non-check-in posts. The blue dots represent the actual Weibo address.
Figure 6. Spatial distributions of (a) precipitation, (b) the total number of Weibo posts, (c) the number of multimodal check-in posts, (d) the number of unimodal check-in posts, (e) the number of multimodal non-check-in posts, and (f) the number of unimodal non-check-in posts. The blue dots represent the actual Weibo address.
Sensors 24 05889 g006
Figure 7. Number of different categories in (a) task 1, (b) task 2, and (c) task 3.
Figure 7. Number of different categories in (a) task 1, (b) task 2, and (c) task 3.
Sensors 24 05889 g007
Figure 8. Daily varying counts of disaster categories in (a) task 2 and (b) task 3.
Figure 8. Daily varying counts of disaster categories in (a) task 2 and (b) task 3.
Sensors 24 05889 g008
Figure 9. Intraday variation in the counts of disaster categories for (a) task 2 and (b) task 3.
Figure 9. Intraday variation in the counts of disaster categories for (a) task 2 and (b) task 3.
Sensors 24 05889 g009
Figure 10. Spatial distributions of disaster categories in (a) task 2 and (b) task 3.
Figure 10. Spatial distributions of disaster categories in (a) task 2 and (b) task 3.
Sensors 24 05889 g010
Figure 11. Visualization of semantic transitions across the three tasks.
Figure 11. Visualization of semantic transitions across the three tasks.
Sensors 24 05889 g011
Table 1. Comparison of studies related to the spatio-temporal analysis of disaster-related social media.
Table 1. Comparison of studies related to the spatio-temporal analysis of disaster-related social media.
ReferenceStudy AreaModalityParadigmGeo InformationMean F1-Score
Yan et al. [29]ProvinceText + imageSupervisedExtracted0.7901
Wang et al. [26]ProvinceTextUnsupervisedCheck-in-
Li et al. [27]CityTextUnsupervisedCheck-in-
Peng et al. [28]CityTextSupervisedExtracted + Check-in0.7529
Li et al. [43]ProvinceTextSupervisedCheck-in0.7044
Qu et al. [44]CityTextUnsupervisedCheck-in-
Wu et al. [45]CityTextUnsupervisedCheck-in-
Karimiziarani et al. [46]ProvinceTextUnsupervisedCheck-in-
This studyProvinceText + imageSupervisedExtracted + Check-in0.8191
Table 2. Examples of location entity extraction from Weibo texts.
Table 2. Examples of location entity extraction from Weibo texts.
TextNERNER + LLM
深圳不愧是打工城市,每天到上班这个点都要下个大暴雨
Shenzhen is a city of laborers, and it rains heavily every day when I go to work.
深圳 (Shenzhen),
打工城市 (city of laborers)
深圳 (Shenzhen)
深圳暴雨如注,武汉阳光明媚
Heavy rain in Shenzhen, sunny in Wuhan.
武汉阳光明媚 (sunny in Wuhan),
深圳 (Shenzhen)
武汉 (Wuhan),
深圳 (Shenzhen)
《韶关江湾镇六个村受山体滑坡影响有群众被困,村民:独自在家的母亲仍失联》4月21日,受强降雨影响,韶关市武江区江湾镇遭受洪涝灾害
«Six villages in Jiangwan Town, Shaoguan are affected by landslides and people are trapped» On 21 April, Jiangwan Town, Wujiang District, Shaoguan City, was affected by flooding due to heavy rainfall.
韶关 (Shaoguan),
韶关市 (Shaoguan City),
江湾镇 (Jiangwan Town),
武江区 (Wujiang District),
六个村 (six villages)
韶关 (Shaoguan),
江湾镇 (Jiangwan Town),
武江区 (Wujiang District)
#女生自驾去广东一路暴雨带闪电#
# Girls driving to Guangdong all the way in torrential rain and lightning #
广东一路 (Guangdong road)广东 (Guangdong)
Table 3. Performance comparison on location entity recognition.
Table 3. Performance comparison on location entity recognition.
ModelPrecisionRecallF1-Score
BERT-BiLSTM-CRF0.98070.52500.6798
BERT-BiLSTM-CRF + LLM0.98460.86500.9186
Table 4. Definitions of disaster categories for the three tasks.
Table 4. Definitions of disaster categories for the three tasks.
TaskCategory
Informativeness task (Task 1)Informative (Class 0)
Not informative (Class 1)
Humanitarian categorization task (Task 2)Infrastructure, utility, or vehicle damage (Class 0)
Other relevant information (Class 1)
Rescue, volunteering, or donation effort (Class 2)
Affected individuals (injured, dead, missing, found) (Class 3)
Not humanitarian (Class 4)
Damage assessment task
(Task 3)
Mild, little, or no damage (Class 0)
Severe damage (Class 1)
Table 5. Detailed information on all models and train datasets.
Table 5. Detailed information on all models and train datasets.
Llama 3BERT-BiLSTM-CRFMultimodal Model
Data sourceManual labeledCCKS 2021 NER challengeCrisisMMD dataset
Data formatText and labelText and labelText–image pair and label
Dataset split200 (testing)8854 (training), 1972 (testing)Task 1: 8814/1101/1103
Task 2: 5503/686/691
Task 3: 2420/302/305
Backbone layers32 layersBERT: 12 layers
BiLSTM: 2 layers
CRF: 1 layer
BERT: 12 layers
DenseNet: 121 layers
Classifier: 1 layer
DeploymentPre-trainedPre-trained + Fine-tunedPre-trained + Fine-tuned
Model selection-5-fold cross-validationHold-out method
Learning rate-3 × 10−5 for BERT, 3 × 10−3 for the rest2 × 10−5
Optimizer-AdamWAdamW
Epochs-1010
Table 6. Performance comparison on the three disaster information classification tasks.
Table 6. Performance comparison on the three disaster information classification tasks.
TaskModalityPrecisionRecallF1-Score
1Multimodal0.90120.90210.9013
Unimodal0.83150.83410.8321
2Multimodal0.84410.84230.8419
Unimodal0.81810.81620.8163
3Multimodal0.71810.71150.7140
Unimodal0.61060.60980.6102
Table 7. Performance of the multimodal model on real-world Weibo data.
Table 7. Performance of the multimodal model on real-world Weibo data.
TaskPrecisionRecallF1-ScoreAUC
10.88210.87500.87130.9339
20.88700.88500.87960.9219
30.89660.71500.78080.7102
Table 8. Some samples and the corresponding predicted labels.
Table 8. Some samples and the corresponding predicted labels.
TextsImagesPredictions
1One day in Chaozhou, because of the heavy rain, we played at a friend’s house all day and learnt to play mahjong.Sensors 24 05889 i0011. Not informative
2. Not humanitarian
3. Mild, little, or no damage
2Half-day trip to Zhaoqing. I was caught in the biggest rainstorm ever on the highway.Sensors 24 05889 i0021. Informative
2. Not humanitarian
3. Mild, little, or no damage
3Aerial photograph of Hanguang town suffering from flooding. Affected by sustained heavy rainfall, Shaoguan and Qingyuan in northern Guangdong province are flooded.Sensors 24 05889 i0031. Informative
2. Infrastructure, utility, or vehicle damage
3. Severe damage
4Chinese Red Cross Foundation provides 2000 relief boxes to support flood-stricken areas in Guangdong.Sensors 24 05889 i0041. Informative
2. Rescue, volunteering, or donation effort
3. Mild, little, or no damage
5Heavy rains in Guangdong have resulted in 4 deaths and 10 missing. (The meaning of the Chinese characters in the figure is: “News Express @ ChinaNet Live”.)Sensors 24 05889 i0051. Informative
2. Affected individuals
3. Mild, little, or no damage
6In the past 24 h, the accumulated rainfall in Jieyang has reached torrential levels, and the rain is still ongoing. (The meaning of the Chinese title in the figure is: “Radar image at 42 min on 28 April 2024”.)Sensors 24 05889 i0061. Informative
2. Other relevant information
3. Mild, little, or no damage
Table 9. Temporal correlations between precipitation and Weibo Counts.
Table 9. Temporal correlations between precipitation and Weibo Counts.
Factor 1Factor 2PCCp-Value
PrecipitationNumber of Weibo0.63670.0079
Number of unimodal Weibo0.61250.0116
Number of multimodal Weibo0.66490.0049
Table 10. Spatial correlations between precipitation and Weibo Counts.
Table 10. Spatial correlations between precipitation and Weibo Counts.
Factor 1Factor 2Moran’s Ip-ValueZ-Score
PrecipitationRweibo0.3650.0023.1154
Rweibo (non-check-in)0.3360.0033.0655
Rweibo (check-in)0.3360.0032.7901
Rweibo (unimodal)0.3390.0022.8922
Rweibo (unimodal, non-check-in)0.3270.0052.7828
Rweibo (unimodal, check-in)0.3100.0052.6523
Rweibo (multimodal)0.3970.0013.4841
Rweibo (multimodal, non-check-in)0.3400.0012.9808
Rweibo (multimodal, check-in)0.3460.0013.0017
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, C.; Wang, Z. Multimodal Social Sensing for the Spatio-Temporal Evolution and Assessment of Nature Disasters. Sensors 2024, 24, 5889. https://doi.org/10.3390/s24185889

AMA Style

Yu C, Wang Z. Multimodal Social Sensing for the Spatio-Temporal Evolution and Assessment of Nature Disasters. Sensors. 2024; 24(18):5889. https://doi.org/10.3390/s24185889

Chicago/Turabian Style

Yu, Chen, and Zhiguo Wang. 2024. "Multimodal Social Sensing for the Spatio-Temporal Evolution and Assessment of Nature Disasters" Sensors 24, no. 18: 5889. https://doi.org/10.3390/s24185889

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop