Hongmin Li, Doina Caragea, & Cornelia Caragea. (2021). Combining Self-training with Deep Learning for Disaster Tweet Classification. In Anouck Adrot, Rob Grace, Kathleen Moore, & Christopher W. Zobel (Eds.), ISCRAM 2021 Conference Proceedings – 18th International Conference on Information Systems for Crisis Response and Management (pp. 719–730). Blacksburg, VA (USA): Virginia Tech.
Abstract: Significant progress has been made towards automated classification of disaster or crisis related tweets using machine learning approaches. Deep learning models, such as Convolutional Neural Networks (CNN), domain adaptation approaches based on self-training, and approaches based on pre-trained language models, such as BERT, have been proposed and used independently for disaster tweet classification. In this paper, we propose to combine self-training with CNN and BERT models, respectively, to improve the performance on the task of identifying crisis related tweets in a target disaster where labeled data is assumed to be unavailable, while unlabeled data is available. We evaluate the resulting self-training models on three crisis tweet collections and find that: 1) the pre-trained language model BERTweet is better than the standard BERT model, when fine-tuned for downstream crisis tweets classification; 2) self-training can help improve the performance of the CNN and BERTweet models for larger unlabeled target datasets, but not for smaller datasets.
|
Tasneem, F., Chakraborty, S., & Chy, A. N. (2023). An Early Synthesis of Deep Neural Networks to Identify Multimodal Informative Disaster Tweets. In Jaziar Radianti, Ioannis Dokas, Nicolas Lalone, & Deepak Khazanchi (Eds.), Proceedings of the 20th International ISCRAM Conference (pp. 428–438). Omaha, USA: University of Nebraska at Omaha.
Abstract: Twitter is always worthwhile in facilitating communication during disasters. It helps in raising situational awareness and undertaking disaster control actions as quickly as possible to alleviate the miseries. But the noisy essence of Twitter causes difficulty in distinguishing relevant information from the heterogeneous contents. Therefore, extracting informative tweets is a substantial task to help in crisis intervention. Analyzing only the text or image content of the tweet often misses necessary insights which might be helpful during disasters. In this paper, we propose a multimodal framework to address the challenges of identifying informative crisis-related tweets containing both texts and images. Our presented approach incorporates an early fusion strategy of BERT-LSTM and ResNet50 networks which effectively learns from the joint representation of texts and images. The experiments and evaluation on the benchmark CrisisMMD dataset show that our fusion method surpasses the baseline by 7% and substantiates its potency over the unimodal systems.
|