Seeing the Storm: Leveraging Multimodal LLMs for Disaster Social Media Video Filtering

Authors

DOI:

https://doi.org/10.59297/f9bnkx60

Keywords:

Social media, Crisis Informatics, Emergency Management, Multimodal data, Large Language Models, Video

Abstract

Emergency management relies on the rapid triage of information to respond appropriately to disaster events. Social media platforms can provide emergency managers with ground-level insights, and videos, in particular, offer an immersive medium for understanding public responses and on-the-ground conditions. However, the overwhelming volume of irrelevant or redundant videos complicates their use for emergency response. This paper investigates the use of multimodal large language models (MLLMs)–specifically the Gemini 1.5 flash model–to automate the identification of relevant videos shared on X (formerly Twitter) during hurricanes. We develop and evaluate a framework to test the accuracy of different prompting styles and question strategies. By identifying the most effective prompting techniques, this study lays the groundwork for a systematic approach to filtering social media videos, enabling emergency managers to focus on the most pertinent content and make timely, informed decisions.

Downloads

Download data is not yet available.

Downloads

Published

2025-05-11

How to Cite

Hughes, A. L., & Clark, H. (2025). Seeing the Storm: Leveraging Multimodal LLMs for Disaster Social Media Video Filtering. Proceedings of the International ISCRAM Conference. https://doi.org/10.59297/f9bnkx60

Similar Articles

1-10 of 152

You may also start an advanced similarity search for this article.

Most read articles by the same author(s)

1 2 3 4 5 6 7 8 9 10 > >>