Multi-modal Deep Learning Approach for Flood Detection
Description
In this paper we propose a multi-modal deep learning approach to detect floods in social media posts. Social media posts normally contain some metadata and/or visual information, therefore in order to detect the floods we use this information. The model is based on a Convolutional Neural Network which extracts the visual features and a bidirectional Long Short-Term Memory network to extract the semantic features from the textual metadata. We validate the method on images extracted from Flickr which contain both visual information and metadata and compare the results when using both, visual information only or metadata only. This work has been done in the context of the MediaEval Multimedia Satellite Task.
Files
07_Mediaeval_2017_paper_14.pdf
Files
(801.2 kB)
Name | Size | Download all |
---|---|---|
md5:dd5b28245359c6d2a0eb2c231bc53200
|
801.2 kB | Preview Download |