Neural relational inference for disaster multimedia retrieval
Research output: Journal contributions › Journal articles › Research › peer-review
Standard
In: Multimedia Tools and Applications, Vol. 79, No. 35-36, 09.2020, p. 26735-26746.
Research output: Journal contributions › Journal articles › Research › peer-review
Harvard
APA
Vancouver
Bibtex
}
RIS
TY - JOUR
T1 - Neural relational inference for disaster multimedia retrieval
AU - Fadel, Samuel G.
AU - Torres, Ricardo da S.
PY - 2020/9
Y1 - 2020/9
N2 - Events around the world are increasingly documented on social media, especially by the people experiencing them, as these platforms become more popular over time. As a consequence, social media turns into a valuable source of data for understanding those events. Due to their destructive potential, natural disasters are among events of particular interest to response operations and environmental monitoring agencies. However, this amount of information also makes it challenging to identify relevant content pertaining to those events. In this paper, we use a relational neural network model for identifying this type of content. The model is particularly suitable for unstructured text, that is, text with no particular arrangement of words, such as tags, which is commonplace in social media data. In addition, our method can be combined with a CNN for handling multimodal data where text and visual data are available. We perform experiments in three different scenarios, where different modalities are evaluated: visual, textual, and both. Our method achieves competitive performance in both modalities by themselves, while significantly outperforms the baseline on the multimodal scenario. We also demonstrate the behavior of the proposed method in different applications by performing additional experiments in the CUB-200-2011 multimodal dataset.
AB - Events around the world are increasingly documented on social media, especially by the people experiencing them, as these platforms become more popular over time. As a consequence, social media turns into a valuable source of data for understanding those events. Due to their destructive potential, natural disasters are among events of particular interest to response operations and environmental monitoring agencies. However, this amount of information also makes it challenging to identify relevant content pertaining to those events. In this paper, we use a relational neural network model for identifying this type of content. The model is particularly suitable for unstructured text, that is, text with no particular arrangement of words, such as tags, which is commonplace in social media data. In addition, our method can be combined with a CNN for handling multimodal data where text and visual data are available. We perform experiments in three different scenarios, where different modalities are evaluated: visual, textual, and both. Our method achieves competitive performance in both modalities by themselves, while significantly outperforms the baseline on the multimodal scenario. We also demonstrate the behavior of the proposed method in different applications by performing additional experiments in the CUB-200-2011 multimodal dataset.
KW - Information retrieval
KW - Machine learning
KW - Multimodal
KW - Natural language processing
KW - Neural networks
KW - Business informatics
UR - http://www.scopus.com/inward/record.url?scp=85088125041&partnerID=8YFLogxK
U2 - 10.1007/s11042-020-09272-z
DO - 10.1007/s11042-020-09272-z
M3 - Journal articles
AN - SCOPUS:85088125041
VL - 79
SP - 26735
EP - 26746
JO - Multimedia Tools and Applications
JF - Multimedia Tools and Applications
SN - 1380-7501
IS - 35-36
ER -