Multi-modal Representation Learning for Social Post Location Inference

被引:0
|
作者
Dai, RuiTing [1 ]
Luo, Jiayi [1 ]
Luo, Xucheng [1 ]
Mo, Lisi [1 ]
Ma, Wanlun [2 ]
Zhou, Fan [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[2] Swinburne Univ Technol, Melbourne, Vic, Australia
来源
ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS | 2023年
关键词
Social geographic location; multi-modal social post dataset; multi-modal representation learning; multi-head attention mechanism; PREDICTION;
D O I
10.1109/ICC45041.2023.10279649
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Inferring geographic locations via social posts is essential for many practical location-based applications such as product marketing, point-of-interest recommendation, and infector tracking for COVID-19. Unlike image-based location retrieval or social-post text embedding-based location inference, the combined effect of multi-modal information (i.e., post images, text, and hashtags) for social post positioning receives less attention. In this work, we collect real datasets of social posts with images, texts, and hashtags from Instagram and propose a novel Multi-modal Representation Learning Framework (MRLF) capable of fusing different modalities of social posts for location inference. MRLF integrates a multi-head attention mechanism to enhance location-salient information extraction while significantly improving location inference compared with single domain-based methods. To overcome the noisy user-generated textual content, we introduce a novel attention-based character-aware module that considers the relative dependencies between characters of social post texts and hashtags for flexible multimodel information fusion. The experimental results show that MRLF can make accurate location predictions and open a new door to understanding the multi-modal data of social posts for online inference tasks.
引用
收藏
页码:6331 / 6336
页数:6
相关论文
共 50 条
  • [1] Enhanced Topic Modeling with Multi-modal Representation Learning
    Zhang, Duoyi
    Wang, Yue
    Abul Bashar, Md
    Nayak, Richi
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT I, 2023, 13935 : 393 - 404
  • [2] Towards a systematic multi-modal representation learning for network data
    Ben Houidi, Zied
    Azorin, Raphael
    Gallo, Massimo
    Finamore, Alessandro
    Rossi, Dario
    THE 21ST ACM WORKSHOP ON HOT TOPICS IN NETWORKS, HOTNETS 2022, 2022, : 181 - 187
  • [3] Lightweight Multi-modal Representation Learning for RGB Salient Object Detection
    Xiao, Yun
    Huang, Yameng
    Li, Chenglong
    Liu, Lei
    Zhou, Aiwu
    Tang, Jin
    COGNITIVE COMPUTATION, 2023, 15 (06) : 1868 - 1883
  • [4] Lightweight Multi-modal Representation Learning for RGB Salient Object Detection
    Yun Xiao
    Yameng Huang
    Chenglong Li
    Lei Liu
    Aiwu Zhou
    Jin Tang
    Cognitive Computation, 2023, 15 : 1868 - 1883
  • [5] Incomplete multi-modal representation learning for Alzheimer's disease diagnosis
    Liu, Yanbei
    Fan, Lianxi
    Zhang, Changqing
    Zhou, Tao
    Xiao, Zhitao
    Geng, Lei
    Shen, Dinggang
    MEDICAL IMAGE ANALYSIS, 2021, 69
  • [6] Geographic mapping with unsupervised multi-modal representation learning from VHR images and POIs
    Bai, Lubin
    Huang, Weiming
    Zhang, Xiuyuan
    Du, Shihong
    Cong, Gao
    Wang, Haoyu
    Liu, Bo
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2023, 201 : 193 - 208
  • [7] Small molecule drug and biotech drug interaction prediction based on multi-modal representation learning
    Dingkai Huang
    Hongjian He
    Jiaming Ouyang
    Chang Zhao
    Xin Dong
    Jiang Xie
    BMC Bioinformatics, 23
  • [8] Small molecule drug and biotech drug interaction prediction based on multi-modal representation learning
    Huang, Dingkai
    He, Hongjian
    Ouyang, Jiaming
    Zhao, Chang
    Dong, Xin
    Xie, Jiang
    BMC BIOINFORMATICS, 2022, 23 (01)
  • [9] Turbo your multi-modal classification with contrastive learning
    Zhang, Zhiyu
    Liu, Da
    Liu, Shengqiang
    Wang, Anna
    Gao, Jie
    Li, Yali
    INTERSPEECH 2023, 2023, : 1848 - 1852
  • [10] Exploiting enhanced and robust RGB-D face representation via progressive multi-modal learning
    Zhu, Yizhe
    Gao, Jialin
    Wu, Tianshu
    Liu, Qiong
    Zhou, Xi
    PATTERN RECOGNITION LETTERS, 2023, 166 : 38 - 45