Supervised Multi-modal Dictionary Learning for Clothing Representation

被引:0
|
作者
Zhao, Qilu [1 ]
Wang, Jiayan [1 ]
Li, Zongmin [1 ]
机构
[1] China Univ Petr East China, 66 Changjiang West Rd, Qingdao, Peoples R China
来源
PROCEEDINGS OF THE FIFTEENTH IAPR INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS - MVA2017 | 2017年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Clothing appearances have complex visual properties, such as color, texture, shape and structure. Different modalities of visual features provide information complementary to each other. Combining multi modal visual features can lead to a comprehensive description of Clothing appearances. Meanwhile, categories provide sufficient semantic information, which can lead to discriminative representations. Clothing categories exhibit hierarchical structure, which could benefit the learning algorithm. In this paper, we propose a multi-view learning algorithm, named Supervised Multi-modal Dictionary Learning (SMMDL), which learns a latent space encoding multi-modal visual properties and semantic relationships between clothing samples. Experiments on the image classification task show that SMMDL outperforms baseline methods.
引用
收藏
页码:51 / 54
页数:4
相关论文
共 50 条
  • [1] Multi-Modal Convolutional Dictionary Learning
    Gao, Fangyuan
    Deng, Xin
    Xu, Mai
    Xu, Jingyi
    Dragotti, Pier Luigi
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 1325 - 1339
  • [2] Multi-modal Network Representation Learning
    Zhang, Chuxu
    Jiang, Meng
    Zhang, Xiangliang
    Ye, Yanfang
    Chawla, Nitesh, V
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 3557 - 3558
  • [3] Multi-Modal Knowledge Representation Learning via Webly-Supervised Relationships Mining
    Nian, Fudong
    Bao, Bing-Kun
    Li, Teng
    Xu, Changsheng
    PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 411 - 419
  • [4] Hierarchical sparse representation with deep dictionary for multi-modal classification
    Wang, Zhengxia
    Teng, Shenghua
    Liu, Guodong
    Zhao, Zengshun
    Wu, Hongli
    NEUROCOMPUTING, 2017, 253 : 65 - 69
  • [5] Mineral: Multi-modal Network Representation Learning
    Kefato, Zekarias T.
    Sheikh, Nasrullah
    Montresor, Alberto
    MACHINE LEARNING, OPTIMIZATION, AND BIG DATA, MOD 2017, 2018, 10710 : 286 - 298
  • [6] Scalable multi-modal representation learning networks
    Zihan Fang
    Ying Zou
    Shiyang Lan
    Shide Du
    Yanchao Tan
    Shiping Wang
    Artificial Intelligence Review, 58 (7)
  • [7] Continual Self-supervised Learning: Towards Universal Multi-modal Medical Data Representation Learning
    Ye, Yiwen
    Xie, Yutong
    Zhang, Jianpeng
    Chen, Ziyang
    Wu, Qi
    Xia, Yong
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 11114 - 11124
  • [8] Comprehensive Semi-Supervised Multi-Modal Learning
    Yang, Yang
    Wang, Ke-Tao
    Zhan, De-Chuan
    Xiong, Hui
    Jiang, Yuan
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4092 - 4098
  • [9] Multi-modal representation learning in retinal imaging using self-supervised learning for enhanced clinical predictions
    Suekei, Emese
    Rumetshofer, Elisabeth
    Schmidinger, Niklas
    Mayr, Andreas
    Schmidt-Erfurth, Ursula
    Klambauer, Guenter
    Bogunovic, Hrvoje
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [10] Fast Multi-Modal Unified Sparse Representation Learning
    Verma, Mridula
    Shukla, Kaushal Kumar
    PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR'17), 2017, : 448 - 452