Personalized Clothing Prediction Algorithm Based on Multi-modal Feature Fusion

被引:0
|
作者
Liu, Rong [1 ,2 ]
Joseph, Annie Anak [1 ]
Xin, Miaomiao [2 ]
Zang, Hongyan [2 ]
Wang, Wanzhen [2 ]
Zhang, Shengqun [2 ]
机构
[1] Univ Malaysia Sarawak, Fac Engn, Kota Samarahan, Sarawak, Malaysia
[2] Qilu Inst Technol, Comp & Informat Engn, Jinan, Peoples R China
关键词
fashion consumers; image; text data; personalized; multi-modal fusion;
D O I
10.46604/ijeti.2024.13394
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
With the popularization of information technology and the improvement of material living standards, fashion consumers are faced with the daunting challenge of making informed choices from massive amounts of data. This study aims to propose deep learning technology and sales data to analyze the personalized preference characteristics of fashion consumers and predict fashion clothing categories, thus empowering consumers to make well-informed decisions. The Visuelle's dataset includes 5,355 apparel products and 45 MB of sales data, and it encompasses image data, text attributes, and time series data. The paper proposes a novel 1DCNN-2DCNN deep convolutional neural network model for the multi-modal fusion of clothing images and sales text data. The experimental findings exhibit the remarkable performance of the proposed model, with accuracy, recall, F1 score, macro average, and weighted average metrics achieving 99.59%, 99.60%, 98.01%, 98.04%, and 98.00%, respectively. Analysis of four hybrid models highlights the superiority of this model in addressing personalized preferences.
引用
收藏
页码:216 / 230
页数:15
相关论文
共 50 条
  • [31] Enhancing multi-modal fusion in visual dialog via sample debiasing and feature interaction
    Lu, Chenyu
    Yin, Jun
    Yang, Hao
    Sun, Shiliang
    INFORMATION FUSION, 2024, 107
  • [32] Deformable Feature Fusion Network for Multi-Modal 3D Object Detection
    Guo, Kun
    Gan, Tong
    Ding, Zhao
    Ling, Qiang
    2024 3RD INTERNATIONAL CONFERENCE ON ROBOTICS, ARTIFICIAL INTELLIGENCE AND INTELLIGENT CONTROL, RAIIC 2024, 2024, : 363 - 367
  • [33] Semantic2Graph: graph-based multi-modal feature fusion for action segmentation in videos
    Junbin Zhang
    Pei-Hsuan Tsai
    Meng-Hsun Tsai
    Applied Intelligence, 2024, 54 : 2084 - 2099
  • [34] A Depression Detection Auxiliary Decision System Based on Multi-Modal Feature-Level Fusion of EEG and Speech
    Ning, Zhaolong
    Hu, Hao
    Yi, Ling
    Qie, Zihan
    Tolba, Amr
    Wang, Xiaojie
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (01) : 3392 - 3402
  • [35] Semantic2Graph: graph-based multi-modal feature fusion for action segmentation in videos
    Zhang, Junbin
    Tsai, Pei-Hsuan
    Tsai, Meng-Hsun
    APPLIED INTELLIGENCE, 2024, 54 (02) : 2084 - 2099
  • [36] On Multi-modal Fusion for Freehand Gesture Recognition
    Schak, Monika
    Gepperth, Alexander
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT I, 2020, 12396 : 862 - 873
  • [37] Multi-Modal Pedestrian Detection Algorithm Based on Deep Learning
    Li X.
    Fu H.
    Niu W.
    Wang P.
    Lü Z.
    Wang W.
    Hsi-An Chiao Tung Ta Hsueh/Journal of Xi'an Jiaotong University, 2022, 56 (10): : 61 - 70
  • [38] COMMUTING CONDITIONAL GANS FOR MULTI-MODAL FUSION
    Roheda, Siddharth
    Krim, Hamid
    Riggan, Benjamin S.
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3197 - 3201
  • [39] Multi-modal deep fusion based fake news detection method
    Jing Q.
    Fan X.
    Wang B.
    Bi J.
    Tan H.
    High Technology Letters, 2022, 32 (04) : 392 - 403
  • [40] Learning-Based Confidence Estimation for Multi-modal Classifier Fusion
    Nadeem, Uzair
    Bennamoun, Mohammed
    Sohel, Ferdous
    Togneri, Roberto
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT II, 2019, 11954 : 299 - 312