Multilingual Semantic Sourcing using Product Images for Cross-lingual Alignment

被引:1
作者
Mangrulkar, Sourab [1 ]
Ankith, M. S. [1 ]
Sembium, Vivek [1 ]
机构
[1] Amazon, Bengaluru, Karnataka, India
来源
COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2022, WWW 2022 COMPANION | 2022年
关键词
Deep Learning; Multilingual; Multimodal; Semantic Sourcing; E-Commerce;
D O I
10.1145/3487553.3524204
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In online retail stores with ever-increasing catalog, product search is the primary means for customers to discover products of their interest. Surfacing irrelevant products can lead to poor customer experience and in extreme situations loss in engagement. With the recent advances in NLP, Deep Learning models are being used to represent queries and products in shared semantic space to enable semantic sourcing. These models require a lot of human annotated (query, product, relevance) tuples to give competitive results which is expensive to generate. The problem becomes more prominent in the emerging marketplaces/languages due to data paucity problem. When expanding to new marketplaces, it becomes imperative to support regional languages to reach a wider customer base and delighting them with good customer experience. Recently, in the NLP domain, approaches using parallel data corpus for training multilingual models have become prominent, but they are expensive to generate. In this work, we learn semantic alignment across languages using product images as an anchor between them. This overcomes the necessity of parallel data corpus. We use the human annotated data from established marketplace to transfer relevance classification knowledge to new/emerging marketplaces to solve the data paucity problem. Our experiments performed on datasets from Amazon reveal that we outperform state-of-the-art baselines with 2.4%-3.65% ROC-AUC lifts on relevance classification task across non-English marketplaces, 34.69%-51.67% Recall@k lifts on language-agnostic retrieval task and 6.25%-13.42% Precision@k lifts on semantic neighborhood quality task, respectively. Our models demonstrate efficient transfer of relevance classification knowledge from data rich marketplaces to new marketplaces by achieving ROC-AUC lifts of 3.74%-6.25% for the relevance classification task in the zero-shot setting where the human annotated relevance data of target marketplace is unavailable during training.
引用
收藏
页码:41 / 51
页数:11
相关论文
共 44 条
[1]   Language-Agnostic Representation Learning for Product Search on E-Commerce Platforms [J].
Ahuja, Aman ;
Rao, Nikhil ;
Katariya, Sumeet ;
Subbian, Karthik ;
Reddy, Chandan K. .
PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM '20), 2020, :7-15
[2]  
Bjerva J., 2017, NODALIDA
[3]   UNITER: UNiversal Image-TExt Representation Learning [J].
Chen, Yen-Chun ;
Li, Linjie ;
Yu, Licheng ;
El Kholy, Ahmed ;
Ahmed, Faisal ;
Gan, Zhe ;
Cheng, Yu ;
Liu, Jingjing .
COMPUTER VISION - ECCV 2020, PT XXX, 2020, 12375 :104-120
[4]  
Clark K., 2020, 8 INT C LEARNING REP, DOI [DOI 10.48550/ARXIV.2003.10555, 10.48550/arXiv.2003.10555]
[5]  
Conneau A., 2020, PROC ACL 2020 C, P8440
[6]  
Conneau A, 2019, ADV NEUR IN, V32
[7]   VirTex: Learning Visual Representations from Textual Annotations [J].
Desai, Karan ;
Johnson, Justin .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :11157-11168
[8]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[9]  
Dosovitskiy A, 2021, ICLR
[10]  
Li LH, 2019, Arxiv, DOI arXiv:1908.03557