Drug-BERT : Pre-trained Language Model Specialized for Korean Drug Crime

被引:0
作者
Lee, Jeong Min [1 ,2 ]
Lee, Suyeon [3 ]
Byon, Sungwon [1 ]
Jung, Eui-Suk [1 ]
Baek, Myung-Sun [1 ,2 ]
机构
[1] Elect & Telecommun Res Inst, Daejeon, South Korea
[2] Univ Sci & Technol, Daejeon, South Korea
[3] Yonsei Univ, Dept Artificial Intelligence, Seoul, South Korea
来源
19TH IEEE INTERNATIONAL SYMPOSIUM ON BROADBAND MULTIMEDIA SYSTEMS AND BROADCASTING, BMSB 2024 | 2024年
关键词
drug slang; natural language processing; pre-trained language model; classification;
D O I
10.1109/BMSB62888.2024.10608314
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose Drug-BERT, a specialized pre-trained language model designed for detecting drug-related content in the Korean language. Given the severity of the current drug issue in South Korea, effective responses are imperative. Focusing on the distinctive features of drug slang, this study seeks to improve the identification and classification of drug-related posts on social media platforms. Recent drug slangs are gathered and used to collect drug-related posts, and the collected data is used to train the language model. The designed pre-trained model is DRUG-BERT. The results show that fine-tuned DRUG-BERT outperforms that of the comparative models, achieving a 99.43% accuracy in classifying drug-relevant posts. Drug-BERT presents a promising solution for combatting drug-related activities, contributing to proactive measures against drug crimes in the Korean context.
引用
收藏
页码:186 / 188
页数:3
相关论文
共 50 条
[41]   BERT4CTR: An Efficient Framework to Combine Pre-trained Language Model with Non-textual Features for CTR Prediction [J].
Wang, Dong ;
Salamatian, Kave ;
Xia, Yunqing ;
Deng, Weiwei ;
Zhang, Qi .
PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, :5039-5050
[42]   Robust Sentiment Classification of Metaverse Services Using a Pre-trained Language Model with Soft Voting [J].
Lee, Haein ;
Jung, Hae Sun ;
Lee, Seon Hong ;
Kim, Jang Hyun .
KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2023, 17 (09) :2334-2347
[43]   Migratable urban street scene sensing method based on vision language pre-trained model [J].
Zhang, Yan ;
Zhang, Fan ;
Chen, Nengcheng .
INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2022, 113
[44]   Electric Power Audit Text Classification With Multi-Grained Pre-Trained Language Model [J].
Meng, Qinglin ;
Song, Yan ;
Mu, Jian ;
Lv, Yuanxu ;
Yang, Jiachen ;
Xu, Liang ;
Zhao, Jin ;
Ma, Junwei ;
Yao, Wei ;
Wang, Rui ;
Xiao, Maoxiang ;
Meng, Qingyu .
IEEE ACCESS, 2023, 11 :13510-13518
[45]   FGeneBERT: function-driven pre-trained gene language model for metagenomics [J].
Duan, Chenrui ;
Zang, Zelin ;
Xu, Yongjie ;
He, Hang ;
Li, Siyuan ;
Liu, Zihan ;
Lei, Zhen ;
Zheng, Ju-Sheng ;
Li, Stan Z. .
BRIEFINGS IN BIOINFORMATICS, 2025, 26 (02)
[46]   NMT Enhancement based on Knowledge Graph Mining with Pre-trained Language Model [J].
Yang, Hao ;
Qin, Ying ;
Deng, Yao ;
Wang, Minghan .
2020 22ND INTERNATIONAL CONFERENCE ON ADVANCED COMMUNICATION TECHNOLOGY (ICACT): DIGITAL SECURITY GLOBAL AGENDA FOR SAFE SOCIETY!, 2020, :185-189
[47]   Prompting disentangled embeddings for knowledge graph completion with pre-trained language model [J].
Geng, Yuxia ;
Chen, Jiaoyan ;
Zeng, Yuhang ;
Chen, Zhuo ;
Zhang, Wen ;
Pan, Jeff Z. ;
Wang, Yuxiang ;
Xu, Xiaoliang .
EXPERT SYSTEMS WITH APPLICATIONS, 2025, 268
[48]   Better Few-Shot Text Classification with Pre-trained Language Model [J].
Chen, Zheng ;
Zhang, Yunchen .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 :537-548
[49]   Explainable reasoning over temporal knowledge graphs by pre-trained language model [J].
Li, Qing ;
Wu, Guanzhong .
INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (01)
[50]   Enhancing pre-trained language model by answering natural questions for event extraction [J].
Zhang, Yuxin ;
Han, Qing .
FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2025, 8