Drug-BERT : Pre-trained Language Model Specialized for Korean Drug Crime

被引:0
|
作者
Lee, Jeong Min [1 ,2 ]
Lee, Suyeon [3 ]
Byon, Sungwon [1 ]
Jung, Eui-Suk [1 ]
Baek, Myung-Sun [1 ,2 ]
机构
[1] Elect & Telecommun Res Inst, Daejeon, South Korea
[2] Univ Sci & Technol, Daejeon, South Korea
[3] Yonsei Univ, Dept Artificial Intelligence, Seoul, South Korea
来源
19TH IEEE INTERNATIONAL SYMPOSIUM ON BROADBAND MULTIMEDIA SYSTEMS AND BROADCASTING, BMSB 2024 | 2024年
关键词
drug slang; natural language processing; pre-trained language model; classification;
D O I
10.1109/BMSB62888.2024.10608314
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose Drug-BERT, a specialized pre-trained language model designed for detecting drug-related content in the Korean language. Given the severity of the current drug issue in South Korea, effective responses are imperative. Focusing on the distinctive features of drug slang, this study seeks to improve the identification and classification of drug-related posts on social media platforms. Recent drug slangs are gathered and used to collect drug-related posts, and the collected data is used to train the language model. The designed pre-trained model is DRUG-BERT. The results show that fine-tuned DRUG-BERT outperforms that of the comparative models, achieving a 99.43% accuracy in classifying drug-relevant posts. Drug-BERT presents a promising solution for combatting drug-related activities, contributing to proactive measures against drug crimes in the Korean context.
引用
收藏
页码:186 / 188
页数:3
相关论文
共 50 条
  • [1] A pre-trained BERT for Korean medical natural language processing
    Kim, Yoojoong
    Kim, Jong-Ho
    Lee, Jeong Moon
    Jang, Moon Joung
    Yum, Yun Jin
    Kim, Seongtae
    Shin, Unsub
    Kim, Young-Min
    Joo, Hyung Joon
    Song, Sanghoun
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [2] A pre-trained BERT for Korean medical natural language processing
    Yoojoong Kim
    Jong-Ho Kim
    Jeong Moon Lee
    Moon Joung Jang
    Yun Jin Yum
    Seongtae Kim
    Unsub Shin
    Young-Min Kim
    Hyung Joon Joo
    Sanghoun Song
    Scientific Reports, 12
  • [3] Author Correction: A pre-trained BERT for Korean medical natural language processing
    Yoojoong Kim
    Jong-Ho Kim
    Jeong Moon Lee
    Moon Joung Jang
    Yun Jin Yum
    Seongtae Kim
    Unsub Shin
    Young-Min Kim
    Hyung Joon Joo
    Sanghoun Song
    Scientific Reports, 13
  • [4] Patent classification with pre-trained Bert model
    Kahraman, Selen Yuecesoy
    Durmusoglu, Alptekin
    Dereli, Tuerkay
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2024, 39 (04): : 2485 - 2496
  • [5] CAM-BERT: Chinese Aerospace Manufacturing Pre-trained Language Model
    Dai, Jinchi
    Wang, Shengren
    Wang, Peiyan
    Li, Ruiting
    Chen, Jiaxin
    Li, Xinrong
    2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, : 361 - 365
  • [6] A pre-trained BERT for Korean medical natural language processing (vol 12, 13847, 2022)
    Kim, Yoojoong
    Kim, Jong-Ho
    Lee, Jeong Moon
    Jang, Moon Joung
    Yum, Yun Jin
    Kim, Seongtae
    Shin, Unsub
    Kim, Young-Min
    Joo, Hyung Joon
    Song, Sanghoun
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [7] μBERT: Mutation Testing using Pre-Trained Language Models
    Degiovanni, Renzo
    Papadakis, Mike
    2022 IEEE 15TH INTERNATIONAL CONFERENCE ON SOFTWARE TESTING, VERIFICATION AND VALIDATION WORKSHOPS (ICSTW 2022), 2022, : 160 - 169
  • [8] CANCN-BERT: A Joint Pre-Trained Language Model for Classical and Modern Chinese
    Ji, Zijing
    Wang, Xin
    Shen, Yuxin
    Rao, Guozheng
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3112 - 3116
  • [9] Hyperbolic Pre-Trained Language Model
    Chen, Weize
    Han, Xu
    Lin, Yankai
    He, Kaichen
    Xie, Ruobing
    Zhou, Jie
    Liu, Zhiyuan
    Sun, Maosong
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3101 - 3112
  • [10] BERT-Log: Anomaly Detection for System Logs Based on Pre-trained Language Model
    Chen, Song
    Liao, Hai
    APPLIED ARTIFICIAL INTELLIGENCE, 2022, 36 (01)