Getting Meta: A Multimodal Approach for Detecting Unsafe Conversations within Instagram Direct Messages of Youth

被引:7
作者
Ali S. [1 ]
Razi A. [2 ]
Kim S. [3 ]
Alsoubai A. [4 ]
Ling C. [1 ]
De Choudhury M. [3 ]
Wisniewski P.J. [4 ]
Stringhini G. [1 ]
机构
[1] Boston University, Boston, MA
[2] Drexel University, Philadelphia, PA
[3] Georgia Institute of Technology, Atlanta, GA
[4] Vanderbilt University, Nashville, TN
基金
美国国家科学基金会;
关键词
end-to-end encryption; ensemble models; instagram; machine learning; online risk detection; social media;
D O I
10.1145/3579608
中图分类号
学科分类号
摘要
Instagram, one of the most popular social media platforms among youth, has recently come under scrutiny for potentially being harmful to the safety and well-being of our younger generations. Automated approaches for risk detection may be one way to help mitigate some of these risks if such algorithms are both accurate and contextual to the types of online harms youth face on social media platforms. However, the imminent switch by Instagram to end-to-end encryption for private conversations will limit the type of data that will be available to the platform to detect and mitigate such risks. In this paper, we investigate which indicators are most helpful in automatically detecting risk in Instagram private conversations, with an eye on high-level metadata, which will still be available in the scenario of end-to-end encryption. Toward this end, we collected Instagram data from 172 youth (ages 13-21) and asked them to identify private message conversations that made them feel uncomfortable or unsafe. Our participants risk-flagged 28,725 conversations that contained 4,181,970 direct messages, including textual posts and images. Based on this rich and multimodal dataset, we tested multiple feature sets (metadata, linguistic cues, and image features) and trained classifiers to detect risky conversations. Overall, we found that the metadata features (e.g., conversation length, a proxy for participant engagement) were the best predictors of risky conversations. However, for distinguishing between risk types, the different linguistic and media cues were the best predictors. Based on our findings, we provide design implications for AI risk detection systems in the presence of end-to-end encryption. More broadly, our work contributes to the literature on adolescent online safety by moving toward more robust solutions for risk detection that directly takes into account the lived risk experiences of youth. © 2023 ACM.
引用
收藏
相关论文
共 134 条
[1]  
Agha Z., Chatlani N., Razi A., Wisniewski P., Towards conducting responsible research with teens and parents regarding online risks, Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1-8, (2020)
[2]  
Ali S., Razi A., Kim S., Alsoubai A., Gracie J., De Choudhury M., Wisniewski P.J., Stringhini G., Understanding the Digital Lives of Youth: Analyzing Media Shared within Safe Versus Unsafe Private Conversations on Instagram, pp. 1-14, (2022)
[3]  
Ishola Aliyu N., Musbau Dogo A., Olajumoke Ajibade F., Abdurauf T., Analysis of cyber bullying on facebook using text mining, Journal of Applied Artificial Intelligence, 1, 1, pp. 1-12, (2020)
[4]  
Alsoubai A., Caddle X.V., Doherty R., Taylor Koehler A., Sanchez E., De Choudhury M., Wisniewski P.J., MOSafely, Is That Sus? A Youth-Centric Online Risk Assessment Dashboard, Companion Publication of the 2022 Conference on Computer Supported Cooperative Work and Social Computing (Virtual Event, Taiwan) (CSCW'22 Companion), pp. 197-200, (2022)
[5]  
Alsoubai A., Song J., Razi A., Naher N., De Choudhury M., Wisniewski P.J., From 'Friends with Benefits' to 'Sextortion: ' A Nuanced Investigation of Adolescents' Online Sexual Risk Experiences, Proc. ACM Hum.-Comput. Interact., 6, CSCW2, (2022)
[6]  
Anderson P., Zuo Z., Yang L., Qu Y., An Intelligent Online Grooming Detection System Using AI Technologies, pp. 1-6, (2019)
[7]  
Ashraf S., Ahmed T., Machine Learning Shrewd Approach For An Imbalanced Dataset Conversion Samples, Journal of Engineering and Technology, 11, (2020)
[8]  
Badillo-Urquiola K., Smriti D., McNally B., Golub E., Bonsignore E., Wisniewski P.J., Stranger danger! social media app features co-designed with children to keep them safe online, pp. 394-406, (2019)
[9]  
Baltrusaitis T., Ahuja C., Morency L., Multimodal Machine Learning: A Survey and Taxonomy, IEEE Transactions on Pattern Analysis and Machine Intelligence, 41, 2, pp. 423-443, (2019)
[10]  
Barbieri F., Ballesteros M., Ronzano F., Saggion H., Multimodal Emoji Prediction, (2018)