共 50 条
- [31] Improving Multi-head Attention with Capsule Networks NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 314 - 326
- [32] Classification of Heads in Multi-head Attention Mechanisms KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2022, PT III, 2022, 13370 : 681 - 692
- [33] Diversifying Multi-Head Attention in the Transformer Model MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (04): : 2618 - 2638
- [34] A new interest extraction method based on multi-head attention mechanism for CTR prediction Knowledge and Information Systems, 2023, 65 : 3337 - 3352
- [36] Multi-head Attention and Graph Convolutional Networks with Regularized Dropout for Biomedical Relation Extraction HEALTH INFORMATION PROCESSING, CHIP 2023, 2023, 1993 : 98 - 111
- [39] Acoustic Scene Analysis with Multi-head Attention Networks INTERSPEECH 2020, 2020, : 1191 - 1195
- [40] Enhancing Attention Models via Multi-head Collaboration 2020 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2020), 2020, : 19 - 23