共 27 条
- [11] Dosovitskiy A, Beyer L, Kolesnikov A, Et al., An image is worth 16×16 words: Transformers for image recognition at scale
- [12] Cao H, Wang Y, Chen J, Et al., Swin-Unet: Unet-like pure Transformer for medical image segmentation
- [13] Wang W, Xie E, Li X, Et al., Pyramid vision Transformer: a Versatile backbone for dense prediction without convolutions, Proceedings of the IEEE/ CVF International Conference on Computer Vision, pp. 568-578, (2021)
- [14] Xie E, Wang W, Yu Z, Et al., SegFormer: simple and efficient design for semantic segmentation with transformers
- [15] Petit O, Thome N, Rambour C, Et al., U-net Transformer: self and cross attention for medical image segmentation[C], International Workshop on Machine Learning in Medical Imaging, pp. 267-276, (2021)
- [16] Islam M A, Jia S, Bruce N D B., How much position information do convolutional neural networks encode?[J/ OL]
- [17] Chu X, Tian Z, Zhang B, Et al., Conditional positional encodings for vision transformers
- [18] Fu J, Liu J, Tian H, Et al., Dual attention network for scene segmentation, Proceedings of the IEEE/ CVF Conference on Computer Vision and Pattern Recognition, pp. 3146-3154, (2019)
- [19] Gutman D, Codella N C F, Celebi E, Et al., Skin lesion analysis toward melanoma detection: a challenge at the international symposium on biomedical imaging (ISBI) 2016, hosted by the international skin imaging collaboration (ISIC) [J/OL]
- [20] Codella N C F, Gutman D, Celebi M E, Et al., Skin lesion analysis toward melanoma detection: a challenge at the 2017 international symposium on biomedical imaging (ISBI), hosted by the international skin imaging collaboration (isic) [C], IEEE 15th International Symposium on Biomedical Imaging(ISBI 2018), pp. 168-172, (2018)