共 35 条
[1]
Borth D., Ji R., Chen T., Breuel T., Chang S.-F., Large-scale visual sentiment ontology and detectors using adjective noun pairs, pp. 223-232, (2013)
[2]
Lu D., Neves L., Carvalho V., Zhang N., Ji H., Visual attention model for name tagging in multimodal social media, pp. 1990-1999, (2018)
[3]
Sun L., Wang J., Su Y., Weng F., Sun Y., Zheng Z., Chen Y., RIVA: A pre-trained tweet multimodal model based on text-image relation for multimodal NER, pp. 1852-1862, (2020)
[4]
Xu N., Mao W., Multisentinet: A deep semantic network for multimodal sentiment analysis, pp. 2399-2402, (2017)
[5]
Xu N., Mao W., Chen G., A co-memory network for multimodal sentiment analysis, pp. 929-932, (2018)
[6]
Yu Y., Lin H., Meng J., Zhao Z., Visual and textual sentiment analysis of a microblog using deep convolutional neural networks, Algorithms, 9, 2, (2016)
[7]
Truong Q., Lauw H.W., Vistanet: Visual aspect attention network for multimodal sentiment analysis, pp. 305-312, (2019)
[8]
Gu Y., Yang K., Fu S., Chen S., Li X., Marsic I., Multimodal affective analysis using hierarchical attention strategy with word-level alignment, pp. 2225-2235, (2018)
[9]
Pham H., Liang Pu P., Manzini T., Morency L.-P., Poczos B., Found in translation: Learning robust joint representations by cyclic translations between modalities, pp. 6892-6899, (2019)
[10]
Zhang J., Yu Y., Tang S., Wu J., Li W., Variational autoencoder with CCA for audio-visual cross-modal retrieval, Corr Abs/2112.02601., (2021)