Geometry-Adaptive Meta-Learning in Mixed-Curvature Spaces

被引:0
作者
Gao, Zhi [1 ]
Wu, Yu-Wei [1 ,2 ]
Jia, Yun-De [1 ,2 ]
机构
[1] Beijing Key Laboratory of Intelligent Information Technology, School of Computer Science & Technology, Beijing Institute of Technology, Beijing
[2] Guangdong Laboratory of Machine Perception and Intelligent Computing, Shenzhen MSU-BIT University, Guangdong, Shenzhen
来源
Jisuanji Xuebao/Chinese Journal of Computers | 2024年 / 47卷 / 10期
基金
中国国家自然科学基金;
关键词
geometry adaptation; meta-learning; mixed-curvature space; Riemannian manifold;
D O I
10.11897/SP.J.1016.2024.02289
中图分类号
学科分类号
摘要
Meta-learning has shown effectiveness in helping learning models quickly adapt to new tasks by learning prior knowledge. In the process of adaptation to new tasks, the matching degree between the geometric structure of space and the geometric structure of data plays an important role in the generalization ability of the model. In many practical applications, data has diverse non-Euclidean structures. For example, natural language has non-Euclidean hierarchical structures, and face images have non-Euclidean cyclical structures. Existing research has shown that the geometric structure of Riemannian manifolds matches the non-Euclidean structures of real-world data, providing theoretical feasibility for modeling data using Riemannian manifolds. In this paper, we propose a geometry-adaptive meta-learning method in mixed-curvature spaces, which uses multiple mixed-curvature spaces to model data and produces matching Riemannian geometry for non-Euclidean structures. We build a multi-mixed-curvature neural network that represents the geometry of mixed-curvature space as curvature, number, and dimensionality of the curvature spaces, through which the geometry adaptation to non-Euclidean structures is achieved via a gradient descent process. We further introduce a geometry initialization generation scheme and geometry updating scheme. Through only a few optimization steps, the geometric structure of the underlying space can quickly match non-Euclidean structures of data, accelerating the gradient descent process. We conduct experiments on few-shot classification, few-shot regression, and image completion to evaluate the effectiveness of our method. Compared with meta-learning methods in Euclidean space, our method improves the accuracy by 3% in few-shot classification tasks, and reduces mean square error by half in few-shot regression tasks, showing the effectiveness of our method. © 2024 Science Press. All rights reserved.
引用
收藏
页码:2289 / 2306
页数:17
相关论文
共 53 条
  • [1] Finn C, Abbeel P, Levine S., Model-agnostic meta-learning for fast adaptation of deep networks, Proceedings of the International Conference on Machine Learning, pp. 1126-1135, (2017)
  • [2] Baik S, Hong S, Lee K., Learning to forget for meta-learning, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2376-2384, (2020)
  • [3] Jamal M, Qi G, Shah M., Task agnostic meta-learning for few-shot learning, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11719-11727, (2019)
  • [4] Li Fan-Zhang, Liu Yang, Wu Peng-Xiang, Et al., A survey on recent advances in meta-learning, Chinese Journal of Computers, 44, 2, pp. 422-446, (2021)
  • [5] Rajeswaran A, Finn C, Kakade S, Et al., Meta-learning with implicit gradients, Advances in Neural Information Processing Systems, pp. 113-124, (2019)
  • [6] Zintgraf L, Shiarlis K, Kurin V, Et al., Fast context adaptation via meta-learning, Proceedings of the International Conference on Machine Learning, pp. 7693-7702, (2019)
  • [7] Li Peng-Fang, Liu Fang, Li Ling-Ling, Et al., Meta-feature relearning with embedded label semantics and reweighting for few-shot object detection, Chinese Journal of Computers, 45, 12, pp. 2561-2575, (2022)
  • [8] Baik S, Choi J, Kim H, Et al., Meta-learning with task-adaptive loss function for few-shot learning, Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9465-9474, (2021)
  • [9] Baik S, Choi M, Choi J, Et al., Meta-learning with adaptive hyperparameters, Advances in Neural Information Processing Systems, pp. 20755-20765, (2020)
  • [10] Jamal M, Wang L, Gong B., A lazy approach to long-horizon gradient-based metaTearning, Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6577-6586, (2021)