计算机工程与应用 ›› 2023, Vol. 59 ›› Issue (14): 151-157.DOI: 10.3778/j.issn.1002-8331.2205-0385

• 模式识别与人工智能 • 上一篇    下一篇

基于特征分布学习的小样本类增量学习

姚光乐,祝钧桃,周文龙,张贵宇,张伟,张谦   

  1. 1.人工智能四川省重点实验室,四川 宜宾 643000
    2.成都理工大学 计算机与网络安全学院,成都 610059
    3.四川轻化工大学 自动化与信息工程学院,四川 宜宾 643000
    4.电子科技大学 信息与通信工程学院,成都 611731
    5.电子信息控制重点实验室,成都 610036
  • 出版日期:2023-07-15 发布日期:2023-07-15

Few-Shot Class-Incremental Learning Based on Feature Distribution Learning

YAO Guangle, ZHU Juntao, ZHOU Wenlong, ZHANG Guiyu, ZHANG Wei, ZHANG Qian   

  1. 1.Artificial Intelligence Key Laboratory of Sichuan Province, Yibin, Sichuan 643000, China
    2.School of Computer and Network Security, Chengdu University of Technology, Chengdu 610059, China
    3.School of Automation & Information Engineering, Sichuan University of Science & Engineering, Yibin, Sichuan 643000, China
    4.School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
    5.Science and Technology on Electronic Information Control Laboratory, Chengdu 610036, China
  • Online:2023-07-15 Published:2023-07-15

摘要: 关注了一个非常具有挑战性的问题:深度神经网络的小样本类增量学习。其中深度神经网络模型可以从少量的样本中逐步学习新知识,同时不会忘记已学习的旧知识。为了平衡模型对旧知识的记忆和对新知识的学习,提出了一个基于特征分布学习的小样本类增量学习方法。在基类上学习模型以获得一个性能良好的特征提取器,并使用每类的特征分布信息来表示知识。将已学习的知识与新类的特征一起映射到一个新的低维子空间中,以统一地回顾旧知识与学习新知识。在子空间内,还为每个新类生成了分类权值初始化,以提高模型对新类的适应性。大量实验表明,该方法可以有效地减轻模型对已学习知识的遗忘,同时提高模型对新知识的适应性。

关键词: 小样本类增量学习, 深度神经网络, 增量学习

Abstract: This paper focuses on a very challenging problem:few-shot class-incremental learning for deep neural networks, where the deep neural network model can gradually learn new knowledge from a small number of samples without forgetting the learned knowledge. To balance the model’s memory of old knowledge and learning of new knowledge, it proposes a few-shot class-incremental learning method based on feature distribution learning. First, it learns the model on the base classes to obtain a well-performing feature extractor and take the feature distribution information as the learned knowledge. Then, it maps the learned knowledge together with the features of the novel classes into a low-dimensional subspace to review old knowledge and learn new knowledge uniformly. Finally, within the subspace, it also generates classification weight initializations for each novel class to improve the adaptability of the model to novel classes. Extensive experiments show that the method can effectively alleviate the model’s forgetting of learned knowledge and improve the model’s adaptability to new knowledge.

Key words: few-shot class-incremental learning, deep neural network, incremental learning