[1]刘展阳,刘进锋.基于因果推理的长尾类增量学习[J].计算机技术与发展,2025,(02):100-106.[doi:10.20165/j.cnki.ISSN1673-629X.2024.0328]
 LIU Zhan-yang,LIU Jin-feng.Long-tailed Class Incremental Learning Based on Casual Inference[J].,2025,(02):100-106.[doi:10.20165/j.cnki.ISSN1673-629X.2024.0328]
点击复制

基于因果推理的长尾类增量学习()

《计算机技术与发展》[ISSN:1006-6977/CN:61-1281/TN]

卷:
期数:
2025年02期
页码:
100-106
栏目:
人工智能
出版日期:
2025-02-10

文章信息/Info

Title:
Long-tailed Class Incremental Learning Based on Casual Inference
文章编号:
1673-629X(2025)02-0100-07
作者:
刘展阳刘进锋
宁夏大学 信息工程学院,宁夏 银川 750021
Author(s):
LIU Zhan-yangLIU Jin-feng
School of Information Engineering,Ningxia University,Yinchuan 750021,China
关键词:
因果推理类增量学习长尾学习灾难性遗忘深度学习
Keywords:
casual inferenceclass incremental learninglong-tailed learningcatastrophic forgettingdeep learning
分类号:
TP391
DOI:
10.20165/j.cnki.ISSN1673-629X.2024.0328
摘要:
针对传统的类增量学习方法只考虑到新任务数据的平衡分布,而忽略了现实世界中数据呈现长尾分布这一问题,提出一种因果推理方法。 该方法旨在增强长尾类增量学习的性能,使模型在长尾数据分布背景下既能够逐步学习新任务的知识,又能够缓解对旧任务知识的灾难性遗忘。 不同于现存的缺少基本理论的启发式方法,该方法通过追求由输入样本引起的直接因果效应,在训练中使用因果干预,在推理中使用反事实推理,以有效改善现存类增量学习模型在长尾分布数据集上的表现。 与当前常用但复杂的两阶段训练方式相比,该方法更加简单有效,既不需要进行额外的微调阶段,也不需要增加额外的网络层,并且能够轻松地整合到绝大多数类增量学习方法中,从而更好地处理长尾类增量学习场景。 在 CIFAR-100 和 ImageNet-Subset 数据集上进行的广泛实验验证了该方法的有效性,显示了它在各种长尾类增量学习设置中相较于之前最佳方法的优越性。
Abstract:
Aiming at the problem that traditional class-incremental learning methods only consider the balanced distribution of data for a new task while ignoring the long - tailed distribution of data in the real world,a method based on casual inference is proposed. The method aims to enhance the performance of long-tailed class incremental learning so that the model is able to both incrementally acquire the knowledge of a new task and mitigate catastrophic forgetting of knowledge for previous tasks in the context of long-tailed data distri-butions. Unlike extant heuristic methods lacking fundamental theory,the proposed method uses causal intervention in training and coun-terfactual reasoning in inference by pursuing direct causal effects induced by input samples, in order to effectively improve the performance of existing class incremental learning models on long-tailed distributed datasets. Compared to the current commonly used but complex two-stage training approaches,the proposed method is simpler and more effective,requiring neither additional fine-tuning stages nor additional network layers. It can be easily integrated into the majority of class incremental learning methods to better handle long-tailed class-incremental learning scenarios. Extensive experiments on the CIFAR-100 and ImageNet-Subset datasets validate the effectiveness of the proposed method,showing its superiority over state-of-the-art approaches in various long-tailed class incremental learning settings.
更新日期/Last Update: 2025-02-10