Sorry, you need to enable JavaScript to visit this website.

ESA: Expert-and-Samples-Aware Incremental Learning under Longtail Distribution

Citation Author(s):
Jie Mei, Jenq-Neng Hwang
Submitted by:
Jie Mei
Last updated:
6 April 2024 - 7:21pm
Document Type:
Poster
Document Year:
2024
Event:
Presenters:
Jie Mei
Paper Code:
MLSP-P29.5
 

Most works in class incremental learning (CIL) assume disjoint sets of classes as tasks. Although a few works deal with overlapped sets of classes, they either assume a balanced data distribution or assume a mild imbalanced distribution. Instead, in this paper, we explore one of the understudied real-world CIL settings where (1) different tasks can share some classes but with new data samples, and (2) the training data of each task follows a long-tail distribution. We call this setting CIL-LT. We hypothesize that previously trained classification heads possess prototype knowledge of seen classes and thus could help learn the new model. Therefore, we propose a method with the multi-expert idea and a dynamic weighting technique to deal with the exacerbated forgetting introduced by the long-tail distribution. Experiments show that the proposed method effectively improves the accuracy in the CIL-LT setup on MNIST, CIFAR10, and CIFAR100. Code and data splits will be released.

up
0 users have voted: