R.I.P.
๐ป
Ghosted
Better Generalized Few-Shot Learning Even Without Base Data
November 29, 2022 ยท Entered Twilight ยท ๐ AAAI Conference on Artificial Intelligence
Repo contents: README.md, configs, datasets, materials, models, test_few_shot.py, train_classifier.py, utils, zero_base_GFSL_concept.png
Authors
Seong-Woong Kim, Dong-Wan Choi
arXiv ID
2211.16095
Category
cs.LG: Machine Learning
Cross-listed
cs.AI,
cs.CV
Citations
9
Venue
AAAI Conference on Artificial Intelligence
Repository
https://github.com/bigdata-inha/Zero-Base-GFSL
โญ 13
Last Checked
1 month ago
Abstract
This paper introduces and studies zero-base generalized few-shot learning (zero-base GFSL), which is an extreme yet practical version of few-shot learning problem. Motivated by the cases where base data is not available due to privacy or ethical issues, the goal of zero-base GFSL is to newly incorporate the knowledge of few samples of novel classes into a pretrained model without any samples of base classes. According to our analysis, we discover the fact that both mean and variance of the weight distribution of novel classes are not properly established, compared to those of base classes. The existing GFSL methods attempt to make the weight norms balanced, which we find helps only the variance part, but discard the importance of mean of weights particularly for novel classes, leading to the limited performance in the GFSL problem even with base data. In this paper, we overcome this limitation by proposing a simple yet effective normalization method that can effectively control both mean and variance of the weight distribution of novel classes without using any base samples and thereby achieve a satisfactory performance on both novel and base classes. Our experimental results somewhat surprisingly show that the proposed zero-base GFSL method that does not utilize any base samples even outperforms the existing GFSL methods that make the best use of base data. Our implementation is available at: https://github.com/bigdata-inha/Zero-Base-GFSL.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Machine Learning
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
R.I.P.
๐ป
Ghosted
Semi-Supervised Classification with Graph Convolutional Networks
R.I.P.
๐ป
Ghosted
Proximal Policy Optimization Algorithms
R.I.P.
๐ป
Ghosted