TY - JOUR
T1 - Cross-domain Few-shot Classification via Invariant-content Feature Reconstruction
AU - Tian, Hongduan
AU - Liu, Feng
AU - Cheung, Ka Chun
AU - Fang, Zhen
AU - See, Simon
AU - Liu, Tongliang
AU - Han, Bo
N1 - HDT and BH were supported by NSFC General Program No. 62376235, RGC Young Collaborative Research Grant No. C2005-24Y, RGC General Research Fund No. 12200725, Guangdong Basic and Applied Basic Research Foundation No. 2024A151501239, and NVIDIA Collaborative Research Grant. FL was supported by the ARC with grant numbers DE240101089, LP240100101, DP230101540 and the NSF&CSIRO Responsible AI program with grant number 2303037. ZF was supported by the ARC with grant number DE250100363. TLL was partially supported by the following Australian Research Council projects: FT220100318, DP220102121.
Open access funding provided by Hong Kong Baptist University Library.
Publisher Copyright:
© The Author(s) 2026.
PY - 2026/2
Y1 - 2026/2
N2 - In cross-domain few-shot classification (CFC), mainstream studies aim to train a simple module (e.g. a linear transformation head) to select or transform features (a.k.a., the high-level semantic features) for previously unseen domains with a few labeled training data available on top of a powerful pre-trained model. These studies usually assume that high-level semantic features are shared across these domains, and just simple feature selection or transformations are enough to adapt features to previously unseen domains. However, in this paper, we find that the simply transformed features are too general to fully cover the key content features regarding each class. Thus, we propose an effective method, invariant-content feature reconstruction (IFR), to train a simple module that simultaneously considers both high-level and fine-grained invariant-content features for the previously unseen domains. Specifically, the fine-grained invariant-content features are considered as a set of informative and discriminative features learned from a few labeled training data of tasks sampled from unseen domains and are extracted by retrieving features that are invariant to style modifications from a set of content-preserving augmented data in pixel level with an attention module. Extensive experiments on the Meta-Dataset benchmark show that IFR achieves good generalization performance on unseen domains, which demonstrates the effectiveness of the fusion of the high-level features and the fine-grained invariant-content features. Specifically, IFR improves the average accuracy on unseen domains by 1.6% and 6.5% respectively under two different cross-domain few-shot classification settings.
AB - In cross-domain few-shot classification (CFC), mainstream studies aim to train a simple module (e.g. a linear transformation head) to select or transform features (a.k.a., the high-level semantic features) for previously unseen domains with a few labeled training data available on top of a powerful pre-trained model. These studies usually assume that high-level semantic features are shared across these domains, and just simple feature selection or transformations are enough to adapt features to previously unseen domains. However, in this paper, we find that the simply transformed features are too general to fully cover the key content features regarding each class. Thus, we propose an effective method, invariant-content feature reconstruction (IFR), to train a simple module that simultaneously considers both high-level and fine-grained invariant-content features for the previously unseen domains. Specifically, the fine-grained invariant-content features are considered as a set of informative and discriminative features learned from a few labeled training data of tasks sampled from unseen domains and are extracted by retrieving features that are invariant to style modifications from a set of content-preserving augmented data in pixel level with an attention module. Extensive experiments on the Meta-Dataset benchmark show that IFR achieves good generalization performance on unseen domains, which demonstrates the effectiveness of the fusion of the high-level features and the fine-grained invariant-content features. Specifically, IFR improves the average accuracy on unseen domains by 1.6% and 6.5% respectively under two different cross-domain few-shot classification settings.
KW - Cross-domain
KW - Feature reconstruction
KW - Few-shot classification
KW - Invariant features
UR - https://www.scopus.com/pages/publications/105027296461
U2 - 10.1007/s11263-025-02601-5
DO - 10.1007/s11263-025-02601-5
M3 - Journal article
AN - SCOPUS:105027296461
SN - 0920-5691
VL - 134
JO - International Journal of Computer Vision
JF - International Journal of Computer Vision
IS - 2
M1 - 54
ER -