TY - GEN
T1 - MOKD: Cross-domain Finetuning for Few-shot Classification via Maximizing Optimized Kernel Dependence
AU - Tian, Hongduan
AU - Liu, Feng
AU - Liu, Tongliang
AU - Du, Bo
AU - Cheung, Yiu-ming
AU - Han, Bo
N1 - HDT and BH were supported by the NSFC General Program No. 62376235, Guangdong Basic and Applied Basic Research Foundation Nos. 2022A1515011652 and 2024A1515012399, HKBU Faculty Niche Research Areas No. RC-FNRA-IG/22-23/SCI/04, and HKBU CSD Departmental Incentive Grant. FL was supported by the Australian Research Council with grant numbers DP230101540 and DE240101089, and the NSF & CSIRO Responsible AI program with grant number 2303037. BD was supported by the National Natural Science Foundation of China under Grants 62225113, the National Key Research and Development Program of China 2023YFC2705700. YMC was supported in part by NSFC/Research Grants Council (RGC) Joint Research Scheme under Grant: N HKBU214/21; in part by the General Research Fund of RGC under Grants: 12201321, 12202622, and 12201323; and in part by RGC Senior Research Fellow Scheme under Grant: SRFS2324-2S02. TLL was partially supported by the following Australian Re- search Council projects: FT220100318, DP220102121, LP220100527, LP220200949, and IC190100031.
PY - 2024/7/21
Y1 - 2024/7/21
N2 - In cross-domain few-shot classification, nearest centroid classifier (NCC) aims to learn representations to construct a metric space where few-shot classification can be performed by measuring the similarities between samples and the prototype of each class. An intuition behind NCC is that each sample is pulled closer to the class centroid it belongs to while pushed away from those of other classes. However, in this paper, we find that there exist high similarities between NCC-learned representations of two samples from different classes. In order to address this problem, we propose a bi-level optimization framework, maximizing optimized kernel dependence (MOKD) to learn a set of class-specific representations that match the cluster structures indicated by labeled data of the given task. Specifically, MOKD first optimizes the kernel adopted in Hilbert-Schmidt independence criterion (HSIC) to obtain the optimized kernel HSIC (opt-HSIC) that can capture the dependence more precisely. Then, an optimization problem regarding the opt-HSIC is addressed to simultaneously maximize the dependence between representations and labels and minimize the dependence among all samples. Extensive experiments on Meta-Dataset demonstrate that MOKD can not only achieve better generalization performance on unseen domains in most cases but also learn better data representation clusters. The project repository of MOKD is available at: https://github.com/tmlr-group/MOKD.
AB - In cross-domain few-shot classification, nearest centroid classifier (NCC) aims to learn representations to construct a metric space where few-shot classification can be performed by measuring the similarities between samples and the prototype of each class. An intuition behind NCC is that each sample is pulled closer to the class centroid it belongs to while pushed away from those of other classes. However, in this paper, we find that there exist high similarities between NCC-learned representations of two samples from different classes. In order to address this problem, we propose a bi-level optimization framework, maximizing optimized kernel dependence (MOKD) to learn a set of class-specific representations that match the cluster structures indicated by labeled data of the given task. Specifically, MOKD first optimizes the kernel adopted in Hilbert-Schmidt independence criterion (HSIC) to obtain the optimized kernel HSIC (opt-HSIC) that can capture the dependence more precisely. Then, an optimization problem regarding the opt-HSIC is addressed to simultaneously maximize the dependence between representations and labels and minimize the dependence among all samples. Extensive experiments on Meta-Dataset demonstrate that MOKD can not only achieve better generalization performance on unseen domains in most cases but also learn better data representation clusters. The project repository of MOKD is available at: https://github.com/tmlr-group/MOKD.
UR - https://proceedings.mlr.press/v235/tian24b.html
UR - https://openreview.net/forum?id=fz9PaJNViP
UR - http://www.scopus.com/inward/record.url?scp=85203801465&partnerID=8YFLogxK
M3 - Conference proceeding
T3 - Proceedings of the International Conference on Machine Learning
SP - 48154
EP - 48185
BT - Proceedings of 41th International Conference on Machine Learning, ICML 2024
A2 - Salakhutdinov, Ruslan
A2 - Kolter, Zico
A2 - Heller, Katherine
A2 - Weller, Adrian
A2 - Oliver, Nuria
A2 - Scarlett, Jonathan
A2 - Berkenkamp, Felix
PB - ML Research Press
T2 - 41st International Conference on Machine Learning, ICML 2024
Y2 - 21 July 2024 through 27 July 2024
ER -