TY - JOUR
T1 - On the Learnability of Out-of-distribution Detection
AU - Fang, Zhen
AU - Li, Yixuan
AU - Liu, Feng
AU - Han, Bo
AU - Lu, Jie
N1 - JL and ZF were supported by the Australian Research Council (ARC) under FL190100149. YL is supported by National Science Foundation (NSF) Award No. IIS-2237037. FL was supported by the ARC with grant numbers DP230101540 and DE240101089, and the NSF&CSIRO Responsible AI program with grant number 2303037. ZF would also like to thank Prof. Peter Bartlett, Dr. Tongliang Liu and Dr. Zhiyong Yang for productive discussions.
Publisher Copyright:
© 2024 Zhen Fang, Yixuan Li, Feng Liu, Bo Han, Jie Lu.
PY - 2024/4
Y1 - 2024/4
N2 - Supervised learning aims to train a classifier under the assumption that training and test data are from the same distribution. To ease the above assumption, researchers have studied a more realistic setting: out-of-distribution (OOD) detection, where test data may come from classes that are unknown during training (i.e., OOD data). Due to the unavailability and diversity of OOD data, good generalization ability is crucial for effective OOD detection algorithms, and corresponding learning theory is still an open problem. To study the generalization of OOD detection, this paper investigates the probably approximately correct (PAC) learning theory of OOD detection that fits the commonly used evaluation metrics in the literature. First, we find a necessary condition for the learnability of OOD detection. Then, using this condition, we prove several impossibility theorems for the learnability of OOD detection under some scenarios. Although the impossibility theorems are frustrating, we find that some conditions of these impossibility theorems may not hold in some practical scenarios. Based on this observation, we next give several necessary and sufficient conditions to characterize the learnability of OOD detection in some practical scenarios. Lastly, we offer theoretical support for representative OOD detection works based on our OOD theory.
AB - Supervised learning aims to train a classifier under the assumption that training and test data are from the same distribution. To ease the above assumption, researchers have studied a more realistic setting: out-of-distribution (OOD) detection, where test data may come from classes that are unknown during training (i.e., OOD data). Due to the unavailability and diversity of OOD data, good generalization ability is crucial for effective OOD detection algorithms, and corresponding learning theory is still an open problem. To study the generalization of OOD detection, this paper investigates the probably approximately correct (PAC) learning theory of OOD detection that fits the commonly used evaluation metrics in the literature. First, we find a necessary condition for the learnability of OOD detection. Then, using this condition, we prove several impossibility theorems for the learnability of OOD detection under some scenarios. Although the impossibility theorems are frustrating, we find that some conditions of these impossibility theorems may not hold in some practical scenarios. Based on this observation, we next give several necessary and sufficient conditions to characterize the learnability of OOD detection in some practical scenarios. Lastly, we offer theoretical support for representative OOD detection works based on our OOD theory.
KW - learnability
KW - out-of-distribution detection
KW - weakly supervised learning
UR - https://www.jmlr.org/papers/v25/23-1257.html
UR - http://www.scopus.com/inward/record.url?scp=85202279228&partnerID=8YFLogxK
M3 - Journal article
SN - 1532-4435
VL - 25
SP - 1
EP - 83
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
IS - 84
ER -