Residential College | false |
Status | 已發表Published |
LCA-on-the-Line: Benchmarking Out-of-Distribution Generalization with Class Taxonomies | |
Shi, Jia1; Gare, Gautam1; Tian, Jinjin1; Chai, Siqi1; Lin, Zhiqiu1; Vasudevan, Arun1; Feng, Di2,3; Ferroni, Francesco2,4; Kong, Shu5,6 | |
2024 | |
Conference Name | 41st International Conference on Machine Learning, ICML 2024 |
Source Publication | Proceedings of Machine Learning Research |
Volume | 235 |
Pages | 44887-44908 |
Conference Date | 21 July 2024through 27 July 2024 |
Conference Place | Vienna |
Publisher | ML Research Press |
Abstract | We tackle the challenge of predicting models' Out-of-Distribution (OOD) performance using in-distribution (ID) measurements without requiring OOD data. Existing evaluations with “Effective robustness”, which use ID accuracy as an indicator of OOD accuracy, encounter limitations when models are trained with diverse supervision and distributions, such as class labels (Vision Models, VMs, on ImageNet) and textual descriptions (Visual-Language Models, VLMs, on LAION). VLMs often generalize better to OOD data than VMs despite having similar or lower ID performance. To improve the prediction of models' OOD performance from ID measurements, we introduce the Lowest Common Ancestor (LCA)on-the-Line framework. This approach revisits the established concept of LCA distance, which measures the hierarchical distance between labels and predictions within a predefined class hierarchy, such as WordNet. We assess 75 models using ImageNet as the ID dataset and five significantly shifted OOD variants, uncovering a strong linear correlation between ID LCA distance and OOD top-1 accuracy. Our method provides a compelling alternative for understanding why VLMs tend to generalize better. Additionally, we propose a technique to construct a taxonomic hierarchy on any dataset using K-means clustering, demonstrating that LCA distance is robust to the constructed taxonomic hierarchy. Moreover, we demonstrate that aligning model predictions with class taxonomies, through soft labels or prompt engineering, can enhance model generalization. Open source code in our Project Page. |
URL | View the original |
Language | 英語English |
Scopus ID | 2-s2.0-85203843463 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | University of Macau |
Affiliation | 1.Carnegie Mellon University, United States 2.Argo AI GmbH, United States 3.Apple, United States 4.Nvidia, United States 5.Texas A&M University, United States 6.University of Macau, Macao |
Recommended Citation GB/T 7714 | Shi, Jia,Gare, Gautam,Tian, Jinjin,et al. LCA-on-the-Line: Benchmarking Out-of-Distribution Generalization with Class Taxonomies[C]:ML Research Press, 2024, 44887-44908. |
APA | Shi, Jia., Gare, Gautam., Tian, Jinjin., Chai, Siqi., Lin, Zhiqiu., Vasudevan, Arun., Feng, Di., Ferroni, Francesco., & Kong, Shu (2024). LCA-on-the-Line: Benchmarking Out-of-Distribution Generalization with Class Taxonomies. Proceedings of Machine Learning Research, 235, 44887-44908. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment