UM
Residential Collegefalse
Status已發表Published
LCA-on-the-Line: Benchmarking Out-of-Distribution Generalization with Class Taxonomies
Shi, Jia1; Gare, Gautam1; Tian, Jinjin1; Chai, Siqi1; Lin, Zhiqiu1; Vasudevan, Arun1; Feng, Di2,3; Ferroni, Francesco2,4; Kong, Shu5,6
2024
Conference Name41st International Conference on Machine Learning, ICML 2024
Source PublicationProceedings of Machine Learning Research
Volume235
Pages44887-44908
Conference Date21 July 2024through 27 July 2024
Conference PlaceVienna
PublisherML Research Press
Abstract

We tackle the challenge of predicting models' Out-of-Distribution (OOD) performance using in-distribution (ID) measurements without requiring OOD data. Existing evaluations with “Effective robustness”, which use ID accuracy as an indicator of OOD accuracy, encounter limitations when models are trained with diverse supervision and distributions, such as class labels (Vision Models, VMs, on ImageNet) and textual descriptions (Visual-Language Models, VLMs, on LAION). VLMs often generalize better to OOD data than VMs despite having similar or lower ID performance. To improve the prediction of models' OOD performance from ID measurements, we introduce the Lowest Common Ancestor (LCA)on-the-Line framework. This approach revisits the established concept of LCA distance, which measures the hierarchical distance between labels and predictions within a predefined class hierarchy, such as WordNet. We assess 75 models using ImageNet as the ID dataset and five significantly shifted OOD variants, uncovering a strong linear correlation between ID LCA distance and OOD top-1 accuracy. Our method provides a compelling alternative for understanding why VLMs tend to generalize better. Additionally, we propose a technique to construct a taxonomic hierarchy on any dataset using K-means clustering, demonstrating that LCA distance is robust to the constructed taxonomic hierarchy. Moreover, we demonstrate that aligning model predictions with class taxonomies, through soft labels or prompt engineering, can enhance model generalization. Open source code in our Project Page.

URLView the original
Language英語English
Scopus ID2-s2.0-85203843463
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionUniversity of Macau
Affiliation1.Carnegie Mellon University, United States
2.Argo AI GmbH, United States
3.Apple, United States
4.Nvidia, United States
5.Texas A&M University, United States
6.University of Macau, Macao
Recommended Citation
GB/T 7714
Shi, Jia,Gare, Gautam,Tian, Jinjin,et al. LCA-on-the-Line: Benchmarking Out-of-Distribution Generalization with Class Taxonomies[C]:ML Research Press, 2024, 44887-44908.
APA Shi, Jia., Gare, Gautam., Tian, Jinjin., Chai, Siqi., Lin, Zhiqiu., Vasudevan, Arun., Feng, Di., Ferroni, Francesco., & Kong, Shu (2024). LCA-on-the-Line: Benchmarking Out-of-Distribution Generalization with Class Taxonomies. Proceedings of Machine Learning Research, 235, 44887-44908.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Shi, Jia]'s Articles
[Gare, Gautam]'s Articles
[Tian, Jinjin]'s Articles
Baidu academic
Similar articles in Baidu academic
[Shi, Jia]'s Articles
[Gare, Gautam]'s Articles
[Tian, Jinjin]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Shi, Jia]'s Articles
[Gare, Gautam]'s Articles
[Tian, Jinjin]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.