Home

Awesome

<h1 align="center">EOE: Envisioning Outlier Exposure</h1> <p align="center"> <a href="https://arxiv.org/pdf/2406.00806"><img src="https://img.shields.io/badge/arXiv-2406.00806-b31b1b.svg" alt="Paper"></a> <a href="https://openreview.net/pdf?id=xZO7SmM12y"> <img alt="License" src="https://img.shields.io/static/v1?label=Pub&message=ICML%2724&color=blue"> </a> </p>

Official code for the paper "Envisioning Outlier Exposure by Large Language Models for Out-of-Distribution Detection" (ICML 2024).

Illustration

Abstract

Detecting out-of-distribution (OOD) samples is essential when deploying machine learning models in open-world scenarios. Zero-shot OOD detection, requiring no training on in-distribution (ID) data, has been possible with the advent of vision-language models like CLIP. Existing methods build a text-based classifier with only closed-set labels. However, this largely restricts the inherent capability of CLIP to recognize samples from large and open label space. In this paper, we propose to tackle this constraint by leveraging the expert knowledge and reasoning capability of large language models (LLM) to Envision potential Outlier Exposure, termed EOE, without access to any actual OOD data. Owing to better adaptation to open-world scenarios, EOE can be generalized to different tasks, including far, near, and fine-grained OOD detection. Technically, we design (1) LLM prompts based on visual similarity to generate potential outlier class labels specialized for OOD detection, as well as (2) a new score function based on potential outlier penalty to distinguish hard OOD samples effectively. Empirically, EOE achieves state-of-the-art performance across different OOD tasks and can be effectively scaled to the ImageNet-1K dataset.

Setup

Dependencies

pip install -r requirements.txt
# what's more, please add your openai/poe key in .env
touch .env
echo "OPENAI_API_KEY=your_openai_api_key_here" >> .env
echo "POE_API_KEY=your_poe_api_key_here" >> .env

Dataset Preparation

In-distribution Datasets

We consider the following ID datasets: CUB-200, Standford-Cars, Food-101, Oxford-Pet, ImageNet-1k, ImageNet-10, ImageNet-20, CUB-100, Stanford-Cars-98, Food-50, Oxford-Pet-18, CIFAR-10, CIFAR-100, ImageNet-C, and ImageNet-Sketch

Specifically,

  1. Please refer to MCM for the preparation of the following datasets:
  1. CUB-100, Stanford-Cars-98, Food-50, Oxford-Pet-18 are randomly extracted from - CUB-200, Standford-Cars, Food-101, Oxford-Pet. We provide the index of the randomly extracted categories used in the paper in CUB-100, Stanford-Cars-98, Food-50, and Oxford-Pet-18
  2. The ImageNet-C and ImageNet-Sketch datasets can be downloaded from ImageNet-C and ImageNet-Sketch, respectively. CIFAR-10/100 will be automatically downloaded when you run the code,

Please put the above datasets into ./datasets.

Out-of-Distribution Datasets

We consider the following OOD datasets: iNaturalist, SUN, Places, Texture, ImageNet10, ImageNet20, CUB-100(OOD), Stanford-Cars-98(OOD), Food-51(OOD), and Oxford-Pet-19(OOD), SSB-hard, NINCO

Specifically,

  1. Please refer to Huang et al. 2021 for the preparation of the following datasets: iNaturalist, SUN, Places, Texture
  2. CUB-100(OOD), Stanford-Cars-98(OOD), Food-51(OOD), and Oxford-Pet-19(OOD) are the remaining categories in the datasets(CUB-200, Standford-Cars, Food-101, Oxford-Pet) and are automatically obtained after loading the ID datasets(CUB-100, Stanford-Cars-98, Food-50, and Oxford-Pet-18).
  3. Please refer to OpenOOD for the preparation of the following datasets: SSB-hard, NINCO

Please put the following datasets (iNaturalist, SUN, Places, Texture, SSB-hard, NINCO) into ./datasets/ImageNet_OOD_dataset.

Quick Start

The main script for evaluating OOD detection performance is eval_ood_detection.py. Here are the list of main arguments:

You can directly use the following script

bash eval.sh

Citation

If you find our work useful, please consider citing our paper:

@inproceedings{cao2024envisioning,
  title={Envisioning Outlier Exposure by Large Language Models for Out-of-Distribution Detection},
  author={Cao, Chentao and Zhong, Zhun and Zhou, Zhanke and Liu, Yang and Liu, Tongliang and Han, Bo},
  booktitle={ICML},
  year={2024}
}

Our implementation is based on MCM by Dr. Ming. Thanks for their great work!

Questions

If you have any questions, please feel free to contact chentaocao1224@gmail.com