Field-Guide-Inspired Zero-Shot Learning

Utkarsh MallBharath HariharanKavita Bala

Cornell University

In ICCV 2021




Abstract

Modern recognition systems require large amounts of supervision to achieve accuracy. Adapting to new domains requires significant data from experts, which is onerous and can become too expensive. Zero-shot learning requires an annotated set of attributes for a novel category. Annotating the full set of attributes for a novel category proves to be a tedious and expensive task in deployment. This is especially the case when the recognition domain is an expert domain. We introduce a new field-guide-inspired approach to zero-shot annotation where the learner model interactively asks for the most useful attributes that define a class. We evaluate our method on classification benchmarks with attribute annotations like CUB, SUN, and AWA2 and show that our model achieves the performance of a model with full annotations at the cost of a significantly fewer number of annotations. Since the time of experts is precious, decreasing annotation cost can be very valuable for real-world deployment.

Overview of our field-guide-inspired workflow. (a) An expert annotator introduces the learner to a novel class by first providing a similar base class. (b) The Learner then interactively asks for values of different attributes using an Acquistion function. (c) The expert annotator provides the values of the attributes and the learner updates its state and class description

Paper

[pdf]   [arXiv]   [supplementary pdf]

Utkarsh Mall, Bharath Hariharan, and Kavita Bala. "Field-Guide-Inspired Zero-Shot Learning". In ICCV, 2021.

@inproceedings{fgzsl-21,
 title={Field-Guide-Inspired Zero-Shot Learning},
 author={Mall, Utkarsh and Hariharan, Bharath and Bala, Kavita},
 booktitle={ICCV},
 year={2021}
}

Poster

Code

The code can be found here

Different runs of pre-trained models for inference can be downloaded here.

[zip] models.zip [3.9G]: Contains pre-trained models in the format {Dataset}_{Run}.pth.tar. Model for {Dataset} dataset (CUB, SUN or AWA2) and one of the random runs.

Data

[txt] Readme.txt

[zip] data.zip: Contains relevant data from (CUB, SUN and AWA2) in uniform pickle (python3) format.

Videos

5-min talk about the paper.

Video showing t-SNE visualization of progression of learning using t-SNE.

Acknowledgements

This work was supported by the TCS, NSF (1900783) and DARPA LwLL Program (HR001118S0044).