description |
---|
By Carnegie Mellon University and CMU Argo AI Center |
CLEAR is a novel continual/lifelong benchmark that captures real-world distribution shifts in Internet image collection (YFCC100M) from 2004 to 2014.
For long, researchers in continual learning (CL) community have been working with artificial CL benchmarks such as "Permuted-MNIST" and "Split-CIFAR", which do not align with practical applications. In reality, distribution shifts are smooth, such as natural temporal evolution of visual concepts.
Below are examples of classes in CLEAR-100 that changed over the past decade:
The CLEAR Benchmark and the CLEAR-10 dataset are first introduced in our NeurIPS 2021 paper.
{% embed url="https://arxiv.org/abs/2201.06289" %} NeurIPS'21 Datasets and Benchmarks Track {% endembed %}
In spirit of the famous CIFAR-10/CIFAR-100 benchmarks for static image classification tasks, we also collected a more challenging CLEAR-100 with a diverse set of 100 classes.
{% hint style="info" %} We hope our CLEAR-10/-100 benchmarks can be the new "CIFAR" as a test stone for continual/lifelong learning community. {% endhint %}
We are also extending CLEAR to an ImageNet-scale benchmark. If you have feedback and insights, feel free to reach out to us!
{% content-ref url="introduction/about-us.md" %} about-us.md {% endcontent-ref %}
In June 2022, the 1st CLEAR Challenge was hosted on CVPR 2022 Open World Vision Workshop, with a total of 15 teams from 21 different countries and regions partcipating. You may find a quick summary of the workshop in the below page:
{% content-ref url="introduction/1st-clear-challenge-cvpr22.md" %} 1st-clear-challenge-cvpr22.md {% endcontent-ref %}
Given the top teams' promising performance on CLEAR-10/-100 benchmarks via utilizing methods that improve generalization, such as sharpness aware minimization, supervised contrastive loss, strong data augmentation, experience replay, etc., we believe there are still a wealth of problems in CLEAR for the community to explore, such as:
- Improving Forward Transfer and Next-Domain Accuracy
- Unsupervised/Online Domain Generalization
- Self-supervised/Semi-Supervised Continous Learning
In the following pages, we will explain the motivation of CLEAR benchmark, how it is curated via visio-linguistic approach, its evaluation protocols, and a walk-through of the 1st CLEAR Challenge on CVPR'22.
You can also jump to the links for downloading CLEAR dataset:
{% content-ref url="documentation/download-clear-10-clear-100.md" %} download-clear-10-clear-100.md {% endcontent-ref %}