This repository contains the revised implementation of the paper: On Multi-Domain Long-Tailed Recognition, Imbalanced Domain Generalization and Beyond (ECCV 2022).
It is also a (living) PyTorch suite containing benchmark datasets and algorithms for Multi-Domain Long-Tailed Recognition (MDLT). Currently we support 8 MDLT datasets (3 synthetic + 5 real), as well as ~20 algorithms that span different learning strategies. Feel free to send us a PR to add your algorithm / dataset for MDLT!
Multi-Domain Long-Tailed Recognition (MDLT) aims to learn from multi-domain imbalanced data, address label imbalance, domain shift, and divergent label distributions across domains, and generalize to all domain-class pairs.
Existing studies on data imbalance focus on single-domain settings, i.e., samples are from the same data distribution. However, natural data can originate from distinct domains, where a minority class in one domain could have abundant instances from other domains. We systematically investigate Multi-Domain Long-Tailed Recognition (MDLT), which learns from multi-domain imbalanced data, addresses label imbalance, domain shift, and divergent label distributions across domains, and generalizes to all domain-class pairs.
We develop the domain-class transferability graph, and show that such transferability governs the success of learning in MDLT. We then propose BoDA, a theoretically grounded learning strategy that tracks the upper bound of transferability statistics, and ensures balanced alignment and calibration across imbalanced domain-class distributions. We curate MDLT benchmark datasets based on widely-used multi-domain datasets, and benchmark ~20 algorithms that span different learning strategies for MDLT.
Further, as a byproduct, we demonstrate that BoDA strengthens Domain Generalization (DG) algorithms, and consistently improves the results on DG benchmarks. Note that all current standard DG benchmarks naturally exhibit heavy class imbalance within domains and label distributions shift across domains, confirming that data imbalance is an intrinsic problem in DG, but has yet been overlooked by past works.
The results shed light on how label imbalance can affect out-of-distribution generalization, and highlight the importance of integrating label imbalance into practical DG algorithm design.
- Download the original datasets, and place them in your
data_path
python -m mdlt.scripts.download --data_dir <data_path>
- Place the
.csv
files of train/val/test splits for each MDLT dataset (provided inmdlt/dataset/split/
) in the corresponding dataset folder under yourdata_path
Step1: Install CUDA Toolkit 11.8 from here
Step2: Run the following commands to install the dependencies
conda create -n mdlt python=3.8
pip install -r requirements.txt
conda env create -f environment.yaml
train.py
: main training scriptsweep.py
: launch a sweep with all selected algorithms (provided inmdlt/learning/algorithms.py
) on all real MDLT datasets (VLCS-MLT, PACS-MLT, OfficeHome-MLT, TerraInc-MLT, DomainNet-MLT)sweep_synthetic.py
: launch a sweep with all selected algorithms on the synthetic MDLT dataset (Digits-MLT)collect_results.py
: collect sweep results to automatically generate result tables (as in the paper)eval_best_hparam.py
&eval_checkpoint.py
: scripts for evaluating trained models
- train.py:
--dataset
: name of chosen MDLT dataset--algorithm
: choose algorithm used for running--data_dir
: data path--output_dir
: output path--output_folder_name
: output folder name (underoutput_dir
) for the current run--hparams_seed
: seed for different hyper-parameters--seed
: seed for different runs--selected_envs
: train on selected envs (only used for Digits-MLT)--imb_type
&--imb_factor
: arguments for customized Digits-MLT label distributions--stage1_folder
&--stage1_algo
: arguments for two-stage algorithms
- sweep.py:
--n_hparams
: how many hparams to run for each <dataset, algorithm> pair--best_hp
&--n_trials
: after sweeping hparams, fix best hparam and run trials with different seeds
python -m mdlt.train --algorithm <algo> --dataset <dset> --output_folder_name <output_folder_name> --data_dir <data_path> --output_dir <output_path>
python -m mdlt.train --algorithm CRT --dataset <dset> --output_folder_name <output_folder_name> --data_dir <data_path> --output_dir <output_path> --stage1_folder <stage1_model_folder> --stage1_algo <stage1_algo>
Note that for stage1_algo
& stage1_folder
python -m mdlt.train --algorithm <algo> --dataset ImbalancedDigits \
--imb_type eee \
--imb_factor 0.01 \
--selected_envs 1 2
Note that for Digits-MLT, we additionally provide MNIST
as another domain.
To maintain the same setting as in paper (2 domains), you only need to set selected_envs
to be 1 2
as above
python -m mdlt.sweep launch --algorithms <...> --dataset <...> --n_hparams <num_of_hparams> --n_trials 1
python -m mdlt.sweep launch --algorithms <...> --dataset <...> --best_hp --input_folder <...> --n_trials <num_of_trials>
python -m mdlt.scripts.collect_results --input_dir <...>
python -u -m mdlt.evaluate.eval_best_hparam --algorithm <...> --dataset <...> --data_dir <...> --output_dir <...> --folder_name <...>
python -u -m mdlt.evaluate.eval_checkpoint --algorithm <...> --dataset <...> --data_dir <...> --checkpoint <...>
Model | VLCS-MLT | PACS-MLT | OfficeHome-MLT | TerraInc-MLT | DomainNet-MLT |
---|---|---|---|---|---|
BoDA (r) | 76.9 / model | 97.0 / model | 81.5 / model | 78.6 / model | 60.1 / model |
BoDA (r,c) | 77.3 / model | 97.2 / model | 82.3 / model | 82.3 / model | 61.7 / model |
- [10/2022] Check out the Oral talk video (10 mins) for our ECCV paper.
- [07/2022] We create a Blog post for this work (version in Chinese is also available here). Check it out for more details!
- [07/2022] Paper accepted to ECCV 2022. We have released the code and models.
- [03/2022] arXiv version posted. The code is currently under cleaning. Please stay tuned for updates.
This code is partly based on the open-source implementations from DomainBed.
If you find this code or idea useful, please cite our work:
@inproceedings{yang2022multi,
title={On Multi-Domain Long-Tailed Recognition, Imbalanced Domain Generalization and Beyond},
author={Yang, Yuzhe and Wang, Hao and Katabi, Dina},
booktitle={European Conference on Computer Vision (ECCV)},
year={2022}
}
If you have any questions, feel free to contact us through email ([email protected]) or Github issues. Enjoy!