📑Paper Link
Authors: Haoran Chen, Micah Goldblum, Zuxuan Wu, Yu-Gang Jiang
- torch 2.0.1
- torchvision 0.15.2
- timm 0.6.12 (Note, for reproducing ARC + DER, we recommend using timm 0.5.4)
- tqdm
- numpy
- scipy
-
Edit the
[MODEL NAME].json
file for global settings and hyperparameters. -
Run:
python main.py --config=./exps/[MODEL NAME].json
- We primarily modified the _eval_cnn function in the files within the models directory compared to the PILOT version.
- Compared to the original conference version, we have re-implemented the framework. In the previous version, the inference process was fixed to a batch size of 1. In the new implementation, we introduce the arc_batch_size hyperparameter, which allows the inference batch size to be adjusted. Empirically, we observe that increasing this parameter tends to reduce overall accuracy while improving inference speed.
- The current version achieves substantially faster inference while maintaining comparable, and in some cases better, performance compared to the original version.
We thank the PILOT repo for providing helpful codebase in our work.
Feel free to contact us if you have any questions or suggestions Email: [email protected]
If you use our code in this repo or find our work helpful, please consider giving a citation:
@inproceedings{chenarc,
title={Adaptive Retention \& Correction: Test-Time Training for Continual Learning},
author={Chen, Haoran and Goldblum, Micah and Wu, Zuxuan and Jiang, Yu-Gang},
booktitle={ICLR 2025},
year={2025}
}