GeneralizeFormer: Layer-Adaptive Model Generation across Test-Time Distribution Shifts

Sameer Ambekar, Zehao Xiao, Xiantong Zhen, Cees Snoek; Proceedings of the Winter Conference on Applications of Computer Vision (WACV), 2025, pp. 6548-6558

Abstract


We consider the problem of test-time domain generalization where a model is trained on several source domains and adjusted on target domains never seen during training. Different from the common methods that fine-tune the model or adjust the classifier parameters online we propose to generate multiple layer parameters on the fly during inference by a lightweight meta-learned transformer which we call GeneralizeFormer. The layer-wise parameters are generated per target batch without fine-tuning or online adjustment. By doing so our method is more effective in dynamic scenarios with multiple target distributions and also avoids forgetting valuable source distribution characteristics. Moreover by considering layer-wise gradients the proposed method adapts itself to various distribution shifts. To reduce the computational and time cost we fix the convolutional parameters while only generating parameters of the Batch Normalization layers and the linear classifier. Experiments on six widely used domain generalization datasets demonstrate the benefits and abilities of the proposed method to efficiently handle various distribution shifts generalize in dynamic scenarios and avoid forgetting.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Ambekar_2025_WACV, author = {Ambekar, Sameer and Xiao, Zehao and Zhen, Xiantong and Snoek, Cees}, title = {GeneralizeFormer: Layer-Adaptive Model Generation across Test-Time Distribution Shifts}, booktitle = {Proceedings of the Winter Conference on Applications of Computer Vision (WACV)}, month = {February}, year = {2025}, pages = {6548-6558} }