Train Till You Drop: Towards Stable and Robust Source-free Unsupervised 3D Domain Adaptation

Evolution of the performance of baselines without degradation prevention strategies as they train over 20k iterations. Our method (TTYDcore) uses an unsupervised criterion to stop training. The horizontal dotted line illustrates that we keep the model obtained at the stopping point (marked with a cross). Models are trained on nuScenes (NS) and unsupervisedly adapted to SemanticKITTI (SK10).
Evolution of the performance of baselines without degradation prevention strategies as they train over 20k iterations. Our method (TTYDcore) uses an unsupervised criterion to stop training. The horizontal dotted line illustrates that we keep the model obtained at the stopping point (marked with a cross). Models are trained on nuScenes (NS) and unsupervisedly adapted to SemanticKITTI (SK10).
We tackle the challenging problem of source-free unsupervised domain adaptation (SFUDA) for 3D semantic segmentation. It amounts to performing domain adaptation on an unlabeled target domain without any access to source data; the available information is a model trained to achieve good performance on the source domain. A common issue with existing SFUDA approaches is that performance degrades after some training time, which is a by-product of an under-constrained and ill-posed problem. We discuss two strategies to alleviate this issue. First, we propose a sensible way to regularize the learning problem. Second, we introduce a novel criterion based on agreement with a reference model. It is used (1) to stop the training when appropriate and (2) as validator to select hyperparameters without any knowledge on the target domain. Our contributions are easy to implement and readily amenable for all SFUDA methods, ensuring stable improvements over all baselines. We validate our findings on various 3D lidar settings, achieving state-of-the-art performance.
Examples of results with TTYDstop: ground truth (GT), initial model trained only on source data, training with our training scheme when using our stopping criterion, and “full” training for 20k iterations. Notable errors due to degradation are marked with a dashed rectangle.
Examples of results with TTYDstop: ground truth (GT), initial model trained only on source data, training with our training scheme when using our stopping criterion, and “full” training for 20k iterations. Notable errors due to degradation are marked with a dashed rectangle.

Published: ECCV, 2024

Björn Michele
Björn Michele
Ph.D. Student

Related