NEO: No-Optimization Test-Time Adaptation through Latent Re-Centering

  • Alexander Murphy*
  • , Michal Danilowski
  • , Soumyajit Chatterjee
  • , Abhirup Ghosh
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Test-Time Adaptation (TTA) methods are often computationally expensive, require a large amount of data for effective adaptation, or are brittle to hyperparameters. Based on a theoretical foundation of the geometry of the latent space, we are able to significantly improve the alignment between source and distribution-shifted samples by re-centering target data embeddings at the origin. This insight motivates NEO -- a hyperparameter-free fully TTA method, that adds no significant compute compared to vanilla inference. NEO is able to improve the classification accuracy of ViT-Base on ImageNet-C from 55.6% to 59.2% after adapting on just one batch of 64 samples. When adapting on 512 samples NEO beats all 7 TTA methods we compare against on ImageNet-C, ImageNet-R and ImageNet-S and beats 6/7 on CIFAR-10-C, while using the least amount of compute. NEO performs well on model calibration metrics and additionally is able to adapt from 1 class to improve accuracy on 999 other classes in ImageNet-C. On Raspberry Pi and Jetson Orin Nano devices, NEO reduces inference time by 63% and memory usage by 9% compared to baselines. Our results based on 3 ViT architectures and 4 datasets show that NEO can be used efficiently and effectively for TTA.
Original languageEnglish
Title of host publicationThe Fourteenth International Conference on Learning Representations (ICLR 2026)
PublisherInternational Conference on Learning Representations, ICLR
Publication statusAccepted/In press - 26 Jan 2026
EventFourteenth International Conference on Learning Representations - Riocentro Convention and Event Center, Rio de Janeiro, Brazil
Duration: 23 Apr 202627 Apr 2026
https://iclr.cc/Conferences/2026

Publication series

NameICLR Proceedings
PublisherInternational Conference on Learning Representations (ICLR)

Conference

ConferenceFourteenth International Conference on Learning Representations
Abbreviated titleICLR 2026
Country/TerritoryBrazil
CityRio de Janeiro
Period23/04/2627/04/26
Internet address

Bibliographical note

Not yet published as of 05/03/2026.

Fingerprint

Dive into the research topics of 'NEO: No-Optimization Test-Time Adaptation through Latent Re-Centering'. Together they form a unique fingerprint.

Cite this