A comprehensive survey on test-time adaptation under distribution shifts

J Liang, R He, T Tan - International Journal of Computer Vision, 2024 - Springer
Abstract Machine learning methods strive to acquire a robust model during the training
process that can effectively generalize to test samples, even in the presence of distribution …

Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks

L Wang, KJ Yoon - IEEE transactions on pattern analysis and …, 2021 - ieeexplore.ieee.org
Deep neural models, in recent years, have been successful in almost every field, even
solving the most complex problem statements. However, these models are huge in size with …

Egoschema: A diagnostic benchmark for very long-form video language understanding

K Mangalam, R Akshulakov… - Advances in Neural …, 2023 - proceedings.neurips.cc
We introduce EgoSchema, a very long-form video question-answering dataset, and
benchmark to evaluate long video understanding capabilities of modern vision and …

Test-time training with masked autoencoders

Y Gandelsman, Y Sun, X Chen… - Advances in Neural …, 2022 - proceedings.neurips.cc
Test-time training adapts to a new test distribution on the fly by optimizing a model for each
test input using self-supervision. In this paper, we use masked autoencoders for this one …

Knowledge distillation: A survey

J Gou, B Yu, SJ Maybank, D Tao - International Journal of Computer Vision, 2021 - Springer
In recent years, deep neural networks have been successful in both industry and academia,
especially for computer vision tasks. The great success of deep learning is mainly due to its …

Test-time training with self-supervision for generalization under distribution shifts

Y Sun, X Wang, Z Liu, J Miller… - … on machine learning, 2020 - proceedings.mlr.press
In this paper, we propose Test-Time Training, a general approach for improving the
performance of predictive models when training and test data come from different …

Learning to (learn at test time): Rnns with expressive hidden states

Y Sun, X Li, K Dalal, J Xu, A Vikram, G Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
Self-attention performs well in long context but has quadratic complexity. Existing RNN
layers have linear complexity, but their performance in long context is limited by the …

Space-time correspondence as a contrastive random walk

A Jabri, A Owens, A Efros - Advances in neural information …, 2020 - proceedings.neurips.cc
This paper proposes a simple self-supervised approach for learning a representation for
visual correspondence from raw video. We cast correspondence as prediction of links in a …

Aligning distillation for cold-start item recommendation

F Huang, Z Wang, X Huang, Y Qian, Z Li… - Proceedings of the 46th …, 2023 - dl.acm.org
Recommending cold items in recommendation systems is a longstanding challenge due to
the inherent differences between warm items, which are recommended based on user …

Liga-stereo: Learning lidar geometry aware representations for stereo-based 3d detector

X Guo, S Shi, X Wang, H Li - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Stereo-based 3D detection aims at detecting 3D object bounding boxes from stereo images
using intermediate depth maps or implicit 3D geometry representations, which provides a …