Contrastive learning has become an important tool in learning representations from unlabeled data mainly relying on the idea of minimizing distance between positive data …
W Ge, Y Yu - Proceedings of the IEEE conference on …, 2017 - openaccess.thecvf.com
Deep neural networks require a large amount of labeled training data during supervised learning. However, collecting and labeling so much data might be infeasible in many cases …
EH Sanchez, M Serrurier, M Ortner - … , Glasgow, UK, August 23–28, 2020 …, 2020 - Springer
In this paper, we investigate the problem of learning disentangled representations. Given a pair of images sharing some attributes, we aim to create a low-dimensional representation …
Self-supervised learning algorithms based on instance discrimination train encoders to be invariant to pre-defined transformations of the same instance. While most methods treat …
We propose an algorithm, guided variational autoencoder (Guided-VAE), that is able to learn a controllable generative model by performing latent representation disentanglement …
B Neyshabur, H Sedghi… - Advances in neural …, 2020 - proceedings.neurips.cc
One desired capability for machines is the ability to transfer their understanding of one domain to another domain where data is (usually) scarce. Despite ample adaptation of …
N Hadad, L Wolf, M Shahar - Proceedings of the IEEE …, 2018 - openaccess.thecvf.com
We address the problem of disentanglement of factors that generate a given data into those that are correlated with the labeling and those that are not. Our solution is simpler than …
We propose a general framework called Network Dissection for quantifying the interpretability of latent representations of CNNs by evaluating the alignment between …
T Xiao, J Hong, J Ma - arXiv preprint arXiv:1711.05415, 2017 - arxiv.org
Disentangling factors of variation has become a very challenging problem on representation learning. Existing algorithms suffer from many limitations, such as unpredictable …