Y Mei, Y Fan, Y Zhou - … of the IEEE/CVF conference on …, 2021 - openaccess.thecvf.com
Both non-local (NL) operation and sparse representation are crucial for Single Image Super- Resolution (SISR). In this paper, we investigate their combinations and propose a novel Non …
Large Transformer models routinely achieve state-of-the-art results on a number of tasks but training these models can be prohibitively costly, especially on long sequences. We …
C Fung, CJM Yoon, I Beschastnikh - 23rd International Symposium on …, 2020 - usenix.org
Federated learning over distributed multi-party data is an emerging paradigm that iteratively aggregates updates from a group of devices to train a globally shared model. Relying on a …
Transformer models have advanced the state of the art in many Natural Language Processing (NLP) tasks. In this paper, we present a new Transformer architecture, Extended …
Quantization based techniques are the current state-of-the-art for scaling maximum inner product search to massive databases. Traditional approaches to quantization aim to …
C Fung, CJM Yoon, I Beschastnikh - arXiv preprint arXiv:1808.04866, 2018 - arxiv.org
Machine learning (ML) over distributed multi-party data is required for a variety of domains. Existing approaches, such as federated learning, collect the outputs computed by a group of …
Deep neural networks (DNNs) enable innovative applications of machine learning like image recognition, machine translation, or malware detection. However, deep learning is …
B Chen, T Dao, E Winsor, Z Song… - Advances in Neural …, 2021 - proceedings.neurips.cc
Recent advances in efficient Transformers have exploited either the sparsity or low-rank properties of attention matrices to reduce the computational and memory bottlenecks of …
JJ Pan, J Wang, G Li - The VLDB Journal, 2024 - Springer
There are now over 20 commercial vector database management systems (VDBMSs), all produced within the past five years. But embedding-based retrieval has been studied for …