Lora: Low-rank adaptation of large language models

EJ Hu, Y Shen, P Wallis, Z Allen-Zhu, Y Li… - arXiv preprint arXiv …, 2021 - arxiv.org
An important paradigm of natural language processing consists of large-scale pre-training
on general domain data and adaptation to particular tasks or domains. As we pre-train larger …

Matrix completion has no spurious local minimum

R Ge, JD Lee, T Ma - Advances in neural information …, 2016 - proceedings.neurips.cc
Matrix completion is a basic machine learning problem that has wide applications,
especially in collaborative filtering and recommender systems. Simple non-convex …

No spurious local minima in nonconvex low rank problems: A unified geometric analysis

R Ge, C Jin, Y Zheng - International Conference on Machine …, 2017 - proceedings.mlr.press
In this paper we develop a new framework that captures the common landscape underlying
the common non-convex low-rank matrix problems including matrix sensing, matrix …

Mr. biq: Post-training non-uniform quantization based on minimizing the reconstruction error

Y Jeon, C Lee, E Cho, Y Ro - Proceedings of the IEEE/CVF …, 2022 - openaccess.thecvf.com
Post-training quantization compresses a neural network within few hours with only a small
unlabeled calibration set. However, so far it has been only discussed and empirically …

An efficient dataset condensation plugin and its application to continual learning

E Yang, L Shen, Z Wang, T Liu… - Advances in Neural …, 2023 - proceedings.neurips.cc
Dataset condensation (DC) distills a large real-world dataset into a small synthetic dataset,
with the goal of training a network from scratch on the latter that performs similarly to the …

Lq-lora: Low-rank plus quantized matrix decomposition for efficient language model finetuning

H Guo, P Greengard, EP Xing, Y Kim - arXiv preprint arXiv:2311.12023, 2023 - arxiv.org
We propose a simple approach for memory-efficient adaptation of pretrained language
models. Our approach uses an iterative algorithm to decompose each pretrained matrix into …

Complete dictionary recovery over the sphere II: Recovery by Riemannian trust-region method

J Sun, Q Qu, J Wright - IEEE Transactions on Information …, 2016 - ieeexplore.ieee.org
We consider the problem of recovering a complete (ie, square and invertible) matrix A 0,
from Y∈ R n× p with Y= A 0 X 0, provided X 0 is sufficiently sparse. This recovery problem is …

Weighted low rank approximations with provable guarantees

I Razenshteyn, Z Song, DP Woodruff - … of the forty-eighth annual ACM …, 2016 - dl.acm.org
The classical low rank approximation problem is: given a matrix A, find a rank-k matrix B
such that the Frobenius norm of A− B is minimized. It can be solved efficiently using, for …

Low rank matrix completion via robust alternating minimization in nearly linear time

Y Gu, Z Song, J Yin, L Zhang - arXiv preprint arXiv:2302.11068, 2023 - arxiv.org
Given a matrix $ M\in\mathbb {R}^{m\times n} $, the low rank matrix completion problem
asks us to find a rank-$ k $ approximation of $ M $ as $ UV^\top $ for $ U\in\mathbb …

Nonconvex Rectangular Matrix Completion via Gradient Descent Without ₂, Regularization

J Chen, D Liu, X Li - IEEE Transactions on Information Theory, 2020 - ieeexplore.ieee.org
The analysis of nonconvex matrix completion has recently attracted much attention in the
community of machine learning thanks to its computational convenience. Existing analysis …