Computational optimal transport: With applications to data science

G Peyré, M Cuturi - Foundations and Trends® in Machine …, 2019 - nowpublishers.com
Optimal transport (OT) theory can be informally described using the words of the French
mathematician Gaspard Monge (1746–1818): A worker with a shovel in hand has to move a …

A survey of recent advances in optimization methods for wireless communications

YF Liu, TH Chang, M Hong, Z Wu… - IEEE Journal on …, 2024 - ieeexplore.ieee.org
Mathematical optimization is now widely regarded as an indispensable modeling and
solution tool for the design of wireless communications systems. While optimization has …

Efficient and modular implicit differentiation

M Blondel, Q Berthet, M Cuturi… - Advances in neural …, 2022 - proceedings.neurips.cc
Automatic differentiation (autodiff) has revolutionized machine learning. Itallows to express
complex computations by composing elementary ones in creativeways and removes the …

Zero-shot hyperspectral sharpening

R Dian, A Guo, S Li - IEEE Transactions on Pattern Analysis …, 2023 - ieeexplore.ieee.org
Fusing hyperspectral images (HSIs) with multispectral images (MSIs) of higher spatial
resolution has become an effective way to sharpen HSIs. Recently, deep convolutional …

Multiview consensus graph clustering

K Zhan, F Nie, J Wang, Y Yang - IEEE Transactions on Image …, 2018 - ieeexplore.ieee.org
A graph is usually formed to reveal the relationship between data points and graph structure
is encoded by the affinity matrix. Most graph-based multiview clustering methods use …

Parseval networks: Improving robustness to adversarial examples

M Cisse, P Bojanowski, E Grave… - International …, 2017 - proceedings.mlr.press
We introduce Parseval networks, a form of deep neural networks in which the Lipschitz
constant of linear, convolutional and aggregation layers is constrained to be smaller than $1 …

Sorting out Lipschitz function approximation

C Anil, J Lucas, R Grosse - International Conference on …, 2019 - proceedings.mlr.press
Training neural networks under a strict Lipschitz constraint is useful for provable adversarial
robustness, generalization bounds, interpretable gradients, and Wasserstein distance …

From softmax to sparsemax: A sparse model of attention and multi-label classification

A Martins, R Astudillo - International conference on machine …, 2016 - proceedings.mlr.press
We propose sparsemax, a new activation function similar to the traditional softmax, but able
to output sparse probabilities. After deriving its properties, we show how its Jacobian can be …

Sparse sequence-to-sequence models

B Peters, V Niculae, AFT Martins - arXiv preprint arXiv:1905.05702, 2019 - arxiv.org
Sequence-to-sequence models are a powerful workhorse of NLP. Most variants employ a
softmax transformation in both their attention mechanism and output layer, leading to dense …

Hyperspectral super-resolution by coupled spectral unmixing

C Lanaras, E Baltsavias… - Proceedings of the IEEE …, 2015 - openaccess.thecvf.com
Hyperspectral cameras capture images with many narrow spectral channels, which densely
sample the electromagnetic spectrum. The detailed spectral resolution is useful for many …