Rwkv: Reinventing rnns for the transformer era

B Peng, E Alcaide, Q Anthony, A Albalak… - arXiv preprint arXiv …, 2023 - arxiv.org
Transformers have revolutionized almost all natural language processing (NLP) tasks but
suffer from memory and computational complexity that scales quadratically with sequence …

Planning with diffusion for flexible behavior synthesis

M Janner, Y Du, JB Tenenbaum, S Levine - arXiv preprint arXiv …, 2022 - arxiv.org
Model-based reinforcement learning methods often use learning only for the purpose of
estimating an approximate dynamics model, offloading the rest of the decision-making work …

scGPT: toward building a foundation model for single-cell multi-omics using generative AI

H Cui, C Wang, H Maan, K Pang, F Luo, N Duan… - Nature …, 2024 - nature.com
Generative pretrained models have achieved remarkable success in various domains such
as language and computer vision. Specifically, the combination of large-scale diverse …

Training a helpful and harmless assistant with reinforcement learning from human feedback

Y Bai, A Jones, K Ndousse, A Askell, A Chen… - arXiv preprint arXiv …, 2022 - arxiv.org
We apply preference modeling and reinforcement learning from human feedback (RLHF) to
finetune language models to act as helpful and harmless assistants. We find this alignment …

Vision gnn: An image is worth graph of nodes

K Han, Y Wang, J Guo, Y Tang… - Advances in neural …, 2022 - proceedings.neurips.cc
Network architecture plays a key role in the deep learning-based computer vision system.
The widely-used convolutional neural network and transformer treat the image as a grid or …

Non-stationary transformers: Exploring the stationarity in time series forecasting

Y Liu, H Wu, J Wang, M Long - Advances in Neural …, 2022 - proceedings.neurips.cc
Transformers have shown great power in time series forecasting due to their global-range
modeling ability. However, their performance can degenerate terribly on non-stationary real …

Tallrec: An effective and efficient tuning framework to align large language model with recommendation

K Bao, J Zhang, Y Zhang, W Wang, F Feng… - Proceedings of the 17th …, 2023 - dl.acm.org
Large Language Models (LLMs) have demonstrated remarkable performance across
diverse domains, thereby prompting researchers to explore their potential for use in …

Videomae: Masked autoencoders are data-efficient learners for self-supervised video pre-training

Z Tong, Y Song, J Wang… - Advances in neural …, 2022 - proceedings.neurips.cc
Pre-training video transformers on extra large-scale datasets is generally required to
achieve premier performance on relatively small datasets. In this paper, we show that video …

Efficientvit: Memory efficient vision transformer with cascaded group attention

X Liu, H Peng, N Zheng, Y Yang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Vision transformers have shown great success due to their high model capabilities.
However, their remarkable performance is accompanied by heavy computation costs, which …

Tensorf: Tensorial radiance fields

A Chen, Z Xu, A Geiger, J Yu, H Su - European conference on computer …, 2022 - Springer
We present TensoRF, a novel approach to model and reconstruct radiance fields. Unlike
NeRF that purely uses MLPs, we model the radiance field of a scene as a 4D tensor, which …