关注
Avanika Narayan
Avanika Narayan
在 stanford.edu 的电子邮件经过验证
标题
引用次数
引用次数
年份
On the opportunities and risks of foundation models
R Bommasani, DA Hudson, E Adeli, R Altman, S Arora, S von Arx, ...
arXiv preprint arXiv:2108.07258, 2021
37572021
Ask me anything: A simple strategy for prompting language models
S Arora, A Narayan, MF Chen, L Orr, N Guha, K Bhatia, I Chami, C Re
The Eleventh International Conference on Learning Representations, 2022
1822022
Can foundation models wrangle your data?
A Narayan, I Chami, L Orr, S Arora, C Ré
arXiv preprint arXiv:2205.09911, 2022
1512022
On the opportunities and risks of foundation models. arXiv 2021
R Bommasani, DA Hudson, E Adeli, R Altman, S Arora, S von Arx, ...
arXiv preprint arXiv:2108.07258, 2023
852023
Rekall: Specifying video events using compositions of spatiotemporal labels
DY Fu, W Crichton, J Hong, X Yao, H Zhang, A Truong, A Narayan, ...
arXiv preprint arXiv:1910.02993, 2019
592019
& Liang, P.(2021). On the opportunities and risks of foundation models
R Bommasani, DA Hudson, E Adeli, R Altman, S Arora, S von Arx, ...
arXiv preprint arXiv:2108.07258, 0
55
On the opportunities and risks of foundation models (2021)
R Bommasani, DA Hudson, E Adeli, R Altman, S Arora, S von Arx, ...
arXiv preprint arXiv:2108.07258, 2022
532022
Perfectly balanced: Improving transfer and robustness of supervised contrastive learning
M Chen, DY Fu, A Narayan, M Zhang, Z Song, K Fatahalian, C Ré
International Conference on Machine Learning, 3090-3122, 2022
452022
Language models enable simple systems for generating structured views of heterogeneous data lakes
S Arora, B Yang, S Eyuboglu, A Narayan, A Hojel, I Trummer, C Ré
arXiv preprint arXiv:2304.09433, 2023
432023
Personalized benchmarking with the ludwig benchmarking toolkit
A Narayan, P Molino, K Goel, W Neiswanger, C Re
arXiv preprint arXiv:2111.04260, 2021
122021
Neural, neural everywhere: Controlled generation meets scaffolded, structured dialogue
EA Chi, C Chiam, T Chang, SK Lim, C Rastogi, A Iyabor, Y He, ...
Alexa Prize Proceedings, 2021
112021
Neural generation meets real people: Building a social, informative open-domain dialogue agent
EA Chi, A Paranjape, A See, C Chiam, T Chang, K Kenealy, SK Lim, ...
arXiv preprint arXiv:2207.12021, 2022
102022
TART: A plug-and-play Transformer module for task-agnostic reasoning
K Bhatia, A Narayan, CM De Sa, C Ré
Advances in Neural Information Processing Systems 36, 9751-9788, 2023
52023
Mistral–a journey towards reproducible language model training
S Karamcheti, L Orr, J Bolton, T Zhang, K Goel, A Narayan, R Bommasani, ...
Palo Alto: Stanford Center for Research on Foundation Models, 2021
52021
Automating the Enterprise with Foundation Models
M Wornow, A Narayan, K Opsahl-Ong, Q McIntyre, NH Shah, C Re
arXiv preprint arXiv:2405.03710, 2024
32024
Do Multimodal Foundation Models Understand Enterprise Workflows? A Benchmark for Business Process Management Tasks
M Wornow, A Narayan, B Viggiano, IS Khare, T Verma, T Thompson, ...
arXiv preprint arXiv:2406.13264, 2024
22024
DAMsL: A Meta-Learning Based Approach for Dialogue State Tracking
A Narayan, J Hedtke
https://web. stanford. edu/class/archive/cs/cs224n/cs224n 1204, 1-10, 0
1
Distributed Collaborative Organisations
A Narayan, J Dietz, G Xethalis, C Crawford, D Bollier, H Shadab, ...
2014
Cookbook: A framework for improving LLM generative abilities via programmatic data generating templates
A Narayan, MF Chen, K Bhatia, C Re
First Conference on Language Modeling, 0
系统目前无法执行此操作,请稍后再试。
文章 1–19