Prometheus: Inducing fine-grained evaluation capability in language models S Kim, J Shin, Y Cho, J Jang, S Longpre, H Lee, S Yun, S Shin, S Kim, ... ICLR 2024, 2023 | 81 | 2023 |
Personalized soups: Personalized large language model alignment via post-hoc parameter merging J Jang, S Kim, BY Lin, Y Wang, J Hessel, L Zettlemoyer, H Hajishirzi, ... arXiv preprint arXiv:2310.11564, 2023 | 57 | 2023 |
The cot collection: Improving zero-shot and few-shot learning of language models via chain-of-thought fine-tuning S Kim, SJ Joo, D Kim, J Jang, S Ye, J Shin, M Seo EMNLP 2023, 2023 | 54 | 2023 |
Flask: Fine-grained language model evaluation based on alignment skill sets S Ye, D Kim, S Kim, H Hwang, S Kim, Y Jo, J Thorne, J Kim, M Seo ICLR 2024, 2023 | 53 | 2023 |
Exploring the benefits of training expert language models over instruction tuning J Jang, S Kim, S Ye, D Kim, L Logeswaran, M Lee, K Lee, M Seo ICML 2023, 2023 | 49 | 2023 |
Prometheus 2: An open source language model specialized in evaluating other language models S Kim, J Suk, S Longpre, BY Lin, J Shin, S Welleck, G Neubig, M Lee, ... EMNLP 2024, 2024 | 36 | 2024 |
Mind the gap! injecting commonsense knowledge for abstractive dialogue summarization S Kim, SJ Joo, H Chae, C Kim, S Hwang, J Yeo COLING 2022, 2022 | 16 | 2022 |
Kmmlu: Measuring massive multitask language understanding in korean G Son, H Lee, S Kim, S Kim, N Muennighoff, T Choi, C Park, KM Yoo, ... arXiv preprint arXiv:2402.11548, 2024 | 14 | 2024 |
Aligning to thousands of preferences via system message generalization S Lee, SH Park, S Kim, M Seo NeurIPS 2024, 2024 | 7 | 2024 |
Prometheusvision: Vision-language model as a judge for fine-grained evaluation S Lee, S Kim, SH Park, G Kim, M Seo ACL 2024 Findings, 2024 | 6 | 2024 |
Cotever: Chain of thought prompting annotation toolkit for explanation verification S Kim, SJ Joo, Y Jang, H Chae, J Yeo EACL 2023, 2023 | 6 | 2023 |
Language models as compilers: Simulating pseudocode execution improves algorithmic reasoning in language models H Chae, Y Kim, S Kim, KT Ong, B Kwak, M Kim, S Kim, T Kwon, J Chung, ... EMNLP 2024, 2024 | 4 | 2024 |
Langbridge: Multilingual reasoning without multilingual supervision D Yoon, J Jang, S Kim, S Kim, S Shafayat, M Seo ACL 2024, 2024 | 4 | 2024 |
Consent in Crisis: The Rapid Decline of the AI Data Commons S Longpre, R Mahari, A Lee, C Lund, H Oderinwale, W Brannon, ... NeurIPS 2024, 2024 | 3 | 2024 |
The BiGGen Bench: A Principled Benchmark for Fine-grained Evaluation of Language Models with Language Models S Kim, J Suk, JY Cho, S Longpre, C Kim, D Yoon, G Son, Y Cho, ... arXiv preprint arXiv:2406.05761, 2024 | 3* | 2024 |
Self-Explore to Avoid the Pit: Improving the Reasoning Capabilities of Language Models with Fine-grained Rewards H Hwang, D Kim, S Kim, S Ye, M Seo EMNLP 2024 Findings, 2024 | 3* | 2024 |
Multi-Task Inference: Can Large Language Models Follow Multiple Instructions at Once? G Son, S Baek, S Nam, I Jeong, S Kim ACL 2024, 2024 | 2 | 2024 |
Can Language Models perform Abductive Commonsense Reasoning? S Kim arXiv preprint arXiv:2207.05155, 2022 | 1 | 2022 |
Can Language Models Evaluate Human Written Text? Case Study on Korean Student Writing for Education S Kim, S Kim arXiv preprint arXiv:2407.17022, 2024 | | 2024 |