Don’t Stop Fine-Tuning: On Training Regimes for Few-Shot Cross-Lingual Transfer with Multilingual Language Models FD Schmidt, I Vulić, G Glavaš Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022 | 13 | 2022 |
SEAGLE: A platform for comparative evaluation of semantic encoders for information retrieval FD Schmidt, M Dietsche, SP Ponzetto, G Glavaš Proceedings of the 2019 Conference on Empirical Methods in Natural Language …, 2019 | 2 | 2019 |
Curb EU enthusiasm: how politicisation shapes bureaucratic responsiveness N Yordanova, A Khokhlova, A Ershova, FD Schmidt, G Glavaš West European Politics, 1-26, 2024 | 1 | 2024 |
Free Lunch: Robust Cross-Lingual Transfer via Model Checkpoint Averaging FD Schmidt, I Vulić, G Glavaš Proceedings of the 61st Annual Meeting of the Association for Computational …, 2023 | 1 | 2023 |
SLICER: Sliced Fine-Tuning for Low-Resource Cross-Lingual Transfer for Named Entity Recognition FD Schmidt, I Vulić, G Glavaš Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022 | 1 | 2022 |
News Without Borders: Domain Adaptation of Multilingual Sentence Embeddings for Cross-lingual News Recommendation A Iana, FD Schmidt, G Glavaš, H Paulheim arXiv preprint arXiv:2406.12634, 2024 | | 2024 |
Self-Distillation for Model Stacking Unlocks Cross-Lingual NLU in 200+ Languages FD Schmidt, P Borchert, I Vulić, G Glavaš arXiv preprint arXiv:2406.12739, 2024 | | 2024 |
Knowledge Distillation vs. Pretraining from Scratch under a Fixed (Computation) Budget MD Bui, FD Schmidt, G Glavaš, K von der Wense arXiv preprint arXiv:2404.19319, 2024 | | 2024 |
One For All & All For One: Bypassing Hyperparameter Tuning with Model Averaging For Cross-Lingual Transfer FD Schmidt, I Vulić, G Glavaš arXiv preprint arXiv:2310.10532, 2023 | | 2023 |