Ahmet Ustun
Ahmet Ustun
Cohere For AI
Verified email at - Homepage
Cited by
Cited by
UDapter: Language Adaptation for Truly Universal Dependency Parsing
A Üstün, A Bisazza, G Bouma, G van Noord
Proceedings of the 2020 Conference on Empirical Methods in Natural Language …, 2020
Massive choice, ample tasks (MaChAmp): A toolkit for multi-task learning in NLP
R van der Goot, A Üstün, A Ramponi, I Sharaf, B Plank
arXiv preprint arXiv:2005.14672, 2020
Characters or morphemes: How to represent words?
A Üstün, M Kurfalı, B Can
Association for Computational Linguistics, 2018
Multilingual unsupervised neural machine translation with denoising adapters
A Üstün, A Berard, L Besacier, M Gallé
arXiv preprint arXiv:2110.10472, 2021
Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning
T Zadouri, A Üstün, A Ahmadian, B Ermiş, A Locatelli, S Hooker
arXiv preprint arXiv:2309.05444, 2023
When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale
M Marion, A Üstün, L Pozzobon, A Wang, M Fadaee, S Hooker
arXiv preprint arXiv:2309.04564, 2023
Siti Oryza Khairunnisa, Mamoru Komachi, and Barbara Plank. 2021. From masked language modeling to translation: Non-English auxiliary tasks improve zero-shot spoken language …
R Van Der Goot, I Sharaf, A Imankulova, A Üstün, M Stepanovic, ...
Proceedings of the 2021 Conference of the North American Chapter of the …, 2021
Automatic judgement forecasting for pending applications of the European Court of Human Rights
M Medvedeva, A Üstün, X Xu, M Vols, M Wieling
Proceedings of the Fifth Workshop on Automatec Semantic Analysis of …, 2021
Unsupervised morphological segmentation using neural word embeddings
A Üstün, B Can
Statistical Language and Speech Processing: 4th International Conference …, 2016
Intriguing properties of quantization at scale
A Ahmadian, S Dash, H Chen, B Venkitesh, ZS Gou, P Blunsom, A Üstün, ...
Advances in Neural Information Processing Systems 36, 34278-34294, 2023
Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer
A Üstün, A Bisazza, G Bouma, G van Noord, S Ruder
arXiv preprint arXiv:2205.12148, 2022
Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model
A Üstün, V Aryabumi, ZX Yong, WY Ko, D D'souza, G Onilude, N Bhandari, ...
arXiv preprint arXiv:2402.07827, 2024
Aya dataset: An open-access collection for multilingual instruction tuning
S Singh, F Vargus, D Dsouza, BF Karlsson, A Mahendiran, WY Ko, ...
arXiv preprint arXiv:2402.06619, 2024
Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs
A Ahmadian, C Cremer, M Gallé, M Fadaee, J Kreutzer, A Üstün, ...
arXiv preprint arXiv:2402.14740, 2024
Turkish pos tagging by reducing sparsity with morpheme tags in small datasets
B Can, A Üstün, M Kurfalı
Computational Linguistics and Intelligent Text Processing: 17th …, 2018
From Masked Language Modeling to Translation: Non-English Auxiliary Tasks Improve Zero-shot Spoken Language Understanding
R van der Goot, I Sharaf, A Imankulova, A Üstün, M Stepanović, ...
arXiv preprint arXiv:2105.07316, 2021
UDapter: Typology-based Language Adapters for Multilingual Dependency Parsing and Sequence Labeling
A Üstün, A Bisazza, G Bouma, G Noord
Computational Linguistics 48 (3), 555-592, 2022
On the Difficulty of Translating Free-Order Case-Marking Languages
A Bisazza, A Üstün, S Sportel
Transactions of the Association for Computational Linguistics 9, 1233-1248, 2021
Unsupervised Translation of German--Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language
L Edman, A Üstün, A Toral, G van Noord
arXiv preprint arXiv:2109.12012, 2021
When does Parameter-Efficient Transfer Learning Work for Machine Translation?
A Üstün, AC Stickland
arXiv preprint arXiv:2205.11277, 2022
The system can't perform the operation now. Try again later.
Articles 1–20