default search action
Ahmet Üstün
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [c22]Shivalika Singh, Freddie Vargus, Daniel D'souza, Börje Karlsson, Abinaya Mahendiran, Wei-Yin Ko, Herumb Shandilya, Jay Patel, Deividas Mataciunas, Laura O'Mahony, Mike Zhang, Ramith Hettiarachchi, Joseph Wilson, Marina Machado, Luisa Souza Moura, Dominik Krzeminski, Hakimeh Fadaei, Irem Ergün, Ifeoma Okoh, Aisha Alaagib, Oshan Mudannayake, Zaid Alyafeai, Minh Vu Chien, Sebastian Ruder, Surya Guthikonda, Emad A. Alghamdi, Sebastian Gehrmann, Niklas Muennighoff, Max Bartolo, Julia Kreutzer, Ahmet Üstün, Marzieh Fadaee, Sara Hooker:
Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning. ACL (1) 2024: 11521-11567 - [c21]Arash Ahmadian, Chris Cremer, Matthias Gallé, Marzieh Fadaee, Julia Kreutzer, Olivier Pietquin, Ahmet Üstün, Sara Hooker:
Back to Basics: Revisiting REINFORCE-Style Optimization for Learning from Human Feedback in LLMs. ACL (1) 2024: 12248-12267 - [c20]Ahmet Üstün, Viraat Aryabumi, Zheng Xin Yong, Wei-Yin Ko, Daniel D'souza, Gbemileke Onilude, Neel Bhandari, Shivalika Singh, Hui-Lee Ooi, Amr Kayid, Freddie Vargus, Phil Blunsom, Shayne Longpre, Niklas Muennighoff, Marzieh Fadaee, Julia Kreutzer, Sara Hooker:
Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model. ACL (1) 2024: 15894-15939 - [c19]John Dang, Arash Ahmadian, Kelly Marchisio, Julia Kreutzer, Ahmet Üstün, Sara Hooker:
RLHF Can Speak Many Languages: Unlocking Multilingual Preference Optimization for LLMs. EMNLP 2024: 13134-13156 - [c18]Kelly Marchisio, Saurabh Dash, Hongyu Chen, Dennis Aumiller, Ahmet Üstün, Sara Hooker, Sebastian Ruder:
How Does Quantization Affect Multilingual LLMs? EMNLP (Findings) 2024: 15928-15947 - [c17]Ted Zadouri, Ahmet Üstün, Arash Ahmadian, Beyza Ermis, Acyr Locatelli, Sara Hooker:
Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning. ICLR 2024 - [i25]Shivalika Singh, Freddie Vargus, Daniel D'souza, Börje F. Karlsson, Abinaya Mahendiran, Wei-Yin Ko, Herumb Shandilya, Jay Patel, Deividas Mataciunas, Laura O'Mahony, Mike Zhang, Ramith Hettiarachchi, Joseph Wilson, Marina Machado, Luisa Souza Moura, Dominik Krzeminski, Hakimeh Fadaei, Irem Ergün, Ifeoma Okoh, Aisha Alaagib, Oshan Mudannayake, Zaid Alyafeai, Minh Chien Vu, Sebastian Ruder, Surya Guthikonda, Emad A. Alghamdi, Sebastian Gehrmann, Niklas Muennighoff, Max Bartolo, Julia Kreutzer, Ahmet Üstün, Marzieh Fadaee, Sara Hooker:
Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning. CoRR abs/2402.06619 (2024) - [i24]Ahmet Üstün, Viraat Aryabumi, Zheng Xin Yong, Wei-Yin Ko, Daniel D'souza, Gbemileke Onilude, Neel Bhandari, Shivalika Singh, Hui-Lee Ooi, Amr Kayid, Freddie Vargus, Phil Blunsom, Shayne Longpre, Niklas Muennighoff, Marzieh Fadaee, Julia Kreutzer, Sara Hooker:
Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model. CoRR abs/2402.07827 (2024) - [i23]Arash Ahmadian, Chris Cremer, Matthias Gallé, Marzieh Fadaee, Julia Kreutzer, Olivier Pietquin, Ahmet Üstün, Sara Hooker:
Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs. CoRR abs/2402.14740 (2024) - [i22]Viraat Aryabumi, John Dang, Dwarak Talupuru, Saurabh Dash, David Cairuz, Hangyu Lin, Bharat Venkitesh, Madeline Smith, Jon Ander Campos, Yi Chern Tan, Kelly Marchisio, Max Bartolo, Sebastian Ruder, Acyr Locatelli, Julia Kreutzer, Nick Frosst, Aidan N. Gomez, Phil Blunsom, Marzieh Fadaee, Ahmet Üstün, Sara Hooker:
Aya 23: Open Weight Releases to Further Multilingual Progress. CoRR abs/2405.15032 (2024) - [i21]John Dang, Arash Ahmadian, Kelly Marchisio, Julia Kreutzer, Ahmet Üstün, Sara Hooker:
RLHF Can Speak Many Languages: Unlocking Multilingual Preference Optimization for LLMs. CoRR abs/2407.02552 (2024) - [i20]Kelly Marchisio, Saurabh Dash, Hongyu Chen, Dennis Aumiller, Ahmet Üstün, Sara Hooker, Sebastian Ruder:
How Does Quantization Affect Multilingual LLMs? CoRR abs/2407.03211 (2024) - [i19]Qizhen Zhang, Nikolas Gritsch, Dwaraknath Gnaneshwar, Simon Guo, David Cairuz, Bharat Venkitesh, Jakob N. Foerster, Phil Blunsom, Sebastian Ruder, Ahmet Üstün, Acyr Locatelli:
BAM! Just Like That: Simple and Efficient Parameter Upcycling for Mixture of Experts. CoRR abs/2408.08274 (2024) - [i18]Viraat Aryabumi, Yixuan Su, Raymond Ma, Adrien Morisot, Ivan Zhang, Acyr Locatelli, Marzieh Fadaee, Ahmet Üstün, Sara Hooker:
To Code, or Not To Code? Exploring Impact of Code in Pre-training. CoRR abs/2408.10914 (2024) - [i17]Nikolas Gritsch, Qizhen Zhang, Acyr Locatelli, Sara Hooker, Ahmet Üstün:
Nexus: Specialization meets Adaptability for Efficiently Training Mixture of Experts. CoRR abs/2408.15901 (2024) - [i16]Abdullatif Köksal, Marion Thaler, Ayyoob Imani, Ahmet Üstün, Anna Korhonen, Hinrich Schütze:
MURI: High-Quality Instruction Tuning Datasets for Low-Resource Languages via Reverse Instructions. CoRR abs/2409.12958 (2024) - 2023
- [c16]Arash Ahmadian, Saurabh Dash, Hongyu Chen, Bharat Venkitesh, Stephen Zhen Gou, Phil Blunsom, Ahmet Üstün, Sara Hooker:
Intriguing Properties of Quantization at Scale. NeurIPS 2023 - [i15]Arash Ahmadian, Saurabh Dash, Hongyu Chen, Bharat Venkitesh, Stephen Gou, Phil Blunsom, Ahmet Üstün, Sara Hooker:
Intriguing Properties of Quantization at Scale. CoRR abs/2305.19268 (2023) - [i14]Max Marion, Ahmet Üstün, Luiza Pozzobon, Alex Wang, Marzieh Fadaee, Sara Hooker:
When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale. CoRR abs/2309.04564 (2023) - [i13]Ted Zadouri, Ahmet Üstün, Arash Ahmadian, Beyza Ermis, Acyr Locatelli, Sara Hooker:
Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning. CoRR abs/2309.05444 (2023) - 2022
- [j3]Ahmet Üstün, Arianna Bisazza, Gosse Bouma, Gertjan van Noord:
UDapter: Typology-based Language Adapters for Multilingual Dependency Parsing and Sequence Labeling. Comput. Linguistics 48(3): 555-592 (2022) - [c15]Ahmet Üstün, Asa Cooper Stickland:
When does Parameter-Efficient Transfer Learning Work for Machine Translation? EMNLP 2022: 7919-7933 - [c14]Ahmet Üstün, Arianna Bisazza, Gosse Bouma, Gertjan van Noord, Sebastian Ruder:
Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer. EMNLP 2022: 7934-7949 - [i12]Ahmet Üstün, Asa Cooper Stickland:
When does Parameter-Efficient Transfer Learning Work for Machine Translation? CoRR abs/2205.11277 (2022) - [i11]Ahmet Üstün, Arianna Bisazza, Gosse Bouma, Gertjan van Noord, Sebastian Ruder:
Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer. CoRR abs/2205.12148 (2022) - 2021
- [j2]Ahmet Üstün, Burcu Can:
Incorporating word embeddings in unsupervised morphological segmentation. Nat. Lang. Eng. 27(5): 609-629 (2021) - [j1]Arianna Bisazza, Ahmet Üstün, Stephan Sportel:
On the Difficulty of Translating Free-Order Case-Marking Languages. Trans. Assoc. Comput. Linguistics 9: 1233-1248 (2021) - [c13]Rob van der Goot, Ahmet Üstün, Alan Ramponi, Ibrahim Sharaf, Barbara Plank:
Massive Choice, Ample Tasks (MaChAmp): A Toolkit for Multi-task Learning in NLP. EACL (System Demonstrations) 2021: 176-197 - [c12]Ahmet Üstün, Alexandre Berard, Laurent Besacier, Matthias Gallé:
Multilingual Unsupervised Neural Machine Translation with Denoising Adapters. EMNLP (1) 2021: 6650-6662 - [c11]Masha Medvedeva, Ahmet Üstün, Xiao Xu, Michel Vols, Martijn Wieling:
Automatic Judgement Forecasting for Pending Applications of the European Court of Human Rights. ASAIL/LegalAIIA@ICAIL 2021: 12-23 - [c10]Rob van der Goot, Ibrahim Sharaf, Aizhan Imankulova, Ahmet Üstün, Marija Stepanovic, Alan Ramponi, Siti Oryza Khairunnisa, Mamoru Komachi, Barbara Plank:
From Masked Language Modeling to Translation: Non-English Auxiliary Tasks Improve Zero-shot Spoken Language Understanding. NAACL-HLT 2021: 2479-2497 - [c9]Lukas Edman, Ahmet Üstün, Antonio Toral, Gertjan van Noord:
Unsupervised Translation of German-Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language. WMT@EMNLP 2021: 982-988 - [i10]Rob van der Goot, Ahmet Üstün, Barbara Plank:
On the Effectiveness of Dataset Embeddings in Mono-lingual, Multi-lingual and Zero-shot Conditions. CoRR abs/2103.01273 (2021) - [i9]Rob van der Goot, Ibrahim Sharaf, Aizhan Imankulova, Ahmet Üstün, Marija Stepanovic, Alan Ramponi, Siti Oryza Khairunnisa, Mamoru Komachi, Barbara Plank:
From Masked Language Modeling to Translation: Non-English Auxiliary Tasks Improve Zero-shot Spoken Language Understanding. CoRR abs/2105.07316 (2021) - [i8]Arianna Bisazza, Ahmet Üstün, Stephan Sportel:
On the Difficulty of Translating Free-Order Case-Marking Languages. CoRR abs/2107.06055 (2021) - [i7]Lukas Edman, Ahmet Üstün, Antonio Toral, Gertjan van Noord:
Unsupervised Translation of German-Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language. CoRR abs/2109.12012 (2021) - [i6]Ahmet Üstün, Alexandre Bérard, Laurent Besacier, Matthias Gallé:
Multilingual Unsupervised Neural Machine Translation with Denoising Adapters. CoRR abs/2110.10472 (2021) - 2020
- [c8]Ahmet Üstün, Arianna Bisazza, Gosse Bouma, Gertjan van Noord:
UDapter: Language Adaptation for Truly Universal Dependency Parsing. EMNLP (1) 2020: 2302-2315 - [c7]Bertelt Braaksma, Richard Scholtens, Stan van Suijlekom, Remy Wang, Ahmet Üstün:
FiSSA at SemEval-2020 Task 9: Fine-tuned for Feelings. SemEval@COLING 2020: 1239-1246 - [i5]Ahmet Üstün, Arianna Bisazza, Gosse Bouma, Gertjan van Noord:
UDapter: Language Adaptation for Truly Universal Dependency Parsing. CoRR abs/2004.14327 (2020) - [i4]Rob van der Goot, Ahmet Üstün, Alan Ramponi, Barbara Plank:
Massive Choice, Ample Tasks (MaChAmp): A Toolkit for Multi-task Learning in NLP. CoRR abs/2005.14672 (2020) - [i3]Bertelt Braaksma, Richard Scholtens, Stan van Suijlekom, Remy Wang, Ahmet Üstün:
FiSSA at SemEval-2020 Task 9: Fine-tuned For Feelings. CoRR abs/2007.12544 (2020)
2010 – 2019
- 2019
- [c6]Tommaso Caselli, Ahmet Üstün:
There and Back Again: Cross-Lingual Transfer Learning for Event Detection. CLiC-it 2019 - [c5]Ahmet Üstün, Gosse Bouma, Gertjan van Noord:
Cross-Lingual Word Embeddings for Morphologically Rich Languages. RANLP 2019: 1222-1228 - 2018
- [c4]Ahmet Üstün, Murathan Kurfali, Burcu Can:
Characters or Morphemes: How to Represent Words? Rep4NLP@ACL 2018: 144-153 - 2017
- [c3]Murathan Kurfali, Ahmet Üstün, Burcu Can:
A Trie-structured Bayesian Model for Unsupervised Morphological Segmentation. CICLing (1) 2017: 87-98 - [i2]Burcu Can, Ahmet Üstün, Murathan Kurfali:
Turkish PoS Tagging by Reducing Sparsity with Morpheme Tags in Small Datasets. CoRR abs/1703.03200 (2017) - [i1]Murathan Kurfali, Ahmet Üstün, Burcu Can:
A Trie-Structured Bayesian Model for Unsupervised Morphological Segmentation. CoRR abs/1704.07329 (2017) - 2016
- [c2]Burcu Can, Ahmet Üstün, Murathan Kurfali:
Turkish PoS Tagging by Reducing Sparsity with Morpheme Tags in Small Datasets. CICLing (1) 2016: 320-331 - [c1]Ahmet Üstün, Burcu Can:
Unsupervised Morphological Segmentation Using Neural Word Embeddings. SLSP 2016: 43-53
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-11-18 21:45 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint