Exploiting cloze questions for few shot text classification and natural language inference T Schick, H Schütze arXiv preprint arXiv:2001.07676, 2020 | 750 | 2020 |
It's not just size that matters: Small language models are also few-shot learners T Schick, H Schütze arXiv preprint arXiv:2009.07118, 2020 | 505 | 2020 |
Bloom: A 176b-parameter open-access multilingual language model TL Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, R Castagné, ... arXiv preprint arXiv:2211.05100, 2022 | 195 | 2022 |
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models A Srivastava, A Rastogi, A Rao, AAM Shoeb, A Abid, A Fisch, AR Brown, ... arXiv preprint arXiv:2206.04615, 2022 | 177 | 2022 |
Self-diagnosis and self-debiasing: A proposal for reducing corpus-based bias in nlp T Schick, S Udupa, H Schütze Transactions of the Association for Computational Linguistics 9, 1408-1424, 2021 | 121 | 2021 |
Automatically identifying words that can serve as labels for few-shot text classification T Schick, H Schmid, H Schütze arXiv preprint arXiv:2010.13641, 2020 | 93 | 2020 |
Toolformer: Language models can teach themselves to use tools T Schick, J Dwivedi-Yu, R Dessì, R Raileanu, M Lomeli, L Zettlemoyer, ... arXiv preprint arXiv:2302.04761, 2023 | 90 | 2023 |
Rare words: A major problem for contextualized embeddings and how to fix it by attentive mimicking T Schick, H Schütze Proceedings of the AAAI Conference on Artificial Intelligence 34 (05), 8766-8774, 2020 | 80 | 2020 |
Atlas: Few-shot learning with retrieval augmented language models G Izacard, P Lewis, M Lomeli, L Hosseini, F Petroni, T Schick, ... arXiv preprint arXiv 2208, 2022 | 74* | 2022 |
Generating datasets with pretrained language models T Schick, H Schütze arXiv preprint arXiv:2104.07540, 2021 | 73 | 2021 |
Few-shot text generation with pattern-exploiting training T Schick, H Schütze arXiv preprint arXiv:2012.11926, 2020 | 62 | 2020 |
Few-shot text generation with natural language instructions T Schick, H Schütze Proceedings of the 2021 Conference on Empirical Methods in Natural Language …, 2021 | 54 | 2021 |
Augmented language models: a survey G Mialon, R Dessì, M Lomeli, C Nalmpantis, R Pasunuru, R Raileanu, ... arXiv preprint arXiv:2302.07842, 2023 | 45 | 2023 |
Attentive mimicking: Better word embeddings by attending to informative contexts T Schick, H Schütze arXiv preprint arXiv:1904.01617, 2019 | 41 | 2019 |
BERTRAM: Improved word embeddings have big impact on contextualized model performance T Schick, H Schütze arXiv preprint arXiv:1910.07181, 2019 | 34 | 2019 |
PEER: A Collaborative Language Model T Schick, J Dwivedi-Yu, Z Jiang, F Petroni, P Lewis, G Izacard, Q You, ... arXiv preprint arXiv:2208.11663, 2022 | 26 | 2022 |
Learning semantic representations for novel words: Leveraging both form and context T Schick, H Schütze Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 6965-6973, 2019 | 26 | 2019 |
Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor O Honovich, T Scialom, O Levy, T Schick arXiv preprint arXiv:2212.09689, 2022 | 21 | 2022 |
True Few-Shot Learning with Prompts—A Real-World Perspective T Schick, H Schütze Transactions of the Association for Computational Linguistics 10, 716-731, 2022 | 20 | 2022 |
Task-aware retrieval with instructions A Asai, T Schick, P Lewis, X Chen, G Izacard, S Riedel, H Hajishirzi, ... arXiv preprint arXiv:2211.09260, 2022 | 13 | 2022 |