Follow
Timo Schick
Timo Schick
NLP Researcher
Verified email at fb.com - Homepage
Title
Cited by
Cited by
Year
Exploiting cloze questions for few shot text classification and natural language inference
T Schick, H Schütze
arXiv preprint arXiv:2001.07676, 2020
7502020
It's not just size that matters: Small language models are also few-shot learners
T Schick, H Schütze
arXiv preprint arXiv:2009.07118, 2020
5052020
Bloom: A 176b-parameter open-access multilingual language model
TL Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, R Castagné, ...
arXiv preprint arXiv:2211.05100, 2022
1952022
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models
A Srivastava, A Rastogi, A Rao, AAM Shoeb, A Abid, A Fisch, AR Brown, ...
arXiv preprint arXiv:2206.04615, 2022
1772022
Self-diagnosis and self-debiasing: A proposal for reducing corpus-based bias in nlp
T Schick, S Udupa, H Schütze
Transactions of the Association for Computational Linguistics 9, 1408-1424, 2021
1212021
Automatically identifying words that can serve as labels for few-shot text classification
T Schick, H Schmid, H Schütze
arXiv preprint arXiv:2010.13641, 2020
932020
Toolformer: Language models can teach themselves to use tools
T Schick, J Dwivedi-Yu, R Dessì, R Raileanu, M Lomeli, L Zettlemoyer, ...
arXiv preprint arXiv:2302.04761, 2023
902023
Rare words: A major problem for contextualized embeddings and how to fix it by attentive mimicking
T Schick, H Schütze
Proceedings of the AAAI Conference on Artificial Intelligence 34 (05), 8766-8774, 2020
802020
Atlas: Few-shot learning with retrieval augmented language models
G Izacard, P Lewis, M Lomeli, L Hosseini, F Petroni, T Schick, ...
arXiv preprint arXiv 2208, 2022
74*2022
Generating datasets with pretrained language models
T Schick, H Schütze
arXiv preprint arXiv:2104.07540, 2021
732021
Few-shot text generation with pattern-exploiting training
T Schick, H Schütze
arXiv preprint arXiv:2012.11926, 2020
622020
Few-shot text generation with natural language instructions
T Schick, H Schütze
Proceedings of the 2021 Conference on Empirical Methods in Natural Language …, 2021
542021
Augmented language models: a survey
G Mialon, R Dessì, M Lomeli, C Nalmpantis, R Pasunuru, R Raileanu, ...
arXiv preprint arXiv:2302.07842, 2023
452023
Attentive mimicking: Better word embeddings by attending to informative contexts
T Schick, H Schütze
arXiv preprint arXiv:1904.01617, 2019
412019
BERTRAM: Improved word embeddings have big impact on contextualized model performance
T Schick, H Schütze
arXiv preprint arXiv:1910.07181, 2019
342019
PEER: A Collaborative Language Model
T Schick, J Dwivedi-Yu, Z Jiang, F Petroni, P Lewis, G Izacard, Q You, ...
arXiv preprint arXiv:2208.11663, 2022
262022
Learning semantic representations for novel words: Leveraging both form and context
T Schick, H Schütze
Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 6965-6973, 2019
262019
Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor
O Honovich, T Scialom, O Levy, T Schick
arXiv preprint arXiv:2212.09689, 2022
212022
True Few-Shot Learning with Prompts—A Real-World Perspective
T Schick, H Schütze
Transactions of the Association for Computational Linguistics 10, 716-731, 2022
202022
Task-aware retrieval with instructions
A Asai, T Schick, P Lewis, X Chen, G Izacard, S Riedel, H Hajishirzi, ...
arXiv preprint arXiv:2211.09260, 2022
132022
The system can't perform the operation now. Try again later.
Articles 1–20