Synthesizer: Rethinking self-attention for transformer models Y Tay, D Bahri, D Metzler, DC Juan, Z Zhao, C Zheng International conference on machine learning, 10183-10192, 2021 | 231 | 2021 |
Structformer: Joint unsupervised induction of dependency and constituency structure from masked language modeling Y Shen, Y Tay, C Zheng, D Bahri, D Metzler, A Courville arXiv preprint arXiv:2012.00857, 2020 | 27 | 2020 |
Choppy: Cut transformer for ranked list truncation D Bahri, Y Tay, C Zheng, D Metzler, A Tomkins Proceedings of the 43rd International ACM SIGIR Conference on Research and …, 2020 | 13 | 2020 |
Reverse engineering configurations of neural text generation models Y Tay, D Bahri, C Zheng, C Brunk, D Metzler, A Tomkins arXiv preprint arXiv:2004.06201, 2020 | 12 | 2020 |
Synthesizer: Rethinking self-attention in transformer models. arxiv 2020 Y Tay, D Bahri, D Metzler, D Juan, Z Zhao, C Zheng arXiv preprint arXiv:2005.00743 2, 2020 | 11 | 2020 |
Generative models are unsupervised predictors of page quality: A colossal-scale study D Bahri, Y Tay, C Zheng, C Brunk, D Metzler, A Tomkins Proceedings of the 14th ACM International Conference on Web Search and Data …, 2021 | 3 | 2021 |
Surprise: Result List Truncation via Extreme Value Theory D Bahri, C Zheng, Y Tay, D Metzler, A Tomkins arXiv preprint arXiv:2010.09797, 2020 | | 2020 |