Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation Y Sun, S Wang, S Feng, S Ding, C Pang, J Shang, J Liu, X Chen, Y Zhao, ... arXiv preprint arXiv:2107.02137, 2021 | 339 | 2021 |
Ernie 3.0 titan: Exploring larger-scale knowledge enhanced pre-training for language understanding and generation S Wang, Y Sun, Y Xiang, Z Wu, S Ding, W Gong, S Feng, J Shang, Y Zhao, ... arXiv preprint arXiv:2112.12731, 2021 | 58 | 2021 |
Learning weakly-supervised contrastive representations YHH Tsai, T Li, W Liu, P Liao, R Salakhutdinov, LP Morency arXiv preprint arXiv:2202.06670, 2022 | 12 | 2022 |
Ernie-tiny: A progressive distillation framework for pretrained transformer compression W Su, X Chen, S Feng, J Liu, W Liu, Y Sun, H Tian, H Wu, H Wang arXiv preprint arXiv:2106.02241, 2021 | 10 | 2021 |
Integrating auxiliary information in self-supervised learning YHH Tsai, T Li, W Liu, P Liao, R Salakhutdinov, LP Morency arXiv preprint arXiv:2106.02869, 2021 | 4 | 2021 |