XLNet: Generalized Autoregressive Pretraining for Language Understanding
Zhilin Yang, et al.
00
2019-06-19
Abstract
This paper introduces and evaluates the idea described in “XLNet: Generalized Autoregressive Pretraining for Language Understanding”, and reports empirical results that helped shape subsequent work in nlp, transformers.
