NLP领域有哪些必读的经典论文? - 知乎 - Zhihu
www.zhihu.com › question › 380703337两种非RNN的seq2seq:Convolutional Sequence to Sequence Learning;Attention Is All You Need; 预训练语言模型:DEEP CONTEXTUALIZED WORD REPRESENTATIONS;Language Models are Unsupervised Multitask Learners;BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding