[{"title":"( 10 个子文件 5.03MB ) NLP自然语言处理10篇论文.zip","children":[{"title":"Deep contextualized word representations.pdf <span style='color:#111;'> 415.90KB </span>","children":null,"spread":false},{"title":"Attention Is All You Need.pdf <span style='color:#111;'> 2.10MB </span>","children":null,"spread":false},{"title":"ALBERT.pdf <span style='color:#111;'> 409.28KB </span>","children":null,"spread":false},{"title":"Don’t Stop Pretraining.pdf <span style='color:#111;'> 1.68MB </span>","children":null,"spread":false},{"title":"Improving Language Understanding by Generative Pre-Training.pdf <span style='color:#111;'> 528.36KB </span>","children":null,"spread":false},{"title":"BERT- Pre-training of Deep Bidirectional Transformers for Language Understanding.pdf <span style='color:#111;'> 757.00KB </span>","children":null,"spread":false},{"title":"Beyond Accuracy- Behavioral Testing of NLP Models with CheckLis.pdf <span style='color:#111;'> 358.95KB </span>","children":null,"spread":false},{"title":"Pre-trained Models for Natural Language Processing- A Survey.pdf <span style='color:#111;'> 505.08KB </span>","children":null,"spread":false},{"title":"RoBERTa-A Robustly Optimized BERT Pretraining Approach.pdf <span style='color:#111;'> 204.76KB </span>","children":null,"spread":false},{"title":"A Neural Probabilistic Language Model.pdf <span style='color:#111;'> 136.81KB </span>","children":null,"spread":false}],"spread":true}]