Don't Stop Pretraining: Adapt Language Models to Domains and Tasks
Paper
• 2004.10964 • Published
Model Description: This model is a fine-tune checkpoint of bert-large-uncased, fine-tuned on SST-2. This model reaches an accuracy of 99.92 on the dev set.