Don't Stop Pretraining: Adapt Language Models to Domains and Tasks
Paper
•
2004.10964
•
Published
Model Description: This model is a fine-tune checkpoint of bert-large-uncased, fine-tuned on SST-2. This model reaches an accuracy of 99.92 on the dev set.