RoBERTa: A Robustly Optimized BERT Pretraining Approach

First published at 16:36 UTC on September 3rd, 2019.
subscribers

This paper shows that the original BERT model, if trained correctly, can outperform all of the improvements that have been proposed lately, raising questions about the necessity and reasoning behind these.

Abstract:
Language model pretraining has le…

MORE
CategoryScience & Technology
SensitivityNormal - Content that is suitable for ages 16 and over
DISCUSS THIS VIDEO