Рет қаралды 11,936
Next Video: • Vision Transformer for...
Bidirectional Encoder Representations from Transformers (BERT) is for pretraining the Transformer models. BERT does not need manually labeled data. BERT can use any books and web documents to automatically generate training data.
Slides: github.com/wangshusen/DeepLea...
Reference:
Devlin, Chang, Lee, and Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In ACL, 2019.