Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
-
Updated
Nov 4, 2021 - Python
setting pretrained_model_name will not only define the model arch but also load the pre-trained checkpoint. We should have another hparam to control whether to load pre-trained checkpoint or not.
Add a description, image, and links to the roberta topic page so that developers can more easily learn about it.
To associate your repository with the roberta topic, visit your repo's landing page and select "manage topics."
Hi, I am so interesting in your project, and wonder if you need contributor and how could I make my own contribution?