Comparison of size in 2 versions. Training. BERT is pretrained on 2 tasks: Masked Language Modeling (MLM) and Next Sentence Prediction (NSP).
確定! 回上一頁