DEFAULT MODEL UPDATES
We have updated our standard text models to more performant ones - the following models have been deprecated.
- DistilBERT
- BERT
- Multilingual BERT
We recommend that you DO NOT start new projects with them without first consulting Support or your CSM
Model Name | MultiLingual BERT (Cased) |
Description | A BERT model trained on Wikipedia in 104 languages |
Use For | Text with multiple languages |
Limitations | Long texts will be truncated to at most 512 tokens |
Graft Default | No |
Reference information
Source | bert-base-multilingual-cased - hugging face |
Trained on | Wikipedia |
Paper | 1801.04805 |
Embedding Dimension | 768 |