A Swedish Bert model
Model description
This model follows the Bert Large model architecture as implemented in Megatron-LM framework. It was trained with a batch size of 512 in 600k steps. The model contains following parameters:
| Hyperparameter | Value |
|---|---|
| 340M | |
| 24 | |
| 16 | |
| 1024 | |
| 30592 |
Training data
The model is pretrained on a Swedish text corpus of around 85 GB from a variety of sources as shown below.
| Dataset | Genre | Size(GB) |
|---|---|---|
| Anföranden | Politics | 0.9 |
| DCEP | Politics | 0.6 |
| DGT | Politics | 0.7 |
| Fass | Medical | 0.6 |
| Författningar | Legal | 0.1 |
| Web data | Misc | 45.0 |
| JRC | Legal | 0.4 |
| Litteraturbanken | Books | 0.3O |
| SCAR | Misc | 28.0 |
| SOU | Politics | 5.3 |
| Subtitles | Drama | 1.3 |
| Wikipedia | Facts | 1.8 |
Intended uses & limitations
The raw model can be used for the usual tasks of masked language modeling or next sentence prediction. It is also often fine-tuned on a downstream task to improve its performance in a specific domain/task.
How to use
from transformers import AutoTokenizer, AutoModelForMaskedLM
tokenizer = AutoTokenizer.from_pretrained("AI-Nordics/bert-large-swedish-cased")
model = AutoModelForMaskedLM.from_pretrained("AI-Nordics/bert-large-swedish-cased")
- Downloads last month
- 11
