File size: 1,017 Bytes
e0e7eb7
 
 
 
 
 
 
 
 
 
 
0c02ee4
3ef661f
 
 
e0e7eb7
 
 
 
 
 
 
 
79dd255
e0e7eb7
 
 
 
 
 
139be99
0a41d94
 
e0e7eb7
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
---
language: 
- da
- en
thumbnail: 
tags:
- named entity recognition
- token criticality
license: Apache 2.0
datasets:
- custom danish dataset
inference: false
metrics:
- array of metric identifiers
---

# DanBERT

## Model description

DanBERT is a danish pre-trained model based on BERT-Base. The pre-trained model has been trained on more than 2 million sentences and 40 millions, danish words. The training has been conducted as part of a thesis.  
The model can be found at:

* [danbert-da](https://huggingface.co/alexanderfalk/danbert-small-cased)

## Intended uses & limitations

#### How to use

```python
from transformers import AutoTokenizer, AutoModel  
tokenizer = AutoTokenizer.from_pretrained("alexanderfalk/danbert-small-cased")  
model = AutoModel.from_pretrained("alexanderfalk/danbert-small-cased") 
```

### BibTeX entry and citation info

```bibtex
@inproceedings{...,
  year={2020},
  title={Anonymization of Danish, Real-Time Data, and Personalized Modelling},
  author={Alexander Falk},
}
```