danbert-small-cased / README.md
alexanderfalk's picture
Update README.md
3ef661f
|
raw
history blame
1.02 kB
metadata
language:
  - da
  - en
thumbnail: null
tags:
  - named entity recognition
  - token criticality
license: Apache 2.0
datasets:
  - custom danish dataset
inference: false
metrics:
  - array of metric identifiers

DanBERT

Model description

DanBERT is a danish pre-trained model based on BERT-Base. The pre-trained model has been trained on more than 2 million sentences and 40 millions, danish words. The training has been conducted as part of a thesis.
The model can be found at:

Intended uses & limitations

How to use

from transformers import AutoTokenizer, AutoModel  
tokenizer = AutoTokenizer.from_pretrained("alexanderfalk/danbert-small-cased")  
model = AutoModel.from_pretrained("alexanderfalk/danbert-small-cased") 

BibTeX entry and citation info

@inproceedings{...,
  year={2020},
  title={Anonymization of Danish, Real-Time Data, and Personalized Modelling},
  author={Alexander Falk},
}