mbrack commited on
Commit
f5cefd1
1 Parent(s): dc331a6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -13,7 +13,7 @@ pipeline_tag: text-generation
13
  > A [polyglot](https://en.wikipedia.org/wiki/Multilingualism#In_individuals) language model for the [Occident](https://en.wikipedia.org/wiki/Occident).
14
  >
15
 
16
- **Occiglot-7B-DE-EN** is a generative language model with 7B parameters for German and English and trained by the [Occiglot Research Collective](occiglot.eu).
17
  It is based on [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) and trained on 114B tokens of additional multilingual and code data with a block size of 8,192 tokens per sample.
18
  Note that the model is a general-purpose base model and was not instruction-fine-tuned nor optimized for chat or other applications. We make an instruction tuned variant available as [occiglot-7b-de-en-instruct](https://huggingface.co/occiglot/occiglot-7b-de-en-instruct)
19
 
@@ -29,7 +29,7 @@ If you want to train a model for your own language or are working on evaluations
29
  - **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.html)
30
  - **Compute resources:** [HessianAI's 42](https://hessian.ai/)
31
  - **Contributors:** Manuel Brack, Patrick Schramowski, Pedro Ortiz, Malte Ostendorff, Fabio Barth, Georg Rehm, Kristian Kersting
32
- - **Research labs:** [Occiglot](https://ociglot.eu) with support from [SAINT](https://www.dfki.de/en/web/research/research-departments/foundations-of-systems-ai) and [SLT](https://www.dfki.de/en/web/research/research-departments/speech-and-language-technology)
33
  - **Contact:** [Discord](https://discord.gg/wUpvYs4XvM) [[email protected]](mailto:[email protected])
34
 
35
  ### How to use
 
13
  > A [polyglot](https://en.wikipedia.org/wiki/Multilingualism#In_individuals) language model for the [Occident](https://en.wikipedia.org/wiki/Occident).
14
  >
15
 
16
+ **Occiglot-7B-DE-EN** is a generative language model with 7B parameters for German and English and trained by the [Occiglot Research Collective](https://occiglot.github.io/occiglot/).
17
  It is based on [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) and trained on 114B tokens of additional multilingual and code data with a block size of 8,192 tokens per sample.
18
  Note that the model is a general-purpose base model and was not instruction-fine-tuned nor optimized for chat or other applications. We make an instruction tuned variant available as [occiglot-7b-de-en-instruct](https://huggingface.co/occiglot/occiglot-7b-de-en-instruct)
19
 
 
29
  - **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.html)
30
  - **Compute resources:** [HessianAI's 42](https://hessian.ai/)
31
  - **Contributors:** Manuel Brack, Patrick Schramowski, Pedro Ortiz, Malte Ostendorff, Fabio Barth, Georg Rehm, Kristian Kersting
32
+ - **Research labs:** [Occiglot](https://occiglot.github.io/occiglot/) with support from [SAINT](https://www.dfki.de/en/web/research/research-departments/foundations-of-systems-ai) and [SLT](https://www.dfki.de/en/web/research/research-departments/speech-and-language-technology)
33
  - **Contact:** [Discord](https://discord.gg/wUpvYs4XvM) [[email protected]](mailto:[email protected])
34
 
35
  ### How to use