Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ Our Swallow-MX-8x7b-NVE-v0.1 model has undergone continuous pre-training from th
|
|
16 |
|
17 |
## Model Details
|
18 |
|
19 |
-
* **Model type**: Please refer to Mixtral technical report for details on the model architecture.
|
20 |
* **Language(s)**: Japanese English
|
21 |
* **Tokenizer**: This model utilizes the same tokenizer as employed by Mixtral-8x7B-Instruct-v0.1.
|
22 |
* **Contact**: swallow[at]nlp.c.titech.ac.jp
|
|
|
16 |
|
17 |
## Model Details
|
18 |
|
19 |
+
* **Model type**: Please refer to [Mixtral technical report](https://arxiv.org/abs/2401.04088) for details on the model architecture.
|
20 |
* **Language(s)**: Japanese English
|
21 |
* **Tokenizer**: This model utilizes the same tokenizer as employed by Mixtral-8x7B-Instruct-v0.1.
|
22 |
* **Contact**: swallow[at]nlp.c.titech.ac.jp
|