AdamLucek commited on
Commit
be5e6b3
1 Parent(s): 1207b9d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -13,7 +13,7 @@ license: apache-2.0
13
 
14
  Lightweight language model based on Gemma2 2B created by merging multiple fine tuned Gemma2-2B-IT versions to test multilingual conversation capabilities in specialized low parameter language models.
15
 
16
- ### Models Merged
17
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
  This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it) as a base.
19
 
@@ -21,7 +21,7 @@ The following models were included in the merge:
21
  * [VAGOsolutions/SauerkrautLM-gemma-2-2b-it](https://huggingface.co/VAGOsolutions/SauerkrautLM-gemma-2-2b-it)
22
  * [stvlynn/Gemma-2-2b-Chinese-it](https://huggingface.co/stvlynn/Gemma-2-2b-Chinese-it)
23
 
24
- ### Configuration
25
 
26
  The following YAML configuration was used to produce this model:
27
 
@@ -36,7 +36,7 @@ dtype: bfloat16
36
 
37
  ```
38
 
39
- ### Usage
40
 
41
  ```python
42
  from transformers import AutoTokenizer, AutoModelForCausalLM
 
13
 
14
  Lightweight language model based on Gemma2 2B created by merging multiple fine tuned Gemma2-2B-IT versions to test multilingual conversation capabilities in specialized low parameter language models.
15
 
16
+ ## Models Merged
17
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
  This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it) as a base.
19
 
 
21
  * [VAGOsolutions/SauerkrautLM-gemma-2-2b-it](https://huggingface.co/VAGOsolutions/SauerkrautLM-gemma-2-2b-it)
22
  * [stvlynn/Gemma-2-2b-Chinese-it](https://huggingface.co/stvlynn/Gemma-2-2b-Chinese-it)
23
 
24
+ ## 🧩 Configuration
25
 
26
  The following YAML configuration was used to produce this model:
27
 
 
36
 
37
  ```
38
 
39
+ ### 💻 Usage
40
 
41
  ```python
42
  from transformers import AutoTokenizer, AutoModelForCausalLM