Kukedlc commited on
Commit
2b087b6
1 Parent(s): 8415b63

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ neuralgemma-2b-slerp.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
37
+ neuralgemma-2b-slerp.bf16.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - google/gemma-2-2b-it
4
+ - Kukedlc/Gemma-2-2B-Spanish-1.0
5
+ tags:
6
+ - merge
7
+ - mergekit
8
+ - lazymergekit
9
+ - google/gemma-2-2b-it
10
+ - Kukedlc/Gemma-2-2B-Spanish-1.0
11
+ - autoquant
12
+ - gguf
13
+ ---
14
+
15
+ # NeuralGemma-2B-Slerp
16
+
17
+ NeuralGemma-2B-Slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
18
+ * [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it)
19
+ * [Kukedlc/Gemma-2-2B-Spanish-1.0](https://huggingface.co/Kukedlc/Gemma-2-2B-Spanish-1.0)
20
+
21
+ ## 🧩 Configuration
22
+
23
+ ```yaml
24
+ models:
25
+ - model: google/gemma-2-2b
26
+ # No parameters necessary for base model
27
+ - model: google/gemma-2-2b-it
28
+ parameters:
29
+ density: 0.55
30
+ weight: 0.6
31
+ - model: Kukedlc/Gemma-2-2B-Spanish-1.0
32
+ parameters:
33
+ density: 0.55
34
+ weight: 0.4
35
+ merge_method: dare_ties
36
+ base_model: google/gemma-2-2b
37
+ parameters:
38
+ int8_mask: true
39
+ dtype: float16
40
+ ```
41
+
42
+ ## 💻 Usage
43
+
44
+ ```python
45
+ !pip install -qU transformers accelerate
46
+
47
+ from transformers import AutoTokenizer
48
+ import transformers
49
+ import torch
50
+
51
+ model = "Kukedlc/NeuralGemma-2B-Slerp"
52
+ messages = [{"role": "user", "content": "What is a large language model?"}]
53
+
54
+ tokenizer = AutoTokenizer.from_pretrained(model)
55
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
56
+ pipeline = transformers.pipeline(
57
+ "text-generation",
58
+ model=model,
59
+ torch_dtype=torch.float16,
60
+ device_map="auto",
61
+ )
62
+
63
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
64
+ print(outputs[0]["generated_text"])
65
+ ```
neuralgemma-2b-slerp.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0213dcdc7e197a0c591872791d8cc303048a69fd52ee409001ba5862484421d
3
+ size 1708582464
neuralgemma-2b-slerp.bf16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89a91c8f931c6a548c14487c04c61d04ba9fce4ba07791ca7645eeda81892adf
3
+ size 5235213888