Update README.md
Browse files
README.md
CHANGED
@@ -14,7 +14,7 @@ language:
|
|
14 |
pipeline_tag: text-generation
|
15 |
---
|
16 |
|
17 |
-
# DiTy/gemma-2-2b-it-function-calling
|
18 |
|
19 |
> NB: If you want to use the model to call functions in complex, long and confusing dialogues, it is better to use a larger model [DiTy/gemma-2-9b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-9b-it-function-calling-GGUF).
|
20 |
|
@@ -22,6 +22,14 @@ This model is a fine-tuned version of [google/gemma-2-2b-it](https://huggingface
|
|
22 |
fully annotated by humans only, on the English version of the <ins>*DiTy/function-calling*</ins> dataset.
|
23 |
<!-- Provide a quick summary of what the model is/does. -->
|
24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
## Model card tree
|
26 |
|
27 |
* [How prepare your functions (tools) for *Function Calling*](#prepare_func_call)
|
@@ -71,13 +79,13 @@ import torch
|
|
71 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
72 |
|
73 |
model = AutoModelForCausalLM.from_pretrained(
|
74 |
-
"DiTy/gemma-2-2b-it-function-calling",
|
75 |
device_map="auto",
|
76 |
torch_dtype=torch.bfloat16, # use float16 or float32 if bfloat16 is not available to you.
|
77 |
cache_dir=PATH_TO_MODEL_DIR, # optional
|
78 |
)
|
79 |
tokenizer = AutoTokenizer.from_pretrained(
|
80 |
-
"DiTy/gemma-2-2b-it-function-calling",
|
81 |
cache_dir=PATH_TO_MODEL_DIR, # optional
|
82 |
)
|
83 |
```
|
@@ -262,7 +270,7 @@ from transformers import pipeline
|
|
262 |
|
263 |
generation_pipeline = pipeline(
|
264 |
"text-generation",
|
265 |
-
model="DiTy/gemma-2-2b-it-function-calling",
|
266 |
model_kwargs={
|
267 |
"torch_dtype": torch.bfloat16, # use float16 or float32 if bfloat16 is not supported for you.
|
268 |
"cache_dir": PATH_TO_MODEL_DIR, # OPTIONAL
|
@@ -401,7 +409,7 @@ During the learning process, the validation error was approximated to the follow
|
|
401 |
| **Model** | **Generation Language** | **Approximately Validation Loss** |
|
402 |
| :-----: | :-----: | :-----: |
|
403 |
| [DiTy/gemma-2-9b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-9b-it-function-calling-GGUF) | EN | 0.5 |
|
404 |
-
| **[DiTy/gemma-2-2b-it-function-calling](https://huggingface.co/DiTy/gemma-2-2b-it-function-calling)** | EN | 0.66 |
|
405 |
|
406 |
## Citation
|
407 |
|
|
|
14 |
pipeline_tag: text-generation
|
15 |
---
|
16 |
|
17 |
+
# DiTy/gemma-2-2b-it-function-calling-GGUF
|
18 |
|
19 |
> NB: If you want to use the model to call functions in complex, long and confusing dialogues, it is better to use a larger model [DiTy/gemma-2-9b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-9b-it-function-calling-GGUF).
|
20 |
|
|
|
22 |
fully annotated by humans only, on the English version of the <ins>*DiTy/function-calling*</ins> dataset.
|
23 |
<!-- Provide a quick summary of what the model is/does. -->
|
24 |
|
25 |
+
In addition to **safetensors**, the model is available in **GGUF** formats (in this case, you need to download only a single file),
|
26 |
+
*however, it should be borne in mind that this model itself was difficult to master "Function Calling", so it is not recommended to use heavily quantized versions*:
|
27 |
+
|
28 |
+
| Filename | Quant type | File Size | Description |
|
29 |
+
| -------- | ---------- | --------- | ----------- |
|
30 |
+
| [gemma-2-2B-it-function-calling-F16.gguf](https://huggingface.co/DiTy/gemma-2-2b-it-function-calling-GGUF/blob/main/gemma-2-2B-it-function-calling-F16.gguf) | F16 | 18.5GB | Base model with float16 *recommended* |
|
31 |
+
| [gemma-2-2B-it-function-calling-Q8_0.gguf](https://huggingface.co/DiTy/gemma-2-2b-it-function-calling-GGUF/blob/main/gemma-2-2B-it-function-calling-Q8_0.gguf) | Q8_0 | 9.83GB | Extremely high quality, generally unneeded but max available quant. |
|
32 |
+
|
33 |
## Model card tree
|
34 |
|
35 |
* [How prepare your functions (tools) for *Function Calling*](#prepare_func_call)
|
|
|
79 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
80 |
|
81 |
model = AutoModelForCausalLM.from_pretrained(
|
82 |
+
"DiTy/gemma-2-2b-it-function-calling-GGUF",
|
83 |
device_map="auto",
|
84 |
torch_dtype=torch.bfloat16, # use float16 or float32 if bfloat16 is not available to you.
|
85 |
cache_dir=PATH_TO_MODEL_DIR, # optional
|
86 |
)
|
87 |
tokenizer = AutoTokenizer.from_pretrained(
|
88 |
+
"DiTy/gemma-2-2b-it-function-calling-GGUF",
|
89 |
cache_dir=PATH_TO_MODEL_DIR, # optional
|
90 |
)
|
91 |
```
|
|
|
270 |
|
271 |
generation_pipeline = pipeline(
|
272 |
"text-generation",
|
273 |
+
model="DiTy/gemma-2-2b-it-function-calling-GGUF",
|
274 |
model_kwargs={
|
275 |
"torch_dtype": torch.bfloat16, # use float16 or float32 if bfloat16 is not supported for you.
|
276 |
"cache_dir": PATH_TO_MODEL_DIR, # OPTIONAL
|
|
|
409 |
| **Model** | **Generation Language** | **Approximately Validation Loss** |
|
410 |
| :-----: | :-----: | :-----: |
|
411 |
| [DiTy/gemma-2-9b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-9b-it-function-calling-GGUF) | EN | 0.5 |
|
412 |
+
| **[DiTy/gemma-2-2b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-2b-it-function-calling-GGUF)** | EN | 0.66 |
|
413 |
|
414 |
## Citation
|
415 |
|