DiTy commited on
Commit
9951bf2
1 Parent(s): e37535f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -5
README.md CHANGED
@@ -14,7 +14,7 @@ language:
14
  pipeline_tag: text-generation
15
  ---
16
 
17
- # DiTy/gemma-2-2b-it-function-calling
18
 
19
  > NB: If you want to use the model to call functions in complex, long and confusing dialogues, it is better to use a larger model [DiTy/gemma-2-9b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-9b-it-function-calling-GGUF).
20
 
@@ -22,6 +22,14 @@ This model is a fine-tuned version of [google/gemma-2-2b-it](https://huggingface
22
  fully annotated by humans only, on the English version of the <ins>*DiTy/function-calling*</ins> dataset.
23
  <!-- Provide a quick summary of what the model is/does. -->
24
 
 
 
 
 
 
 
 
 
25
  ## Model card tree
26
 
27
  * [How prepare your functions (tools) for *Function Calling*](#prepare_func_call)
@@ -71,13 +79,13 @@ import torch
71
  from transformers import AutoTokenizer, AutoModelForCausalLM
72
 
73
  model = AutoModelForCausalLM.from_pretrained(
74
- "DiTy/gemma-2-2b-it-function-calling",
75
  device_map="auto",
76
  torch_dtype=torch.bfloat16, # use float16 or float32 if bfloat16 is not available to you.
77
  cache_dir=PATH_TO_MODEL_DIR, # optional
78
  )
79
  tokenizer = AutoTokenizer.from_pretrained(
80
- "DiTy/gemma-2-2b-it-function-calling",
81
  cache_dir=PATH_TO_MODEL_DIR, # optional
82
  )
83
  ```
@@ -262,7 +270,7 @@ from transformers import pipeline
262
 
263
  generation_pipeline = pipeline(
264
  "text-generation",
265
- model="DiTy/gemma-2-2b-it-function-calling",
266
  model_kwargs={
267
  "torch_dtype": torch.bfloat16, # use float16 or float32 if bfloat16 is not supported for you.
268
  "cache_dir": PATH_TO_MODEL_DIR, # OPTIONAL
@@ -401,7 +409,7 @@ During the learning process, the validation error was approximated to the follow
401
  | **Model** | **Generation Language** | **Approximately Validation Loss** |
402
  | :-----: | :-----: | :-----: |
403
  | [DiTy/gemma-2-9b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-9b-it-function-calling-GGUF) | EN | 0.5 |
404
- | **[DiTy/gemma-2-2b-it-function-calling](https://huggingface.co/DiTy/gemma-2-2b-it-function-calling)** | EN | 0.66 |
405
 
406
  ## Citation
407
 
 
14
  pipeline_tag: text-generation
15
  ---
16
 
17
+ # DiTy/gemma-2-2b-it-function-calling-GGUF
18
 
19
  > NB: If you want to use the model to call functions in complex, long and confusing dialogues, it is better to use a larger model [DiTy/gemma-2-9b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-9b-it-function-calling-GGUF).
20
 
 
22
  fully annotated by humans only, on the English version of the <ins>*DiTy/function-calling*</ins> dataset.
23
  <!-- Provide a quick summary of what the model is/does. -->
24
 
25
+ In addition to **safetensors**, the model is available in **GGUF** formats (in this case, you need to download only a single file),
26
+ *however, it should be borne in mind that this model itself was difficult to master "Function Calling", so it is not recommended to use heavily quantized versions*:
27
+
28
+ | Filename | Quant type | File Size | Description |
29
+ | -------- | ---------- | --------- | ----------- |
30
+ | [gemma-2-2B-it-function-calling-F16.gguf](https://huggingface.co/DiTy/gemma-2-2b-it-function-calling-GGUF/blob/main/gemma-2-2B-it-function-calling-F16.gguf) | F16 | 18.5GB | Base model with float16 *recommended* |
31
+ | [gemma-2-2B-it-function-calling-Q8_0.gguf](https://huggingface.co/DiTy/gemma-2-2b-it-function-calling-GGUF/blob/main/gemma-2-2B-it-function-calling-Q8_0.gguf) | Q8_0 | 9.83GB | Extremely high quality, generally unneeded but max available quant. |
32
+
33
  ## Model card tree
34
 
35
  * [How prepare your functions (tools) for *Function Calling*](#prepare_func_call)
 
79
  from transformers import AutoTokenizer, AutoModelForCausalLM
80
 
81
  model = AutoModelForCausalLM.from_pretrained(
82
+ "DiTy/gemma-2-2b-it-function-calling-GGUF",
83
  device_map="auto",
84
  torch_dtype=torch.bfloat16, # use float16 or float32 if bfloat16 is not available to you.
85
  cache_dir=PATH_TO_MODEL_DIR, # optional
86
  )
87
  tokenizer = AutoTokenizer.from_pretrained(
88
+ "DiTy/gemma-2-2b-it-function-calling-GGUF",
89
  cache_dir=PATH_TO_MODEL_DIR, # optional
90
  )
91
  ```
 
270
 
271
  generation_pipeline = pipeline(
272
  "text-generation",
273
+ model="DiTy/gemma-2-2b-it-function-calling-GGUF",
274
  model_kwargs={
275
  "torch_dtype": torch.bfloat16, # use float16 or float32 if bfloat16 is not supported for you.
276
  "cache_dir": PATH_TO_MODEL_DIR, # OPTIONAL
 
409
  | **Model** | **Generation Language** | **Approximately Validation Loss** |
410
  | :-----: | :-----: | :-----: |
411
  | [DiTy/gemma-2-9b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-9b-it-function-calling-GGUF) | EN | 0.5 |
412
+ | **[DiTy/gemma-2-2b-it-function-calling-GGUF](https://huggingface.co/DiTy/gemma-2-2b-it-function-calling-GGUF)** | EN | 0.66 |
413
 
414
  ## Citation
415