sayhan commited on
Commit
302fb77
1 Parent(s): c35332e

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -0
README.md ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/gemma-7b-it
3
+ language:
4
+ - en
5
+ pipeline_tag: text-generation
6
+ license: other
7
+ model_type: gemma
8
+ library_name: transformers
9
+ inference: false
10
+ ---
11
+ ![image/webp](https://cdn-uploads.huggingface.co/production/uploads/65aa2d4b356bf23b4a4da247/NQAvp6NRHlNILyWWFlrA7.webp)
12
+ ## Google Gemma 7B Instruct
13
+ - **Model creator:** [Google](https://huggingface.co/google)
14
+ - **Original model:** [gemma-7b-it](https://huggingface.co/google/gemma-7b-it)
15
+ - [**Terms of use**](https://www.kaggle.com/models/google/gemma/license/consent)
16
+ <!-- description start -->
17
+ ## Description
18
+ This repo contains GGUF format model files for [Google's Gemma 7B Instruct](https://huggingface.co/google/gemma-7b-it)
19
+
20
+ ## Original model
21
+ - **Developed by:** [Google](https://huggingface.co/google)
22
+
23
+ ### Description
24
+ Gemma is a family of lightweight, state-of-the-art open models from Google,
25
+ built from the same research and technology used to create the Gemini models.
26
+ They are text-to-text, decoder-only large language models, available in English,
27
+ with open weights, pre-trained variants, and instruction-tuned variants. Gemma
28
+ models are well-suited for a variety of text generation tasks, including
29
+ question answering, summarization, and reasoning. Their relatively small size
30
+ makes it possible to deploy them in environments with limited resources such as
31
+ a laptop, desktop or your own cloud infrastructure, democratizing access to
32
+ state of the art AI models and helping foster innovation for everyone.
33
+
34
+ ## Quantizon types
35
+ | quantization method | bits | size | description | recommended |
36
+ |---------------------|------|----------|-----------------------------------------------------|-------------|
37
+ | Q3_K_S | 3 | 20.4 GB | very small, high quality loss | ❌ |
38
+ | Q3_K_L | 3 | 26.4 GB | small, substantial quality loss | ❌ |
39
+ | Q4_0 | 4 | 26.4 GB | legacy; small, very high quality loss | ❌ |
40
+ | Q4_K_M | 4 | 28.4 GB | medium, balanced quality | ✅ |
41
+ | Q5_0 | 5 | 33.2 GB | legacy; medium, balanced quality | ❌ |
42
+ | Q5_K_S | 5 | 32.2 GB | large, low quality loss | ✅ |
43
+ | Q5_K_M | 5 | 33.2 GB | large, very low quality loss | ✅ |
44
+ | Q6_K | 6 | 38.4 GB | very large, extremely low quality loss | ❌ |
45
+ | Q8_0 | 8 | 49.6 GB | very large, extremely low quality loss | ❌ |
46
+
47
+ ## Usage
48
+ You can use this model with the latest builds of LM Studio and llama.cpp.
49
+ If you're new to the world of large language models, I recommend starting with LM Studio.
50
+ <!-- description end -->