Anthonyg5005 commited on
Commit
4c0c61d
1 Parent(s): 9bf2fb9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +75 -12
README.md CHANGED
@@ -1,12 +1,75 @@
1
- # Exl2 quants for [L3-8B-Niitama-v1](https://huggingface.co/Sao10K/L3-8B-Niitama-v1)
2
-
3
- ## Automatically quantized using the auto quant script from [hf-scripts](https://huggingface.co/anthonyg5005/hf-scripts)
4
-
5
- Would recommend Anthonyg5005 to change up this README to include more info.
6
-
7
- ### BPW:
8
-
9
- [6.0](https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2/tree/6.0bpw)\
10
- [6.5](https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2/tree/6.5bpw)\
11
- [8.0](https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2/tree/8.0bpw)\
12
- [measurement.json](https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2/blob/main/measurement.json)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ pipeline_tag: text-generation
5
+ library_name: ExLlamaV2
6
+ tags:
7
+ - llama
8
+ - llama-3
9
+ license: other
10
+ license_name: llama3
11
+ license_link: https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE
12
+ base_model: R136a1/BeyondInfinity-4x7B
13
+ ---
14
+ # Exl2 quants for [L3-8B-Niitama-v1](https://huggingface.co/Sao10K/L3-8B-Niitama-v1)
15
+
16
+ ## Automatically quantized using the auto quant script from [hf-scripts](https://huggingface.co/anthonyg5005/hf-scripts)
17
+
18
+ Niitama is a model created by [Sao10k](https://huggingface.co/Sao10K). There's currently no card for this model other than the fact it was a horde model.
19
+
20
+ ### BPW:
21
+
22
+ [6.0](https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2/tree/6.0bpw)\
23
+ [6.5](https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2/tree/6.5bpw)\
24
+ [8.0](https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2/tree/8.0bpw)\
25
+ [measurement.json](https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2/blob/main/measurement.json)
26
+
27
+ # How to download:
28
+
29
+ ### oobabooga's downloader
30
+
31
+ use something like [download-model.py](https://github.com/oobabooga/text-generation-webui/blob/main/download-model.py) to download with python requests.\
32
+ Install requirements:
33
+
34
+ ```shell
35
+ pip install requests tqdm
36
+ ```
37
+
38
+ Example for downloading 8bpw:
39
+
40
+ ```shell
41
+ python download-model.py Anthonyg5005/L3-8B-Niitama-v1-exl2-exl2:8.0bpw
42
+ ```
43
+
44
+ ### huggingface-cli
45
+
46
+ You may also use huggingface-cli\
47
+ To install it, install python hf-hub
48
+
49
+ ```shell
50
+ pip install huggingface-hub
51
+ ```
52
+
53
+ Example for 8bpw:
54
+
55
+ ```shell
56
+ huggingface-cli download Anthonyg5005/L3-8B-Niitama-v1-exl2 --local-dir L3-8B-Niitama-v1-exl2-8bpw --revision 8.0bpw
57
+ ```
58
+ ### Git LFS (not recommended)
59
+
60
+ I would recommend the http downloaders over using git, they can resume downloads if failed and are much easier to work with.\
61
+ Make sure to have git and git LFS installed.\
62
+ Example for 8bpw download with git:
63
+
64
+ Have LFS file skip disabled
65
+ ```shell
66
+ # windows
67
+ set GIT_LFS_SKIP_SMUDGE=0
68
+ # linux
69
+ export GIT_LFS_SKIP_SMUDGE=0
70
+ ```
71
+
72
+ Clone repo branch
73
+ ```shell
74
+ git clone https://huggingface.co/Anthonyg5005/L3-8B-Niitama-v1-exl2 -b 8.0bpw
75
+ ```