Anthonyg5005
commited on
Commit
•
157ee7d
1
Parent(s):
14324a5
head bits 8 default
Browse filesdoesn't make much of a vram difference for better results
auto-exl2-upload/auto-exl2-upload.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46c9216642f00007b56581890c3ca89426a6ce7148bbca4f19a5ec0a04ac5e16
|
3 |
+
size 7202
|
auto-exl2-upload/exl2-quant.py
CHANGED
@@ -168,7 +168,7 @@ for bpw in bpwvalue:
|
|
168 |
os.makedirs(f"{model}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
|
169 |
subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
|
170 |
#more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
|
171 |
-
result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {model}-exl2-{bpw}bpw -b {bpw}{mskip}", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
|
172 |
if result.returncode != 0:
|
173 |
print("Quantization failed.")
|
174 |
sys.exit("Exiting...")
|
|
|
168 |
os.makedirs(f"{model}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
|
169 |
subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
|
170 |
#more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
|
171 |
+
result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {model}-exl2-{bpw}bpw -b {bpw}{mskip} -hb 8", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
|
172 |
if result.returncode != 0:
|
173 |
print("Quantization failed.")
|
174 |
sys.exit("Exiting...")
|
exl2-multi-quant-local/exl2-multi-quant-local.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:030dc763b103c4d58919a31ed238ff839bd0b1b0ec5dd00a0bcdd1571eed778b
|
3 |
+
size 6211
|
exl2-multi-quant-local/exl2-quant.py
CHANGED
@@ -122,7 +122,7 @@ for bpw in bpwvalue:
|
|
122 |
os.makedirs(f"{modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
|
123 |
subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
|
124 |
#more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
|
125 |
-
result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw -b {bpw}{mskip}", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
|
126 |
if result.returncode != 0:
|
127 |
print("Quantization failed.")
|
128 |
sys.exit("Exiting...")
|
|
|
122 |
os.makedirs(f"{modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
|
123 |
subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
|
124 |
#more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
|
125 |
+
result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw -b {bpw}{mskip} -hb 8", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
|
126 |
if result.returncode != 0:
|
127 |
print("Quantization failed.")
|
128 |
sys.exit("Exiting...")
|