build: AutoGPTQ for meta-llama/Llama-2-13b-chat-hf: 4bits, gr128, desc_act=False
Browse files- gptq_model-4bit-128g.bin +3 -0
gptq_model-4bit-128g.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fbea079ace0870618c480b61a652203ccd76d17e707da91a3f5c0693b8d29ff1
|
3 |
+
size 7259678987
|