build: AutoGPTQ for meta-llama/Llama-2-13b-chat-hf: 4bits, gr128, desc_act=False
Browse files- gptq_model-4bit-128g.bin +2 -2
gptq_model-4bit-128g.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e78da402d57ed928bd4ce95f9c8a40bbf6cac7055f54bed221d73ad78ed76d77
|
3 |
+
size 7259817703
|