Upload glm-4-9b-chat-1m-Q6_K.gguf with huggingface_hub
Browse files- .gitattributes +1 -0
- glm-4-9b-chat-1m-Q6_K.gguf +3 -0
.gitattributes
CHANGED
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
glm-4-9b-chat-1m-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
glm-4-9b-chat-1m-Q6_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
glm-4-9b-chat-1m-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
glm-4-9b-chat-1m-Q6_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
glm-4-9b-chat-1m-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
glm-4-9b-chat-1m-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b2653e5d4bebeff7e4985370f791f2a42ee1af75360573699d8bdd13dfd1087f
|
3 |
+
size 8330924832
|