ZeroWw commited on
Commit
5fe621e
1 Parent(s): e362aca

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ ghost-8b-beta-1608.fq8.gguf filter=lfs diff=lfs merge=lfs -text
37
+ ghost-8b-beta-1608.silly.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ license: mit
4
+ language:
5
+ - en
6
+ pipeline_tag: text-generation
7
+ ---
8
+
9
+ ZeroWw 'SILLY' version.
10
+ The original model has been quantized (fq8 version)
11
+ and a percentage of it's tensors have
12
+ been modified adding some noise.
13
+
14
+ Full colab: https://colab.research.google.com/drive/1a7seagBzu5l3k3FL4SFk0YJocl7nsDJw?usp=sharing
15
+
16
+ Fast colab: https://colab.research.google.com/drive/1SDD7ox21di_82Y9v68AUoy0PhkxwBVvN?usp=sharing
17
+
18
+ Original reddit post: https://www.reddit.com/r/LocalLLaMA/comments/1ec0s8p/i_made_a_silly_test/
19
+
20
+ I created a program to randomize the weights of a model. The program has 2 parameters: the percentage of weights to modify and the percentage of the original value to randmly apply to each weight.
21
+
22
+ At the end I check the resulting GGUF file for binary differences.
23
+ In this example I set to modify 100% of the weights of Mistral 7b Instruct v0.3 by a maximum of 15% deviation.
24
+
25
+ Since the deviation is calculated on the F32 weights, when quantized to Q8\_0 this changes.
26
+ So, in the end I got a file that compared to the original has:
27
+
28
+ Bytes Difference percentage: 73.04%
29
+
30
+ Average value divergence: 2.98%
31
+
32
+ The cool thing is that chatting with the model I see no apparent difference and the model still works nicely as the original.
33
+
34
+ Since I am running everything on CPU, I could not run perplexity scores or anything computing intensive.
35
+
36
+ As a small test, I asked the model a few questions (like the history of the roman empire) and then fact check its answer using a big model. No errors were detected.
37
+
38
+ Update: all procedure tested and created on COLAB.
39
+
40
+ Created on: Tue Aug 20, 19:24:36
ghost-8b-beta-1608.fq8.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7772f2e30221f1f2cc0b6e35449419e8b70fb38d60162dfe555dc397b41482c2
3
+ size 9525778368
ghost-8b-beta-1608.silly.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bb59be21255518307ede4b14f24e318a4b11902a1e8a269d59700e7122be9ca
3
+ size 9525778368