sydonayrex commited on
Commit
204ccef
1 Parent(s): 91a6e89

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -47
README.md CHANGED
@@ -7,52 +7,37 @@ tags:
7
  - transformers
8
  - unsloth
9
  - mistral
10
- - trl
11
- - llama-cpp
12
- - gguf-my-repo
13
- base_model: sydonayrex/Barely-Regal-10.7B
14
  ---
15
 
16
- # sydonayrex/Barely-Regal-10.7B-Q6_K-GGUF
17
- This model was converted to GGUF format from [`sydonayrex/Barely-Regal-10.7B`](https://huggingface.co/sydonayrex/Barely-Regal-10.7B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
18
- Refer to the [original model card](https://huggingface.co/sydonayrex/Barely-Regal-10.7B) for more details on the model.
19
-
20
- ## Use with llama.cpp
21
- Install llama.cpp through brew (works on Mac and Linux)
22
-
23
- ```bash
24
- brew install llama.cpp
25
-
26
- ```
27
- Invoke the llama.cpp server or the CLI.
28
-
29
- ### CLI:
30
- ```bash
31
- llama --hf-repo sydonayrex/Barely-Regal-10.7B-Q6_K-GGUF --hf-file barely-regal-10.7b-q6_k.gguf -p "The meaning to life and the universe is"
32
- ```
33
-
34
- ### Server:
35
- ```bash
36
- llama-server --hf-repo sydonayrex/Barely-Regal-10.7B-Q6_K-GGUF --hf-file barely-regal-10.7b-q6_k.gguf -c 2048
37
- ```
38
-
39
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
40
-
41
- Step 1: Clone llama.cpp from GitHub.
42
- ```
43
- git clone https://github.com/ggerganov/llama.cpp
44
- ```
45
-
46
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
47
- ```
48
- cd llama.cpp && LLAMA_CURL=1 make
49
- ```
50
-
51
- Step 3: Run inference through the main binary.
52
- ```
53
- ./main --hf-repo sydonayrex/Barely-Regal-10.7B-Q6_K-GGUF --hf-file barely-regal-10.7b-q6_k.gguf -p "The meaning to life and the universe is"
54
- ```
55
- or
56
- ```
57
- ./server --hf-repo sydonayrex/Barely-Regal-10.7B-Q6_K-GGUF --hf-file barely-regal-10.7b-q6_k.gguf -c 2048
58
- ```
 
7
  - transformers
8
  - unsloth
9
  - mistral
10
+ - gguf
11
+ base_model: sydonayrex/AI-M3-10.7Bv2
12
+ library_name: transformers
13
+ pipeline_tag: text-generation
14
  ---
15
 
16
+
17
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63603f4a3605bd411c196eef/cU9HbxD9OYFlpSxj6k6Es.jpeg)
18
+
19
+ The name of the model is not a ding against the models performance, more of a commentary of our current ISP infrastructure within the U.S. and the fact that many of our ISPs have not moved into the era of AI yet. They still are primarily monopolies, they put forth arbitrary caps on data transfer, and they do not invest in improving services for customers in moderate to low density population areas. Some of us hit those arbitrary data caps just trying to upload and download models.
20
+
21
+ The base of this model is Mistral Instruct 0.3 that has been supersized using task arithmetic to combine layers, when folding it in on itself. This new model I call Artificial Innovation - Mistral 3, which will show as AI-M3-10.7B as the base model on hub. In just my basic testing, this seems to have worked better than simple passthrough merging of layers, as the LLM has had less issues.
22
+
23
+ In addition to the layer merging, the model has been further fine tuned using SFT using Unsloth to act as a base for further training and experimentation with DPO or ORPO (current DPO project in the process of being trained using Axolotl.)
24
+
25
+ If you find the LLM is acting as if it has had a stroke, see if you have flash attn turned off and enable it is so. This seemed to correct any issues I had when running the model in LM Studio.
26
+
27
+ GGUFs are available here:
28
+
29
+ Q4_K_M and Q8: https://huggingface.co/sydonayrex/Barely-Regal-10.7B-Q6_K-GGUF
30
+
31
+ Q5_K_M: https://huggingface.co/sydonayrex/Barely-Regal-10.7B-Q5_K_M-GGUF
32
+
33
+ Q6_K: https://huggingface.co/sydonayrex/Barely-Regal-10.7B-Q6_K-GGUF
34
+
35
+ # Uploaded model
36
+
37
+ - **Developed by:** sydonayrex
38
+ - **License:** apache-2.0
39
+ - **Finetuned from model :** sydonayrex/AI-M3-10.7Bv2
40
+
41
+ This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
42
+
43
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)