grimjim commited on
Commit
8bfb3e5
1 Parent(s): d8ea2ff

Update README.md

Browse files

Updated with links to GGUF quants

Files changed (1) hide show
  1. README.md +61 -58
README.md CHANGED
@@ -1,58 +1,61 @@
1
- ---
2
- base_model:
3
- - mlabonne/NeuralDaredevil-8B-abliterated
4
- - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
5
- - Hastagaras/Halu-OAS-8B-Llama3
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
- license: llama3
11
- license_link: LICENSE
12
- pipeline_tag: text-generation
13
- ---
14
- # Llama-3-Oasis-v1-OAS-8B
15
-
16
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
-
18
- Each merge component was already subjected to Orthogonal Activation Steering (OAS) to mitigate refusals. The resulting text completion model should be versatile for both positive and negative roleplay scenarios and storytelling. Care should be taken when using this model.
19
-
20
- - mlabonne/NeuralDaredevil-8B-abliterated : high MMLU for reasoning
21
- - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS : focus on roleplay
22
- - Hastagaras/Halu-OAS-8B-Llama3 : focus on storytelling
23
-
24
- Built with Meta Llama 3.
25
-
26
- ## Merge Details
27
- ### Merge Method
28
-
29
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [mlabonne/NeuralDaredevil-8B-abliterated](https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated) as a base.
30
-
31
- ### Models Merged
32
-
33
- The following models were also included in the merge:
34
- * [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS)
35
- * [Hastagaras/Halu-OAS-8B-Llama3](https://huggingface.co/Hastagaras/Halu-OAS-8B-Llama3)
36
-
37
- ### Configuration
38
-
39
- The following YAML configuration was used to produce this model:
40
-
41
- ```yaml
42
- base_model: mlabonne/NeuralDaredevil-8B-abliterated
43
- dtype: bfloat16
44
- merge_method: task_arithmetic
45
- slices:
46
- - sources:
47
- - layer_range: [0, 32]
48
- model: mlabonne/NeuralDaredevil-8B-abliterated
49
- - layer_range: [0, 32]
50
- model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
51
- parameters:
52
- weight: 0.3
53
- - layer_range: [0, 32]
54
- model: Hastagaras/Halu-OAS-8B-Llama3
55
- parameters:
56
- weight: 0.3
57
-
58
- ```
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - mlabonne/NeuralDaredevil-8B-abliterated
4
+ - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
5
+ - Hastagaras/Halu-OAS-8B-Llama3
6
+ library_name: transformers
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ license: llama3
11
+ license_link: LICENSE
12
+ pipeline_tag: text-generation
13
+ ---
14
+ # Llama-3-Oasis-v1-OAS-8B
15
+
16
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
+
18
+ Each merge component was already subjected to Orthogonal Activation Steering (OAS) to mitigate refusals. The resulting text completion model should be versatile for both positive and negative roleplay scenarios and storytelling. Care should be taken when using this model.
19
+
20
+ - mlabonne/NeuralDaredevil-8B-abliterated : high MMLU for reasoning
21
+ - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS : focus on roleplay
22
+ - Hastagaras/Halu-OAS-8B-Llama3 : focus on storytelling
23
+
24
+ [static GGUF quants c/o mradermacher](https://huggingface.co/mradermacher/Llama-3-Oasis-v1-OAS-8B-GGUF)
25
+ [weighted/imatrix GGUF quants c/o mradermacher](https://huggingface.co/mradermacher/Llama-3-Oasis-v1-OAS-8B-i1-GGUF)
26
+
27
+ Built with Meta Llama 3.
28
+
29
+ ## Merge Details
30
+ ### Merge Method
31
+
32
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [mlabonne/NeuralDaredevil-8B-abliterated](https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated) as a base.
33
+
34
+ ### Models Merged
35
+
36
+ The following models were also included in the merge:
37
+ * [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS)
38
+ * [Hastagaras/Halu-OAS-8B-Llama3](https://huggingface.co/Hastagaras/Halu-OAS-8B-Llama3)
39
+
40
+ ### Configuration
41
+
42
+ The following YAML configuration was used to produce this model:
43
+
44
+ ```yaml
45
+ base_model: mlabonne/NeuralDaredevil-8B-abliterated
46
+ dtype: bfloat16
47
+ merge_method: task_arithmetic
48
+ slices:
49
+ - sources:
50
+ - layer_range: [0, 32]
51
+ model: mlabonne/NeuralDaredevil-8B-abliterated
52
+ - layer_range: [0, 32]
53
+ model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
54
+ parameters:
55
+ weight: 0.3
56
+ - layer_range: [0, 32]
57
+ model: Hastagaras/Halu-OAS-8B-Llama3
58
+ parameters:
59
+ weight: 0.3
60
+
61
+ ```