gghfez commited on
Commit
56e137f
1 Parent(s): 1836672

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ language:
4
+ - en
5
+ base_model:
6
+ - mistralai/Mistral-Large-Instruct-2407
7
+ - NeverSleep/Lumimaid-v0.2-123B
8
+ ---
9
+
10
+ # SmartMaid-123b
11
+
12
+ This `experimental model` is a hybrid creation combining aspects of [Mistral-Large-Instruct-2407](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407) and [Lumimaid-v0.2-123B](https://huggingface.co/NeverSleep/Lumimaid-v0.2-123B) using LoRA (Low-Rank Adaptation) on the mlp.down_proj module.
13
+
14
+ ## Model Details
15
+
16
+ - **Base Model**: Mistral-Large-Instruct-2407
17
+ - **Influence Model**: Lumimaid-v0.2-123B
18
+ - **Method**: LoRA extraction from Lumimaid and targetted application to Mistral-Large
19
+ - **LoRA Configuration**:
20
+ - Rank: 32
21
+ - Alpha: 64
22
+ - Target Module: `mlp.down_proj`
23
+
24
+ ## Process
25
+
26
+ 1. A LoRA was extracted from Lumimaid-v0.2-123B
27
+ 2. This LoRA was then applied to a fresh instance of Mistral-Large-Instruct-2407 targetting only the mlp.down_proj modules.
28
+ 3. The resulting model was merged to create this standalone version.
29
+
30
+ ## Purpose
31
+
32
+ The aim of this model is to incorporate the enhanced prose qualities of Lumimaid-v0.2-123B while retaining the core intelligence and capabilities of Mistral-Large.
33
+ By applying the LoRA to the `mlp.down_proj` module, we sought to influence the model's language generation style without significantly altering its underlying knowledge and reasoning abilities.
34
+
35
+ ## Prompt Template
36
+
37
+ <s>[INST] {input} [/INST] {output}</s>