File size: 1,421 Bytes
1176794
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40

---
license: other
language:
- en
base_model:
- mistralai/Mistral-Large-Instruct-2407
- NeverSleep/Lumimaid-v0.2-123B
---

# SmartMaid-123b

This **experimental model** is a hybrid creation combining aspects of [Mistral-Large-Instruct-2407](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407) and [Lumimaid-v0.2-123B](https://huggingface.co/NeverSleep/Lumimaid-v0.2-123B) using LoRA (Low-Rank Adaptation) on the `mlp.down_proj` module.

## Model Details

- **Base Model**: Mistral-Large-Instruct-2407
- **Influence Model**: Lumimaid-v0.2-123B
- **Method**: LoRA extraction from Lumimaid and targetted application to Mistral-Large
- **LoRA Configuration**:
  - Rank: 32
  - Alpha: 64
  - Target Module: `mlp.down_proj`

## Process

1. A LoRA was extracted from Lumimaid-v0.2-123B
2. This LoRA was then applied to a fresh instance of Mistral-Large-Instruct-2407 targetting only the mlp.down_proj modules.
3. The resulting model was merged to create this standalone version.

## Purpose

The aim of this model is to incorporate the enhanced prose qualities of Lumimaid-v0.2-123B while retaining the core intelligence and capabilities of Mistral-Large.
By applying the LoRA to the `mlp.down_proj` module, we sought to influence the model's language generation style without significantly altering its underlying knowledge and reasoning abilities.

## Prompt Template
```
<s>[INST] {input} [/INST] {output}</s>
```