--- license: other language: - en base_model: - mistralai/Mistral-Large-Instruct-2407 - NeverSleep/Lumimaid-v0.2-123B --- # SmartMaid-123b This `experimental model` is a hybrid creation combining aspects of [Mistral-Large-Instruct-2407](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407) and [Lumimaid-v0.2-123B](https://huggingface.co/NeverSleep/Lumimaid-v0.2-123B) using LoRA (Low-Rank Adaptation) on the mlp.down_proj module. ## Model Details - **Base Model**: Mistral-Large-Instruct-2407 - **Influence Model**: Lumimaid-v0.2-123B - **Method**: LoRA extraction from Lumimaid and targetted application to Mistral-Large - **LoRA Configuration**: - Rank: 32 - Alpha: 64 - Target Module: `mlp.down_proj` ## Process 1. A LoRA was extracted from Lumimaid-v0.2-123B 2. This LoRA was then applied to a fresh instance of Mistral-Large-Instruct-2407 targetting only the mlp.down_proj modules. 3. The resulting model was merged to create this standalone version. ## Purpose The aim of this model is to incorporate the enhanced prose qualities of Lumimaid-v0.2-123B while retaining the core intelligence and capabilities of Mistral-Large. By applying the LoRA to the `mlp.down_proj` module, we sought to influence the model's language generation style without significantly altering its underlying knowledge and reasoning abilities. ## Prompt Template [INST] {input} [/INST] {output}