qwenselfbaseinstruct
This is a merge of pre-trained language models created using mergekit.
Re-injected base model into instruct model in the intermediate layers while keeping input and output layers the same (sophosympatheia gradient).
While this did degrade the overall score of the model compared to instruct in EQ-bench testing (76.9195 down to 73.8068), it removed its issue with misspelling some of the emotion responses and remains notably higher than the base model (60.1027 but without any syntax errors). It did throw in one non-mispelled "didn't match reference" syntax error, I presume it replaced the emotion entirely or used a similar grammatically correct one.
Looking at this as research evidence, it seems like the instruct model picked up something hurting the spelling occasionally specifically in the intermediate layers?
I don't know if there's any other gain from this merge compared to using one or both components, this was for curiosity. Might still be useful as more-compact merge materials if you wanted both base and instruct anyway.
Merge Details
Merge Method
This model was merged using the SLERP merge method.
Models Merged
The following models were included in the merge:
Configuration
The following YAML configuration was used to produce this model:
models:
- model: Qwen/Qwen2.5-14B
merge_method: slerp
base_model: Qwen/Qwen2.5-14B-Instruct
parameters:
t:
- value: [0, 0, 0.3, 0.4, 0.5, 0.6, 0.5, 0.4, 0.3, 0, 0]
dtype: bfloat16
- Downloads last month
- 11