--- base_model: hibana2077/Pioneer-2x7B inference: false library_name: transformers merged_models: - HuggingFaceH4/mistral-7b-grok - OpenPipe/mistral-ft-optimized-1218 pipeline_tag: text-generation quantized_by: Suparious tags: - mergekit - merge - 4-bit - AWQ - text-generation - autotrain_compatible - endpoints_compatible --- # hibana2077/Pioneer-2x7B AWQ - Model creator: [hibana2077](https://huggingface.co/hibana2077) - Original model: [Pioneer-2x7B](https://huggingface.co/hibana2077/Pioneer-2x7B) ## Model Summary This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). This model was merged using the SLERP merge method. The following models were included in the merge: * [HuggingFaceH4/mistral-7b-grok](https://huggingface.co/HuggingFaceH4/mistral-7b-grok) * [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218)