--- license: other language: - en pipeline_tag: text-generation inference: false tags: - transformers - gguf - imatrix - Borealis-10.7B-DPO --- Quantizations of https://huggingface.co/Undi95/Borealis-10.7B-DPO ### Inference Clients/UIs * [llama.cpp](https://github.com/ggerganov/llama.cpp) * [KoboldCPP](https://github.com/LostRuins/koboldcpp) * [ollama](https://github.com/ollama/ollama) * [text-generation-webui](https://github.com/oobabooga/text-generation-webui) * [GPT4All](https://github.com/nomic-ai/gpt4all) * [jan](https://github.com/janhq/jan) --- # From original readme ## Borealis ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/ieXwQCpyL7AX659Z7ivxA.png) Borealis-10.7B-DPO is a 10.7B model made of 48 Mistral 7B layers, finetuned for +70h on 2xA6000 on a big RP and Conversational dataset with llama2 configuration of Axolotl, like SOLAR. This variant had a DPO train on top of it. ## Description This repo contains fp16 files of Borealis-10.7B-DPO, a conversational model. The goal of this model isn't to break all benchmark, but to have a better RP/ERP/Conversational model. It was trained on multiple basic dataset to make it intelligent, but majority of the dataset was basic conversations. ## Dataset used - NobodyExistsOnTheInternet/ToxicQAFinal - teknium/openhermes - unalignment/spicy-3.1 - Doctor-Shotgun/no-robots-sharegpt - Undi95/toxic-dpo-v0.1-sharegpt - Aesir [1], [2], [3-SFW], [3-NSFW] - lemonilia/LimaRP - Squish42/bluemoon-fandom-1-1-rp-cleaned - Undi95/ConversationChronicles-sharegpt-SHARDED (2 sets, modified) ## DPO Dataset used - Intel/orca_dpo_pairs - NobodyExistsOnTheInternet/ToxicDPOqa - Undi95/toxic-dpo-v0.1-NoWarning ## Prompt format: NsChatml ``` <|im_system|> {sysprompt}<|im_end|> <|im_user|> {input}<|im_end|> <|im_bot|> {output}<|im_end|> ```