Exllamav2 quant (exl2 / 3.75 bpw) made with ExLlamaV2 v0.1.3
Other EXL2 quants:
Quant | Model Size | lm_head |
---|---|---|
Phi-3-mini-128k-instruct- abliterated-v3 -geminified
Credit to u/Anduin1357 on reddit for the name who wrote this comment
My Jupyter "cookbook" to replicate the methodology can be found here, refined library coming soon
What's this?
Well, after my abliterated models, I figured I should cover all the possible ground of such work and introduce a model that acts like the polar opposite of them. This is the result of that, and I feel it lines it up in performance to a certain search engine's AI model series.
Summary
This is microsoft/Phi-3-mini-128k-instruct with orthogonalized bfloat16 safetensor weights, generated with a refined methodology based on that which was described in the preview paper/blog post: 'Refusal in LLMs is mediated by a single direction' which I encourage you to read to understand more.
This model has been orthogonalized to act more like certain rhymes-with-Shmemini models.
- Downloads last month
- 6