Edit model card
Configuration Parsing Warning: In config.json: "quantization_config.bits" must be an integer

Exllamav2 quant (exl2 / 4.25 bpw) made with ExLlamaV2 v0.1.3

Other EXL2 quants:

Quant Model Size lm_head
2.2
1217 MB
6
2.5
1342 MB
6
3.0
1558 MB
6
3.5
1774 MB
6
3.75
1882 MB
6
4.0
1990 MB
6
4.25
2099 MB
6
5.0
2423 MB
6
6.0
2870 MB
8
6.5
3089 MB
8
8.0
3620 MB
8

Phi-3-mini-128k-instruct- abliterated-v3 -geminified

Credit to u/Anduin1357 on reddit for the name who wrote this comment

My Jupyter "cookbook" to replicate the methodology can be found here, refined library coming soon

What's this?

Well, after my abliterated models, I figured I should cover all the possible ground of such work and introduce a model that acts like the polar opposite of them. This is the result of that, and I feel it lines it up in performance to a certain search engine's AI model series.

Summary

This is microsoft/Phi-3-mini-128k-instruct with orthogonalized bfloat16 safetensor weights, generated with a refined methodology based on that which was described in the preview paper/blog post: 'Refusal in LLMs is mediated by a single direction' which I encourage you to read to understand more.

This model has been orthogonalized to act more like certain rhymes-with-Shmemini models.

Downloads last month
6
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.