Text Generation
Safetensors
qwen2
chat
conversational
Eval Results
4-bit precision
luigi86 commited on
Commit
d9f97ff
1 Parent(s): e98fa4c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -8
README.md CHANGED
@@ -121,14 +121,6 @@ model-index:
121
 
122
  Quantized to 4 bpw precision and tested using the `mlx_lm` utility on a 64GiB URAM M1 Max.
123
 
124
- ## Notes on using:
125
-
126
- Requires and optimized for Apple Silicon. Fast enough for rapid back-and-forth as long as it fits on your URAM.
127
-
128
- I tried to serve this with `mlx_lm.serve` per usual, but I got python string indexing errors no matter what I did. It works fine with LM Studio in OpenAI mode.
129
-
130
- I used this with SillyTavern, it worked well.
131
-
132
  See [original model](https://huggingface.co/anthracite-org/magnum-v2-72b) for further details.
133
 
134
  Larger, 8bpw quants available at [mlx-community](https://huggingface.co/mlx-community/magnum-v2-72b).
 
121
 
122
  Quantized to 4 bpw precision and tested using the `mlx_lm` utility on a 64GiB URAM M1 Max.
123
 
 
 
 
 
 
 
 
 
124
  See [original model](https://huggingface.co/anthracite-org/magnum-v2-72b) for further details.
125
 
126
  Larger, 8bpw quants available at [mlx-community](https://huggingface.co/mlx-community/magnum-v2-72b).