Update README.md
Browse files
README.md
CHANGED
@@ -220,7 +220,8 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
220 |
trust_remote_code=True,
|
221 |
|
222 |
# If you have installed flash attention 2
|
223 |
-
# attn_implementation="flash_attention_2",
|
|
|
224 |
)
|
225 |
model.to(device)
|
226 |
tokenizer = AutoTokenizer.from_pretrained("OuteAI/Lite-Oute-2-Mamba2Attn-Instruct")
|
|
|
220 |
trust_remote_code=True,
|
221 |
|
222 |
# If you have installed flash attention 2
|
223 |
+
# attn_implementation="flash_attention_2",
|
224 |
+
# torch_dtype=torch.bfloat16,
|
225 |
)
|
226 |
model.to(device)
|
227 |
tokenizer = AutoTokenizer.from_pretrained("OuteAI/Lite-Oute-2-Mamba2Attn-Instruct")
|