|
from peft import AutoPeftModelForCausalLM |
|
from transformers import AutoTokenizer |
|
|
|
from huggingface_hub import login |
|
login("YOUR TOKEN HERE") |
|
|
|
|
|
PROMPT = "[INST]YOUR PROMPT HERE[/INST]" |
|
MAX_LENGTH = 32768 |
|
DEVICE = "cpu" |
|
|
|
|
|
model_id = "agarkovv/CryptoTrader-LM" |
|
base_model_id = "mistralai/Ministral-8B-Instruct-2410" |
|
|
|
model = AutoPeftModelForCausalLM.from_pretrained(model_id) |
|
tokenizer = AutoTokenizer.from_pretrained(base_model_id) |
|
|
|
model = model.to(DEVICE) |
|
model.eval() |
|
inputs = tokenizer( |
|
PROMPT, return_tensors="pt", padding=False, max_length=MAX_LENGTH, truncation=True |
|
) |
|
inputs = {key: value.to(model.device) for key, value in inputs.items()} |
|
|
|
res = model.generate( |
|
**inputs, |
|
use_cache=True, |
|
max_new_tokens=MAX_LENGTH, |
|
) |
|
output = tokenizer.decode(res[0], skip_special_tokens=True) |
|
print(output) |
|
|