LLaMA-2-NL: Fine-tuned using LoRa and a translated tokenizer
from transformers import AutoModelForCausalLM, AutoTokenizer
# take the new tokenizer
tokenizer = AutoTokenizer.from_pretrained('llama-2-nl/Llama-2-7b-hf-lora-tokentrans')
model = AutoModelForCausalLM.from_pretrained('llama-2-nl/Llama-2-7b-hf-lora-tokentrans')
- Downloads last month
- 0
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.