Model Card for Model ID
Lora weights after finetuning Switch-base-128 on WMT16 datasets.
To use Lora weights
from peft import PeftModel
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
base_model = AutoModelForSeq2SeqLM.from_pretrained('google/switch-base-128', use_safetensors=False)
lora_model = PeftModel.from_pretrained(base_model, 'marsggbo/wmt-switch-base-128-lora-ckpt35000')
merged_model = lora_model.merge_and_unload()
merged_model.save_pretrained('./switch-base-128-wmt', state_dict=merged_model.state_dict(), safe_serialization=False)
To use merged model weights
model = AutoModelForSeq2SeqLM.from_pretrained('./switch-base-128-wmt')