Continue Pretraining
#7 opened about 1 month ago
by
HuggySSO
Embedding from transformers
#6 opened about 2 months ago
by
tillwenke
"[...] mixture of full fine-tuning and LoRA was used to provide better generalization."
#5 opened 2 months ago
by
bobox