Edit model card

llama3-sinhala

LLaMA3 (8B) model instruction finetuned to understand and respond in Sinhala language. meta-llama/Meta-Llama-3-8B-Instruct is finetuned on a reletively large dataset in Sinhala compiled by translating English datasets such as ELI5 and Alpaca. The dataset in hosted in Hugging Face Datasets hub (sinhala-instruction-finetune-large)

The original model is 4-bit quantized and finetuned with a causal language modelling (CLM) objective by adding LoRA adapters with a rank of 16 and a scaling factor of 32.

The finetuned llama3-sinhala model generates better responses in Sinhala compared to the original instruction finetuned model released by Meta. See the github repo llama3-sinhala for more details.

Downloads last month
10
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ihalage/llama3-sinhala

Quantizations
1 model

Dataset used to train ihalage/llama3-sinhala