Edit model card

Finetuning

This model is a fine-tuned version of Varsha00/finetuned-opusmt-en-to-ta on the samanantar & WMT News dataset. source group: English target group: Gujarati model: transformer

Model description

This model is a sequentially finetuned version of the Helsinki-NLP/opus-mt-en-mul model, designed for translating between English and Gujarati. The model was initially finetuned on the Tamil language using a substantial dataset and subsequently finetuned on Gujarati using a smaller dataset. This approach, known as sequential finetuning or cascaded finetuning, allows the model to leverage the knowledge gained from Hindi to improve its performance on Gujarati translations, despite the limited data available for the latter.

Training and evaluation data

ai4bharath/samanantar WMT-News

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-5
  • warmup_steps: 500
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 16
  • num_epochs: 10

Benchamark Evaluation

  • BLEU score on Tatoeba: 26.2690989442415
  • BLUE score on IN-22: 5.81105590832792

Framework versions

  • Transformers 4.42.3
  • Pytorch 2.1.2
  • Datasets 2.20.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
77M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train Varsha00/finetuned-opusmt-en-ta-gu