Edit model card

Fietje banner

Fietje 2 Instruct

An open and efficient LLM for Dutch

👱‍♀️ Base version - 🤖 Instruct version (this one) - 💬 Chat version - 🚀 GGUF of Instruct

Chat with Fietje here!

This is the instruct version of Fietje, an SFT-tuned (instruction-tuned) variant of the base model. Fietje is an adapated version of microsoft/phi-2, tailored to Dutch text generation by training on 28B tokens. It is small and efficient with a size of 2.7 billion parameters while performing almost on par with more powerful Dutch LLMs of twice its size like GEITje 7B Ultra.

A thorough description of the creation and evaluation of Fietje as well as usage examples are available in this Github repository.

Intended uses & limitations

The same limitations as phi-2, and LLMs in general, apply here. LLMs hallucinate, make mistakes, and should not be trusted. Use at your own risk!

Training and evaluation data

Fietje 2 instruct was finetuned from the base model on the following datasets. Number of training samples per dataset given in brackets, totalling 201,579 samples.

Training procedure

I am thankful to the Flemish Supercomputer Center (VSC) for providing the computational power to accomplish this project. Accounting for waiting for jobs, training took around a day on four nodes of 4x A100 80GB each (16 total). I cannot find the exact time anymore and I do not think that the runtime in all_results.json accounts for interrupted-and-continued jobs.

Training was done with the wonderful alignment-handbook, using DeepSpeed as a back-end. Exact training recipes and SLURM script are given in the Github repository.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 6e-05
  • train_batch_size: 42
  • eval_batch_size: 42
  • seed: 42
  • distributed_type: multi-GPU
  • num_devices: 16
  • total_train_batch_size: 672
  • total_eval_batch_size: 672
  • optimizer: Adam with betas=(0.9,0.98) and epsilon=1e-07
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 3.0

Training results

Training Loss Epoch Step Validation Loss
0.9325 1.0 178 0.9060
0.8687 2.0 356 0.8850
0.8385 3.0 534 0.8818

Framework versions

  • Transformers 4.39.1
  • Pytorch 2.1.2+cu121
  • Datasets 2.18.0
  • Tokenizers 0.15.2

Open LLM Leaderboard Evaluation Results

Results for the English Open LLM Leaderboard. For results specific to Dutch, check out ScandEval.

Detailed results can be found here

Metric Value
Avg. 10.20
IFEval (0-Shot) 27.90
BBH (3-Shot) 17.57
MATH Lvl 5 (4-Shot) 0.53
GPQA (0-shot) 0.00
MuSR (0-shot) 2.91
MMLU-PRO (5-shot) 12.26
Downloads last month
1,028
Safetensors
Model size
2.78B params
Tensor type
BF16
·
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for BramVanroy/fietje-2-instruct

Base model

microsoft/phi-2
Finetuned
(8)
this model
Adapters
2 models
Finetunes
2 models
Quantizations
2 models

Datasets used to train BramVanroy/fietje-2-instruct

Collection including BramVanroy/fietje-2-instruct