Edit model card

Model Card : gdpr_gemma-2-2b

This model is a fine-tuned version of google/gemma-2b-it on GDPR compliance data using Direct Preference Optimization (DPO).

Model Details

  • Developed by: cycloevan
  • Model type: Causal Language Model
  • Language(s): English
  • License: Apache 2.0 (inherited from base model)
  • Finetuned from model: google/gemma-2b-it
  • GitHub Code: gdpr-gemma2

Uses

This model is designed to assist with GDPR compliance queries and provide information related to data protection regulations.

Training Details

Training Data

The model was fine-tuned on the sims2k/GDPR_QA_instruct_dataset.

Training Procedure

  • Fine-tuning method: Direct Preference Optimization (DPO)
  • Optimizer: AdamW (paged_adamw_32bit)
  • Learning rate: 5e-6
  • Batch size: 1
  • Gradient accumulation steps: 3
  • Number of epochs: 10
  • LR scheduler: Cosine
  • Warmup steps: 2
  • Training regime: LoRA (Low-Rank Adaptation)

LoRA Hyperparameters

  • r: 16
  • lora_alpha: 32
  • lora_dropout: 0.05
  • Target modules: all-linear

Limitations and Bias

  • The model's knowledge is limited to the training data and may not cover all aspects of GDPR or recent updates.
  • The model may occasionally generate incorrect or inconsistent information.
Downloads last month
6
Safetensors
Model size
2.61B params
Tensor type
BF16
·
Inference API
Unable to determine this model's library. Check the docs .

Dataset used to train cycloevan/gdpr_gemma-2-2b