Edit model card

gemma-prompt

This model is a fine-tuned version of google/gemma-2b on 2 datasets- dolly-15k for general knowledge and a curated m-a-p/MusicPile

Model description

This model is a completed trained model used for music knowledge and prompt automation from musical vibes.

Intended uses & limitations

Intended use for the model is to have it generate prompts for music that takes into account elements of the surrounding environment, such as the types of buildings nearby, the weather, time of day, and nearby landmarks.

Training and evaluation data

The datasets used to help train the model are the dolly-15k dataset for general purpose of answering questions and following commands, and a second curated m-a-p/MusicPile data used to fine-tune the model specifically for musical vibe and descriptions of different objects, places, and things. We evaluate our model using a portion of the m-a-p/MusicPile.

Training procedure

Split dataset from MusicPile to focus on distilled music knowledge Used dolly for general finetuning

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 2
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 8
  • total_train_batch_size: 16
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_ratio: 0.05
  • training_steps: 888

Framework versions

  • PEFT 0.10.0
  • Transformers 4.40.0
  • Pytorch 2.0.1a0+cxx11.abi
  • Datasets 2.19.0
  • Tokenizers 0.19.1
Downloads last month
2
Unable to determine this model’s pipeline type. Check the docs .

Adapter for