a43992899 commited on
Commit
5ba1103
1 Parent(s): 6ebafbe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -9
README.md CHANGED
@@ -90,8 +90,6 @@ dump_wav('cm_music_piece.wav', audio, sample_rate=44100, use_int16=True)
90
 
91
  ## Example Prompts
92
 
93
- **Our model produces symbolic music(ABC notation) well in the following prompts.** Here are some musical tasks.
94
-
95
  ### Function: Chord Conditioned Music Generation
96
  ```
97
  Develop a musical piece using the given chord progression.
@@ -193,18 +191,12 @@ ChatMusician is pretrained on the 🤗 [MusicPile](https://huggingface.co/datase
193
  And supervised finetuned on 1.1M samples(2:1 ratio between music scores
194
  and music knowledge & music summary data) from MusicPile. Check our [paper](http://arxiv.org/abs/2402.16153) for more details.
195
 
196
- ## Training Procedure
197
-
198
- We initialized a fp16-precision ChatMusician-Base from the LLaMA2-7B-Base weights, and applied a continual pre-training plus fine-tuning pipeline. LoRA adapters were integrated into the attention and MLP layers, with additional training on embeddings and all linear layers. The maximum sequence length
199
- was 2048. We utilized 16 80GB-A800 GPUs for one epoch pre-training and 8 32GB-V100 GPUs for two epoch fine-tuning. DeepSpeed was employed for memory efficiency, and the AdamW optimizer was used with a 1e-4 learning rate and a 5% warmup cosine scheduler. Gradient clipping was set at 1.0. The LoRA parameters dimension, alpha, and
200
- dropout were set to 64, 16, and 0.1, with a batch size of 8.
201
-
202
 
203
  ## Evaluation
204
 
205
  1. Music understanding abilities are evaluated on the [MusicTheoryBench](https://huggingface.co/datasets/m-a-p/MusicTheoryBench). The following figure is zero-shot accuracy on MusicTheoryBench.
206
  We included GPT-3.5, GPT-4, LLaMA2-7B-Base, ChatMusician-Base, and ChatMusician. The blue bar represents the performance on the music knowledge metric, and the red bar represents the music reasoning metric. The dashed line corresponds to a random baseline, with a score of 25%.![MusicTheoryBench_result](./MusicTheoryBench_result_plt.png)
207
- 2. General language abilities of ChatMusician are evaluated on the [Massive Multitask Language Understanding (MMLU) dataset](https://huggingface.co/datasets/lukaemon/mmlu).
208
 
209
 
210
 
 
90
 
91
  ## Example Prompts
92
 
 
 
93
  ### Function: Chord Conditioned Music Generation
94
  ```
95
  Develop a musical piece using the given chord progression.
 
191
  And supervised finetuned on 1.1M samples(2:1 ratio between music scores
192
  and music knowledge & music summary data) from MusicPile. Check our [paper](http://arxiv.org/abs/2402.16153) for more details.
193
 
 
 
 
 
 
 
194
 
195
  ## Evaluation
196
 
197
  1. Music understanding abilities are evaluated on the [MusicTheoryBench](https://huggingface.co/datasets/m-a-p/MusicTheoryBench). The following figure is zero-shot accuracy on MusicTheoryBench.
198
  We included GPT-3.5, GPT-4, LLaMA2-7B-Base, ChatMusician-Base, and ChatMusician. The blue bar represents the performance on the music knowledge metric, and the red bar represents the music reasoning metric. The dashed line corresponds to a random baseline, with a score of 25%.![MusicTheoryBench_result](./MusicTheoryBench_result_plt.png)
199
+ 2. General language abilities of ChatMusician are evaluated on the [Massive Multitask Language Understanding (MMLU) dataset](https://huggingface.co/datasets/lukaemon/mmlu).
200
 
201
 
202