Update README.md
Browse files
README.md
CHANGED
@@ -90,8 +90,6 @@ dump_wav('cm_music_piece.wav', audio, sample_rate=44100, use_int16=True)
|
|
90 |
|
91 |
## Example Prompts
|
92 |
|
93 |
-
**Our model produces symbolic music(ABC notation) well in the following prompts.** Here are some musical tasks.
|
94 |
-
|
95 |
### Function: Chord Conditioned Music Generation
|
96 |
```
|
97 |
Develop a musical piece using the given chord progression.
|
@@ -193,18 +191,12 @@ ChatMusician is pretrained on the 🤗 [MusicPile](https://huggingface.co/datase
|
|
193 |
And supervised finetuned on 1.1M samples(2:1 ratio between music scores
|
194 |
and music knowledge & music summary data) from MusicPile. Check our [paper](http://arxiv.org/abs/2402.16153) for more details.
|
195 |
|
196 |
-
## Training Procedure
|
197 |
-
|
198 |
-
We initialized a fp16-precision ChatMusician-Base from the LLaMA2-7B-Base weights, and applied a continual pre-training plus fine-tuning pipeline. LoRA adapters were integrated into the attention and MLP layers, with additional training on embeddings and all linear layers. The maximum sequence length
|
199 |
-
was 2048. We utilized 16 80GB-A800 GPUs for one epoch pre-training and 8 32GB-V100 GPUs for two epoch fine-tuning. DeepSpeed was employed for memory efficiency, and the AdamW optimizer was used with a 1e-4 learning rate and a 5% warmup cosine scheduler. Gradient clipping was set at 1.0. The LoRA parameters dimension, alpha, and
|
200 |
-
dropout were set to 64, 16, and 0.1, with a batch size of 8.
|
201 |
-
|
202 |
|
203 |
## Evaluation
|
204 |
|
205 |
1. Music understanding abilities are evaluated on the [MusicTheoryBench](https://huggingface.co/datasets/m-a-p/MusicTheoryBench). The following figure is zero-shot accuracy on MusicTheoryBench.
|
206 |
We included GPT-3.5, GPT-4, LLaMA2-7B-Base, ChatMusician-Base, and ChatMusician. The blue bar represents the performance on the music knowledge metric, and the red bar represents the music reasoning metric. The dashed line corresponds to a random baseline, with a score of 25%.![MusicTheoryBench_result](./MusicTheoryBench_result_plt.png)
|
207 |
-
2. General language abilities of ChatMusician are evaluated
|
208 |
|
209 |
|
210 |
|
|
|
90 |
|
91 |
## Example Prompts
|
92 |
|
|
|
|
|
93 |
### Function: Chord Conditioned Music Generation
|
94 |
```
|
95 |
Develop a musical piece using the given chord progression.
|
|
|
191 |
And supervised finetuned on 1.1M samples(2:1 ratio between music scores
|
192 |
and music knowledge & music summary data) from MusicPile. Check our [paper](http://arxiv.org/abs/2402.16153) for more details.
|
193 |
|
|
|
|
|
|
|
|
|
|
|
|
|
194 |
|
195 |
## Evaluation
|
196 |
|
197 |
1. Music understanding abilities are evaluated on the [MusicTheoryBench](https://huggingface.co/datasets/m-a-p/MusicTheoryBench). The following figure is zero-shot accuracy on MusicTheoryBench.
|
198 |
We included GPT-3.5, GPT-4, LLaMA2-7B-Base, ChatMusician-Base, and ChatMusician. The blue bar represents the performance on the music knowledge metric, and the red bar represents the music reasoning metric. The dashed line corresponds to a random baseline, with a score of 25%.![MusicTheoryBench_result](./MusicTheoryBench_result_plt.png)
|
199 |
+
2. General language abilities of ChatMusician are evaluated on the [Massive Multitask Language Understanding (MMLU) dataset](https://huggingface.co/datasets/lukaemon/mmlu).
|
200 |
|
201 |
|
202 |
|