talargv commited on
Commit
872b96f
1 Parent(s): beee6bb

Model save

Browse files
Files changed (2) hide show
  1. README.md +6 -7
  2. generation_config.json +10 -0
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
- base_model: facebook/musicgen-large
3
- library_name: peft
4
  license: cc-by-nc-4.0
 
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -14,7 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # musicgen-finetune-aav
16
 
17
- This model is a fine-tuned version of [facebook/musicgen-large](https://huggingface.co/facebook/musicgen-large) on an unknown dataset.
18
 
19
  ## Model description
20
 
@@ -33,7 +33,7 @@ More information needed
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
- - learning_rate: 0.0002
37
  - train_batch_size: 2
38
  - eval_batch_size: 8
39
  - seed: 42
@@ -41,7 +41,7 @@ The following hyperparameters were used during training:
41
  - total_train_batch_size: 16
42
  - optimizer: Use adamw_torch with betas=(0.9,0.99) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
- - num_epochs: 2
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
@@ -50,8 +50,7 @@ The following hyperparameters were used during training:
50
 
51
  ### Framework versions
52
 
53
- - PEFT 0.13.2
54
  - Transformers 4.47.0.dev0
55
  - Pytorch 2.1.2+cu121
56
  - Datasets 3.1.0
57
- - Tokenizers 0.20.1
 
1
  ---
2
+ library_name: transformers
 
3
  license: cc-by-nc-4.0
4
+ base_model: facebook/musicgen-small
5
  tags:
6
  - generated_from_trainer
7
  model-index:
 
14
 
15
  # musicgen-finetune-aav
16
 
17
+ This model is a fine-tuned version of [facebook/musicgen-small](https://huggingface.co/facebook/musicgen-small) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
+ - learning_rate: 0.0001
37
  - train_batch_size: 2
38
  - eval_batch_size: 8
39
  - seed: 42
 
41
  - total_train_batch_size: 16
42
  - optimizer: Use adamw_torch with betas=(0.9,0.99) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
+ - num_epochs: 4
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
 
50
 
51
  ### Framework versions
52
 
 
53
  - Transformers 4.47.0.dev0
54
  - Pytorch 2.1.2+cu121
55
  - Datasets 3.1.0
56
+ - Tokenizers 0.20.2
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 2048,
4
+ "decoder_start_token_id": 2048,
5
+ "do_sample": true,
6
+ "guidance_scale": 3.0,
7
+ "max_length": 1500,
8
+ "pad_token_id": 2048,
9
+ "transformers_version": "4.47.0.dev0"
10
+ }