AlekseyKorshuk commited on
Commit
ad9a638
1 Parent(s): 03fc98d

huggingartists

Browse files
README.md CHANGED
@@ -45,15 +45,15 @@ from datasets import load_dataset
45
  dataset = load_dataset("huggingartists/eminem")
46
  ```
47
 
48
- [Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/rzzk4mti/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline.
49
 
50
  ## Training procedure
51
 
52
  The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on Eminem's lyrics.
53
 
54
- Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/3goym9q5) for full transparency and reproducibility.
55
 
56
- At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/3goym9q5/artifacts) is logged and versioned.
57
 
58
  ## How to use
59
 
 
45
  dataset = load_dataset("huggingartists/eminem")
46
  ```
47
 
48
+ [Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/i4fhzbrj/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline.
49
 
50
  ## Training procedure
51
 
52
  The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on Eminem's lyrics.
53
 
54
+ Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/1poi2h7i) for full transparency and reproducibility.
55
 
56
+ At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/1poi2h7i/artifacts) is logged and versioned.
57
 
58
  ## How to use
59
 
config.json CHANGED
@@ -37,7 +37,7 @@
37
  }
38
  },
39
  "torch_dtype": "float32",
40
- "transformers_version": "4.16.2",
41
  "use_cache": true,
42
  "vocab_size": 50257
43
  }
 
37
  }
38
  },
39
  "torch_dtype": "float32",
40
+ "transformers_version": "4.17.0",
41
  "use_cache": true,
42
  "vocab_size": 50257
43
  }
evaluation.txt CHANGED
@@ -1 +1 @@
1
- {"eval_loss": 0.2587912380695343, "eval_runtime": 28.9992, "eval_samples_per_second": 21.207, "eval_steps_per_second": 2.655, "epoch": 5.0}
 
1
+ {"eval_loss": 0.21638885140419006, "eval_runtime": 7.4926, "eval_samples_per_second": 75.808, "eval_steps_per_second": 9.476, "epoch": 16.0}
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6502370464fa8cd462a1120ff3ee1dec9d5cb90a566bf81d4ec4ec2871aee3ee
3
  size 497764120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b53cf55c5742dfd44f30083b86ae36459beabcbd11195bdaac41d1187bf6903
3
  size 497764120
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea78d837462bd33d0b6f54deeb72fa1e3764e83693c5898b20be034af45e293d
3
  size 995604017
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dbdcbfbbed9454138a13e687af3e4629daf4c9011020de0b8620c5336aaafe5
3
  size 995604017
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb985697bef42b35a4af505c6b0feac38014cf1700a081b2fee7115c3ec54778
3
- size 510403817
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aca2cc621206c043e63592cae95f51494033a0b1b40cab5e23ba5095d51fbf90
3
+ size 510404393
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7de0ea18d422a4a904e0e321d0648ef7e938ccc8b5f8f3cda6bbcafb21a6903
3
  size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e8aeaafb3f4610c17ece97e4f8b321dfc495ac314a95aa40c9bfe6baff30177
3
  size 14567
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da390c8a01b24385c79a732c429de5ff0108495a6ebece5eee2c39f888d039ff
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf464022adc7eb0c26cafda89e5bc206138120150f9a55ed481cde8b86e0da26
3
  size 623
trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.2587912380695343,
3
- "best_model_checkpoint": "output/eminem/checkpoint-458",
4
  "epoch": 1.0,
5
- "global_step": 458,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -560,11 +560,25 @@
560
  "eval_samples_per_second": 21.243,
561
  "eval_steps_per_second": 2.66,
562
  "step": 458
 
 
 
 
 
 
 
 
 
 
 
 
 
 
563
  }
564
  ],
565
- "max_steps": 2290,
566
- "num_train_epochs": 5,
567
- "total_flos": 478425710592000.0,
568
  "trial_name": null,
569
  "trial_params": null
570
  }
 
1
  {
2
+ "best_metric": 0.21638885140419006,
3
+ "best_model_checkpoint": "output/eminem/checkpoint-464",
4
  "epoch": 1.0,
5
+ "global_step": 464,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
560
  "eval_samples_per_second": 21.243,
561
  "eval_steps_per_second": 2.66,
562
  "step": 458
563
+ },
564
+ {
565
+ "epoch": 0.99,
566
+ "learning_rate": 8.656213912947592e-06,
567
+ "loss": 0.6179,
568
+ "step": 460
569
+ },
570
+ {
571
+ "epoch": 1.0,
572
+ "eval_loss": 0.21638885140419006,
573
+ "eval_runtime": 7.4743,
574
+ "eval_samples_per_second": 75.994,
575
+ "eval_steps_per_second": 9.499,
576
+ "step": 464
577
  }
578
  ],
579
+ "max_steps": 7424,
580
+ "num_train_epochs": 16,
581
+ "total_flos": 484304781312000.0,
582
  "trial_name": null,
583
  "trial_params": null
584
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac7a77ffb86041404ae25e05b2064145909205e5d64035e10a124f7b561b78a7
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f57a3894986a975d80afe4e8230dbc97b7c7cf3a77bfec1e24d21961b8e53884
3
  size 3055