mgfrantz commited on
Commit
d87f480
1 Parent(s): 38cb89c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -0
README.md CHANGED
@@ -1,3 +1,32 @@
1
  ---
 
 
 
 
 
2
  license: mit
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - "en"
4
+ thumbnail: "https://styles.redditmedia.com/t5_2to41/styles/communityIcon_qedoavxzocr61.png?width=256&s=9c7c19b81474c3788279b8d6d6823e791d0524fc"
5
+ datasets:
6
+ - "reddit_tifu (subset: short)"
7
  license: mit
8
  ---
9
+
10
+ # mgfrantz/distilgpt2-finetuned-reddit-tifu
11
+
12
+ This model was trained to as practice for fine-tuning a causal language model.
13
+ There was no intended use case for this model besides having some fun seeing how different things might be screwed up.
14
+
15
+ ## Data
16
+
17
+ This model was trained on "short" subset of [`reddit_tifu`](https://huggingface.co/datasets/reddit_tifu) dataset.
18
+ The data was split into 90% train and 10% validation using `dataset.train_test_split`, with a seed of 0.
19
+
20
+ To prepare the data for training, the `"tldr"` and `"documents"` fields were joined by `"\n\n"`.
21
+ When multiple items were in the `"tldr"` or `"documents"` fields, only the first item was selected for joining.
22
+ These joined documents were tokenized using the `"distilgpt2"` tokenizer.
23
+
24
+ Finally, tokenized texts were concatenated end-to-end and split into blocks of 128 tokens.
25
+
26
+ **TODO:** Add a different separation token between documents that can be used to stop generation.
27
+
28
+ ## Training
29
+
30
+ This model was trained in Colab by fine-tuning [`distilgpt2`](https://huggingface.co/distilgpt2) for 174390 steps (3 epochs).
31
+ Default training arguments were used, except for `learning_rate=2e-5` and `weight_decay=0.01`.
32
+ At the conclusion of training, a training loss of 3.52 and a validation loss of 3.44 were observed.