imelike commited on
Commit
862b2b0
1 Parent(s): a71b95a

Training in progress epoch 0

Browse files
Files changed (3) hide show
  1. README.md +3 -3
  2. tf_model.h5 +1 -1
  3. tokenizer_config.json +4 -0
README.md CHANGED
@@ -15,8 +15,8 @@ probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 5.6114
19
- - Validation Loss: 5.4982
20
  - Epoch: 0
21
 
22
  ## Model description
@@ -43,7 +43,7 @@ The following hyperparameters were used during training:
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
- | 5.6114 | 5.4982 | 0 |
47
 
48
 
49
  ### Framework versions
 
15
 
16
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Train Loss: 6.2977
19
+ - Validation Loss: 5.4911
20
  - Epoch: 0
21
 
22
  ## Model description
 
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
+ | 6.2977 | 5.4911 | 0 |
47
 
48
 
49
  ### Framework versions
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32336fea53fd9cf3b2df9ec38433aec94f911679a1c9baca437e16dc52dcf141
3
  size 503289936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:736640d570117172a8829fff44475c03e33d8d931815e91960e358ef18540eb7
3
  size 503289936
tokenizer_config.json CHANGED
@@ -15,8 +15,12 @@
15
  "clean_up_tokenization_spaces": true,
16
  "eos_token": "<|endoftext|>",
17
  "errors": "replace",
 
18
  "model_max_length": 1024,
19
  "pad_token": "<|endoftext|>",
 
20
  "tokenizer_class": "GPT2Tokenizer",
 
 
21
  "unk_token": "<|endoftext|>"
22
  }
 
15
  "clean_up_tokenization_spaces": true,
16
  "eos_token": "<|endoftext|>",
17
  "errors": "replace",
18
+ "max_length": 40,
19
  "model_max_length": 1024,
20
  "pad_token": "<|endoftext|>",
21
+ "stride": 0,
22
  "tokenizer_class": "GPT2Tokenizer",
23
+ "truncation_side": "right",
24
+ "truncation_strategy": "longest_first",
25
  "unk_token": "<|endoftext|>"
26
  }