w05230505 commited on
Commit
4210e4e
1 Parent(s): df212f1

End of training

Browse files
Files changed (6) hide show
  1. README.md +24 -6
  2. all_results.json +10 -10
  3. config.json +1 -0
  4. eval_results.json +7 -7
  5. train_results.json +3 -3
  6. trainer_state.json +3 -3
README.md CHANGED
@@ -1,15 +1,33 @@
1
  ---
2
  library_name: transformers
 
 
3
  license: apache-2.0
4
  base_model: bert-base-uncased
5
  tags:
6
  - generated_from_trainer
 
 
7
  metrics:
8
  - accuracy
9
  - f1
10
  model-index:
11
  - name: bert-base-uncased-finetuned-mrpc
12
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -17,12 +35,12 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # bert-base-uncased-finetuned-mrpc
19
 
20
- This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Accuracy: 0.8529
23
- - Combined Score: 0.8744
24
- - F1: 0.8958
25
- - Loss: 0.4322
26
 
27
  ## Model description
28
 
 
1
  ---
2
  library_name: transformers
3
+ language:
4
+ - en
5
  license: apache-2.0
6
  base_model: bert-base-uncased
7
  tags:
8
  - generated_from_trainer
9
+ datasets:
10
+ - glue
11
  metrics:
12
  - accuracy
13
  - f1
14
  model-index:
15
  - name: bert-base-uncased-finetuned-mrpc
16
+ results:
17
+ - task:
18
+ name: Text Classification
19
+ type: text-classification
20
+ dataset:
21
+ name: GLUE MRPC
22
+ type: glue
23
+ args: mrpc
24
+ metrics:
25
+ - name: Accuracy
26
+ type: accuracy
27
+ value: 0.8480392156862745
28
+ - name: F1
29
+ type: f1
30
+ value: 0.8923611111111112
31
  ---
32
 
33
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
35
 
36
  # bert-base-uncased-finetuned-mrpc
37
 
38
+ This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the GLUE MRPC dataset.
39
  It achieves the following results on the evaluation set:
40
+ - Loss: 0.4198
41
+ - Accuracy: 0.8480
42
+ - F1: 0.8924
43
+ - Combined Score: 0.8702
44
 
45
  ## Model description
46
 
all_results.json CHANGED
@@ -1,18 +1,18 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_accuracy": 0.8529411764705882,
4
- "eval_combined_score": 0.8743872549019608,
5
- "eval_f1": 0.8958333333333334,
6
- "eval_loss": 0.4279242157936096,
7
  "eval_model_preparation_time": 0.0025,
8
- "eval_runtime": 60.4806,
9
  "eval_samples": 408,
10
- "eval_samples_per_second": 6.746,
11
- "eval_steps_per_second": 0.843,
12
  "total_flos": 2895274053181440.0,
13
  "train_loss": 0.0,
14
- "train_runtime": 0.0014,
15
  "train_samples": 3668,
16
- "train_samples_per_second": 7792355.431,
17
- "train_steps_per_second": 488615.526
18
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_accuracy": 0.8480392156862745,
4
+ "eval_combined_score": 0.8702001633986929,
5
+ "eval_f1": 0.8923611111111112,
6
+ "eval_loss": 0.4198153018951416,
7
  "eval_model_preparation_time": 0.0025,
8
+ "eval_runtime": 60.8121,
9
  "eval_samples": 408,
10
+ "eval_samples_per_second": 6.709,
11
+ "eval_steps_per_second": 0.839,
12
  "total_flos": 2895274053181440.0,
13
  "train_loss": 0.0,
14
+ "train_runtime": 0.0013,
15
  "train_samples": 3668,
16
+ "train_samples_per_second": 8422284.893,
17
+ "train_steps_per_second": 528114.92
18
  }
config.json CHANGED
@@ -23,6 +23,7 @@
23
  "num_hidden_layers": 12,
24
  "pad_token_id": 0,
25
  "position_embedding_type": "absolute",
 
26
  "torch_dtype": "float32",
27
  "transformers_version": "4.45.0.dev0",
28
  "type_vocab_size": 2,
 
23
  "num_hidden_layers": 12,
24
  "pad_token_id": 0,
25
  "position_embedding_type": "absolute",
26
+ "problem_type": "single_label_classification",
27
  "torch_dtype": "float32",
28
  "transformers_version": "4.45.0.dev0",
29
  "type_vocab_size": 2,
eval_results.json CHANGED
@@ -1,11 +1,11 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_accuracy": 0.8529411764705882,
4
- "eval_combined_score": 0.8743872549019608,
5
- "eval_f1": 0.8958333333333334,
6
- "eval_loss": 0.4279242157936096,
7
- "eval_runtime": 60.4806,
8
  "eval_samples": 408,
9
- "eval_samples_per_second": 6.746,
10
- "eval_steps_per_second": 0.843
11
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_accuracy": 0.8480392156862745,
4
+ "eval_combined_score": 0.8702001633986929,
5
+ "eval_f1": 0.8923611111111112,
6
+ "eval_loss": 0.4198153018951416,
7
+ "eval_runtime": 60.8121,
8
  "eval_samples": 408,
9
+ "eval_samples_per_second": 6.709,
10
+ "eval_steps_per_second": 0.839
11
  }
train_results.json CHANGED
@@ -2,8 +2,8 @@
2
  "epoch": 3.0,
3
  "total_flos": 2895274053181440.0,
4
  "train_loss": 0.0,
5
- "train_runtime": 0.0014,
6
  "train_samples": 3668,
7
- "train_samples_per_second": 7792355.431,
8
- "train_steps_per_second": 488615.526
9
  }
 
2
  "epoch": 3.0,
3
  "total_flos": 2895274053181440.0,
4
  "train_loss": 0.0,
5
+ "train_runtime": 0.0013,
6
  "train_samples": 3668,
7
+ "train_samples_per_second": 8422284.893,
8
+ "train_steps_per_second": 528114.92
9
  }
trainer_state.json CHANGED
@@ -67,9 +67,9 @@
67
  "step": 690,
68
  "total_flos": 2895274053181440.0,
69
  "train_loss": 0.0,
70
- "train_runtime": 0.0014,
71
- "train_samples_per_second": 7792355.431,
72
- "train_steps_per_second": 488615.526
73
  }
74
  ],
75
  "logging_steps": 500,
 
67
  "step": 690,
68
  "total_flos": 2895274053181440.0,
69
  "train_loss": 0.0,
70
+ "train_runtime": 0.0013,
71
+ "train_samples_per_second": 8422284.893,
72
+ "train_steps_per_second": 528114.92
73
  }
74
  ],
75
  "logging_steps": 500,