abhishek HF staff commited on
Commit
03ba52c
1 Parent(s): 3062d47

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - image-classification
6
+ widget:
7
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
8
+ example_title: Tiger
9
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
10
+ example_title: Teapot
11
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
12
+ example_title: Palace
13
+ datasets:
14
+ - abhishek/autotrain-data-bxw3-9299-s4ce
15
+ ---
16
+
17
+ # Model Trained Using AutoTrain
18
+
19
+ - Problem type: Image Classification
20
+
21
+ ## Validation Metricsg
22
+ loss: 0.05566494166851044
23
+
24
+ f1_macro: 0.9822762607526727
25
+
26
+ f1_micro: 0.9818181818181818
27
+
28
+ f1_weighted: 0.9817521242226859
29
+
30
+ precision_macro: 0.9820894227603638
31
+
32
+ precision_micro: 0.9818181818181818
33
+
34
+ precision_weighted: 0.981850377976082
35
+
36
+ recall_macro: 0.9826135468293302
37
+
38
+ recall_micro: 0.9818181818181818
39
+
40
+ recall_weighted: 0.9818181818181818
41
+
42
+ accuracy: 0.9818181818181818
checkpoint-414/config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224",
3
+ "_num_labels": 5,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "daisy",
14
+ "1": "dandelion",
15
+ "2": "rose",
16
+ "3": "sunflower",
17
+ "4": "tulip"
18
+ },
19
+ "image_size": 224,
20
+ "initializer_range": 0.02,
21
+ "intermediate_size": 3072,
22
+ "label2id": {
23
+ "daisy": 0,
24
+ "dandelion": 1,
25
+ "rose": 2,
26
+ "sunflower": 3,
27
+ "tulip": 4
28
+ },
29
+ "layer_norm_eps": 1e-12,
30
+ "model_type": "vit",
31
+ "num_attention_heads": 12,
32
+ "num_channels": 3,
33
+ "num_hidden_layers": 12,
34
+ "patch_size": 16,
35
+ "problem_type": "single_label_classification",
36
+ "qkv_bias": true,
37
+ "torch_dtype": "float32",
38
+ "transformers_version": "4.35.2"
39
+ }
checkpoint-414/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ff39ca08220c5e2e9e17ebd0f9eb6be478de5c243ca0e990c4bf77a12479000
3
+ size 343233204
checkpoint-414/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ef6e7e23b10d8308dd47ca3e9d22b71bca80c3cf368046157b7a2c98b2fffed
3
+ size 686587322
checkpoint-414/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67710ddba44e5a4d45e04bc8cdda0559a2eee9946ee2fa882d5d147b68006421
3
+ size 14512
checkpoint-414/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd692783033de947a75ee26477cc7c712b2bb5361afd74eeb139795a3822f726
3
+ size 14512
checkpoint-414/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ba837861a61d574f101651884ef9685f2f78c0eec6c57d4e963a2642c4e3b94
3
+ size 1064
checkpoint-414/trainer_state.json ADDED
@@ -0,0 +1,259 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.05566494166851044,
3
+ "best_model_checkpoint": "output/bxw3-9299-s4ce-0/checkpoint-414",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 414,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.09,
13
+ "learning_rate": 1.3095238095238096e-05,
14
+ "loss": 1.6774,
15
+ "step": 13
16
+ },
17
+ {
18
+ "epoch": 0.19,
19
+ "learning_rate": 2.7380952380952383e-05,
20
+ "loss": 1.4563,
21
+ "step": 26
22
+ },
23
+ {
24
+ "epoch": 0.28,
25
+ "learning_rate": 4.2857142857142856e-05,
26
+ "loss": 0.8852,
27
+ "step": 39
28
+ },
29
+ {
30
+ "epoch": 0.38,
31
+ "learning_rate": 4.905913978494624e-05,
32
+ "loss": 0.4858,
33
+ "step": 52
34
+ },
35
+ {
36
+ "epoch": 0.47,
37
+ "learning_rate": 4.731182795698925e-05,
38
+ "loss": 0.3189,
39
+ "step": 65
40
+ },
41
+ {
42
+ "epoch": 0.57,
43
+ "learning_rate": 4.556451612903226e-05,
44
+ "loss": 0.1496,
45
+ "step": 78
46
+ },
47
+ {
48
+ "epoch": 0.66,
49
+ "learning_rate": 4.381720430107527e-05,
50
+ "loss": 0.2137,
51
+ "step": 91
52
+ },
53
+ {
54
+ "epoch": 0.75,
55
+ "learning_rate": 4.206989247311828e-05,
56
+ "loss": 0.1153,
57
+ "step": 104
58
+ },
59
+ {
60
+ "epoch": 0.85,
61
+ "learning_rate": 4.032258064516129e-05,
62
+ "loss": 0.1762,
63
+ "step": 117
64
+ },
65
+ {
66
+ "epoch": 0.94,
67
+ "learning_rate": 3.8575268817204305e-05,
68
+ "loss": 0.1755,
69
+ "step": 130
70
+ },
71
+ {
72
+ "epoch": 1.0,
73
+ "eval_accuracy": 0.9727272727272728,
74
+ "eval_f1_macro": 0.972396986523243,
75
+ "eval_f1_micro": 0.9727272727272728,
76
+ "eval_f1_weighted": 0.9726060520073766,
77
+ "eval_loss": 0.08732765913009644,
78
+ "eval_precision_macro": 0.9722370619929125,
79
+ "eval_precision_micro": 0.9727272727272728,
80
+ "eval_precision_weighted": 0.9728649069611348,
81
+ "eval_recall_macro": 0.9729742025670353,
82
+ "eval_recall_micro": 0.9727272727272728,
83
+ "eval_recall_weighted": 0.9727272727272728,
84
+ "eval_runtime": 1.0879,
85
+ "eval_samples_per_second": 505.581,
86
+ "eval_steps_per_second": 16.546,
87
+ "step": 138
88
+ },
89
+ {
90
+ "epoch": 1.04,
91
+ "learning_rate": 3.682795698924731e-05,
92
+ "loss": 0.1259,
93
+ "step": 143
94
+ },
95
+ {
96
+ "epoch": 1.13,
97
+ "learning_rate": 3.508064516129033e-05,
98
+ "loss": 0.1674,
99
+ "step": 156
100
+ },
101
+ {
102
+ "epoch": 1.22,
103
+ "learning_rate": 3.3333333333333335e-05,
104
+ "loss": 0.1101,
105
+ "step": 169
106
+ },
107
+ {
108
+ "epoch": 1.32,
109
+ "learning_rate": 3.158602150537634e-05,
110
+ "loss": 0.1049,
111
+ "step": 182
112
+ },
113
+ {
114
+ "epoch": 1.41,
115
+ "learning_rate": 2.9838709677419357e-05,
116
+ "loss": 0.0704,
117
+ "step": 195
118
+ },
119
+ {
120
+ "epoch": 1.51,
121
+ "learning_rate": 2.8091397849462365e-05,
122
+ "loss": 0.0785,
123
+ "step": 208
124
+ },
125
+ {
126
+ "epoch": 1.6,
127
+ "learning_rate": 2.6344086021505376e-05,
128
+ "loss": 0.0617,
129
+ "step": 221
130
+ },
131
+ {
132
+ "epoch": 1.7,
133
+ "learning_rate": 2.4596774193548387e-05,
134
+ "loss": 0.1625,
135
+ "step": 234
136
+ },
137
+ {
138
+ "epoch": 1.79,
139
+ "learning_rate": 2.28494623655914e-05,
140
+ "loss": 0.0238,
141
+ "step": 247
142
+ },
143
+ {
144
+ "epoch": 1.88,
145
+ "learning_rate": 2.110215053763441e-05,
146
+ "loss": 0.0708,
147
+ "step": 260
148
+ },
149
+ {
150
+ "epoch": 1.98,
151
+ "learning_rate": 1.935483870967742e-05,
152
+ "loss": 0.0856,
153
+ "step": 273
154
+ },
155
+ {
156
+ "epoch": 2.0,
157
+ "eval_accuracy": 0.9818181818181818,
158
+ "eval_f1_macro": 0.9820711118316969,
159
+ "eval_f1_micro": 0.9818181818181818,
160
+ "eval_f1_weighted": 0.9817660734040653,
161
+ "eval_loss": 0.0563533715903759,
162
+ "eval_precision_macro": 0.9816901091597003,
163
+ "eval_precision_micro": 0.9818181818181818,
164
+ "eval_precision_weighted": 0.9818836715959347,
165
+ "eval_recall_macro": 0.9826135468293302,
166
+ "eval_recall_micro": 0.9818181818181818,
167
+ "eval_recall_weighted": 0.9818181818181818,
168
+ "eval_runtime": 1.0889,
169
+ "eval_samples_per_second": 505.086,
170
+ "eval_steps_per_second": 16.53,
171
+ "step": 276
172
+ },
173
+ {
174
+ "epoch": 2.07,
175
+ "learning_rate": 1.7607526881720432e-05,
176
+ "loss": 0.0957,
177
+ "step": 286
178
+ },
179
+ {
180
+ "epoch": 2.17,
181
+ "learning_rate": 1.586021505376344e-05,
182
+ "loss": 0.0648,
183
+ "step": 299
184
+ },
185
+ {
186
+ "epoch": 2.26,
187
+ "learning_rate": 1.4112903225806454e-05,
188
+ "loss": 0.1037,
189
+ "step": 312
190
+ },
191
+ {
192
+ "epoch": 2.36,
193
+ "learning_rate": 1.2365591397849464e-05,
194
+ "loss": 0.0823,
195
+ "step": 325
196
+ },
197
+ {
198
+ "epoch": 2.45,
199
+ "learning_rate": 1.0618279569892473e-05,
200
+ "loss": 0.1222,
201
+ "step": 338
202
+ },
203
+ {
204
+ "epoch": 2.54,
205
+ "learning_rate": 8.870967741935484e-06,
206
+ "loss": 0.0418,
207
+ "step": 351
208
+ },
209
+ {
210
+ "epoch": 2.64,
211
+ "learning_rate": 7.1236559139784956e-06,
212
+ "loss": 0.0513,
213
+ "step": 364
214
+ },
215
+ {
216
+ "epoch": 2.73,
217
+ "learning_rate": 5.376344086021506e-06,
218
+ "loss": 0.0941,
219
+ "step": 377
220
+ },
221
+ {
222
+ "epoch": 2.83,
223
+ "learning_rate": 3.6290322580645166e-06,
224
+ "loss": 0.023,
225
+ "step": 390
226
+ },
227
+ {
228
+ "epoch": 2.92,
229
+ "learning_rate": 1.881720430107527e-06,
230
+ "loss": 0.0559,
231
+ "step": 403
232
+ },
233
+ {
234
+ "epoch": 3.0,
235
+ "eval_accuracy": 0.9818181818181818,
236
+ "eval_f1_macro": 0.9822762607526727,
237
+ "eval_f1_micro": 0.9818181818181818,
238
+ "eval_f1_weighted": 0.9817521242226859,
239
+ "eval_loss": 0.05566494166851044,
240
+ "eval_precision_macro": 0.9820894227603638,
241
+ "eval_precision_micro": 0.9818181818181818,
242
+ "eval_precision_weighted": 0.981850377976082,
243
+ "eval_recall_macro": 0.9826135468293302,
244
+ "eval_recall_micro": 0.9818181818181818,
245
+ "eval_recall_weighted": 0.9818181818181818,
246
+ "eval_runtime": 1.0922,
247
+ "eval_samples_per_second": 503.548,
248
+ "eval_steps_per_second": 16.48,
249
+ "step": 414
250
+ }
251
+ ],
252
+ "logging_steps": 13,
253
+ "max_steps": 414,
254
+ "num_train_epochs": 3,
255
+ "save_steps": 500,
256
+ "total_flos": 5.133207467857019e+17,
257
+ "trial_name": null,
258
+ "trial_params": null
259
+ }
checkpoint-414/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cddb551093ed3c10aeb207b87f813b50be28d3b8a4fe3fb659b7d8946e621e33
3
+ size 4536
config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224",
3
+ "_num_labels": 5,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "daisy",
14
+ "1": "dandelion",
15
+ "2": "rose",
16
+ "3": "sunflower",
17
+ "4": "tulip"
18
+ },
19
+ "image_size": 224,
20
+ "initializer_range": 0.02,
21
+ "intermediate_size": 3072,
22
+ "label2id": {
23
+ "daisy": 0,
24
+ "dandelion": 1,
25
+ "rose": 2,
26
+ "sunflower": 3,
27
+ "tulip": 4
28
+ },
29
+ "layer_norm_eps": 1e-12,
30
+ "model_type": "vit",
31
+ "num_attention_heads": 12,
32
+ "num_channels": 3,
33
+ "num_hidden_layers": 12,
34
+ "patch_size": 16,
35
+ "problem_type": "single_label_classification",
36
+ "qkv_bias": true,
37
+ "torch_dtype": "float32",
38
+ "transformers_version": "4.35.2"
39
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ff39ca08220c5e2e9e17ebd0f9eb6be478de5c243ca0e990c4bf77a12479000
3
+ size 343233204
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cddb551093ed3c10aeb207b87f813b50be28d3b8a4fe3fb659b7d8946e621e33
3
+ size 4536
training_params.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "abhishek/autotrain-data-bxw3-9299-s4ce",
3
+ "model": "google/vit-base-patch16-224",
4
+ "lr": 5e-05,
5
+ "epochs": 3,
6
+ "batch_size": 8,
7
+ "warmup_ratio": 0.1,
8
+ "gradient_accumulation": 1,
9
+ "optimizer": "adamw_torch",
10
+ "scheduler": "linear",
11
+ "weight_decay": 0.0,
12
+ "max_grad_norm": 1.0,
13
+ "seed": 42,
14
+ "train_split": "train",
15
+ "valid_split": "validation",
16
+ "logging_steps": -1,
17
+ "project_name": "output/bxw3-9299-s4ce-0",
18
+ "auto_find_batch_size": false,
19
+ "mixed_precision": "fp16",
20
+ "save_total_limit": 1,
21
+ "save_strategy": "epoch",
22
+ "push_to_hub": true,
23
+ "repo_id": "abhishek/bxw3-9299-s4ce-0",
24
+ "evaluation_strategy": "epoch",
25
+ "image_column": "autotrain_image",
26
+ "target_column": "autotrain_label",
27
+ "log": "none"
28
+ }