tzhao3 commited on
Commit
91ed8a6
1 Parent(s): 1f26bb1

Upload 8 files

Browse files
config.json ADDED
@@ -0,0 +1,228 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "apple",
13
+ "1": "aquarium_fish",
14
+ "2": "baby",
15
+ "3": "bear",
16
+ "4": "beaver",
17
+ "5": "bed",
18
+ "6": "bee",
19
+ "7": "beetle",
20
+ "8": "bicycle",
21
+ "9": "bottle",
22
+ "10": "bowl",
23
+ "11": "boy",
24
+ "12": "bridge",
25
+ "13": "bus",
26
+ "14": "butterfly",
27
+ "15": "camel",
28
+ "16": "can",
29
+ "17": "castle",
30
+ "18": "caterpillar",
31
+ "19": "cattle",
32
+ "20": "chair",
33
+ "21": "chimpanzee",
34
+ "22": "clock",
35
+ "23": "cloud",
36
+ "24": "cockroach",
37
+ "25": "couch",
38
+ "26": "cra",
39
+ "27": "crocodile",
40
+ "28": "cup",
41
+ "29": "dinosaur",
42
+ "30": "dolphin",
43
+ "31": "elephant",
44
+ "32": "flatfish",
45
+ "33": "forest",
46
+ "34": "fox",
47
+ "35": "girl",
48
+ "36": "hamster",
49
+ "37": "house",
50
+ "38": "kangaroo",
51
+ "39": "keyboard",
52
+ "40": "lamp",
53
+ "41": "lawn_mower",
54
+ "42": "leopard",
55
+ "43": "lion",
56
+ "44": "lizard",
57
+ "45": "lobster",
58
+ "46": "man",
59
+ "47": "maple_tree",
60
+ "48": "motorcycle",
61
+ "49": "mountain",
62
+ "50": "mouse",
63
+ "51": "mushroom",
64
+ "52": "oak_tree",
65
+ "53": "orange",
66
+ "54": "orchid",
67
+ "55": "otter",
68
+ "56": "palm_tree",
69
+ "57": "pear",
70
+ "58": "pickup_truck",
71
+ "59": "pine_tree",
72
+ "60": "plain",
73
+ "61": "plate",
74
+ "62": "poppy",
75
+ "63": "porcupine",
76
+ "64": "possum",
77
+ "65": "rabbit",
78
+ "66": "raccoon",
79
+ "67": "ray",
80
+ "68": "road",
81
+ "69": "rocket",
82
+ "70": "rose",
83
+ "71": "sea",
84
+ "72": "seal",
85
+ "73": "shark",
86
+ "74": "shrew",
87
+ "75": "skunk",
88
+ "76": "skyscraper",
89
+ "77": "snail",
90
+ "78": "snake",
91
+ "79": "spider",
92
+ "80": "squirrel",
93
+ "81": "streetcar",
94
+ "82": "sunflower",
95
+ "83": "sweet_pepper",
96
+ "84": "table",
97
+ "85": "tank",
98
+ "86": "telephone",
99
+ "87": "television",
100
+ "88": "tiger",
101
+ "89": "tractor",
102
+ "90": "train",
103
+ "91": "trout",
104
+ "92": "tulip",
105
+ "93": "turtle",
106
+ "94": "wardrobe",
107
+ "95": "whale",
108
+ "96": "willow_tree",
109
+ "97": "wolf",
110
+ "98": "woman",
111
+ "99": "worm"
112
+ },
113
+ "image_size": 224,
114
+ "initializer_range": 0.02,
115
+ "intermediate_size": 3072,
116
+ "label2id": {
117
+ "apple": 0,
118
+ "aquarium_fish": 1,
119
+ "baby": 2,
120
+ "bear": 3,
121
+ "beaver": 4,
122
+ "bed": 5,
123
+ "bee": 6,
124
+ "beetle": 7,
125
+ "bicycle": 8,
126
+ "bottle": 9,
127
+ "bowl": 10,
128
+ "boy": 11,
129
+ "bridge": 12,
130
+ "bus": 13,
131
+ "butterfly": 14,
132
+ "camel": 15,
133
+ "can": 16,
134
+ "castle": 17,
135
+ "caterpillar": 18,
136
+ "cattle": 19,
137
+ "chair": 20,
138
+ "chimpanzee": 21,
139
+ "clock": 22,
140
+ "cloud": 23,
141
+ "cockroach": 24,
142
+ "couch": 25,
143
+ "cra": 26,
144
+ "crocodile": 27,
145
+ "cup": 28,
146
+ "dinosaur": 29,
147
+ "dolphin": 30,
148
+ "elephant": 31,
149
+ "flatfish": 32,
150
+ "forest": 33,
151
+ "fox": 34,
152
+ "girl": 35,
153
+ "hamster": 36,
154
+ "house": 37,
155
+ "kangaroo": 38,
156
+ "keyboard": 39,
157
+ "lamp": 40,
158
+ "lawn_mower": 41,
159
+ "leopard": 42,
160
+ "lion": 43,
161
+ "lizard": 44,
162
+ "lobster": 45,
163
+ "man": 46,
164
+ "maple_tree": 47,
165
+ "motorcycle": 48,
166
+ "mountain": 49,
167
+ "mouse": 50,
168
+ "mushroom": 51,
169
+ "oak_tree": 52,
170
+ "orange": 53,
171
+ "orchid": 54,
172
+ "otter": 55,
173
+ "palm_tree": 56,
174
+ "pear": 57,
175
+ "pickup_truck": 58,
176
+ "pine_tree": 59,
177
+ "plain": 60,
178
+ "plate": 61,
179
+ "poppy": 62,
180
+ "porcupine": 63,
181
+ "possum": 64,
182
+ "rabbit": 65,
183
+ "raccoon": 66,
184
+ "ray": 67,
185
+ "road": 68,
186
+ "rocket": 69,
187
+ "rose": 70,
188
+ "sea": 71,
189
+ "seal": 72,
190
+ "shark": 73,
191
+ "shrew": 74,
192
+ "skunk": 75,
193
+ "skyscraper": 76,
194
+ "snail": 77,
195
+ "snake": 78,
196
+ "spider": 79,
197
+ "squirrel": 80,
198
+ "streetcar": 81,
199
+ "sunflower": 82,
200
+ "sweet_pepper": 83,
201
+ "table": 84,
202
+ "tank": 85,
203
+ "telephone": 86,
204
+ "television": 87,
205
+ "tiger": 88,
206
+ "tractor": 89,
207
+ "train": 90,
208
+ "trout": 91,
209
+ "tulip": 92,
210
+ "turtle": 93,
211
+ "wardrobe": 94,
212
+ "whale": 95,
213
+ "willow_tree": 96,
214
+ "wolf": 97,
215
+ "woman": 98,
216
+ "worm": 99
217
+ },
218
+ "layer_norm_eps": 1e-12,
219
+ "model_type": "vit",
220
+ "num_attention_heads": 12,
221
+ "num_channels": 3,
222
+ "num_hidden_layers": 12,
223
+ "patch_size": 16,
224
+ "problem_type": "single_label_classification",
225
+ "qkv_bias": true,
226
+ "torch_dtype": "float32",
227
+ "transformers_version": "4.29.2"
228
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cec63a72d55fe2d19bb452a5e4501ceadd435316fec4792753ab880c8ee39cf
3
+ size 687121797
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d79ee5b99d5a21206df139f6dcfc731cc9d97147459238647d41389b2c20f9c
3
+ size 343570157
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82bdb3348630fd0c85da6a17901be79efaea14773a56d2c79fe5b0074f9069a1
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:371420dc8d96d1664bdfe0e2f08c8877254f3b5eab59c06c0025aae16e65b70b
3
+ size 627
trainer_state.json ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.87,
3
+ "best_model_checkpoint": "fine-tune-vit-cifar100\\checkpoint-2816",
4
+ "epoch": 5.0,
5
+ "global_step": 3520,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.71,
12
+ "learning_rate": 0.00019715909090909094,
13
+ "loss": 2.6561,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 1.0,
18
+ "eval_accuracy": 0.824,
19
+ "eval_loss": 0.7328090071678162,
20
+ "eval_runtime": 24.9237,
21
+ "eval_samples_per_second": 200.612,
22
+ "eval_steps_per_second": 3.17,
23
+ "step": 704
24
+ },
25
+ {
26
+ "epoch": 1.42,
27
+ "learning_rate": 0.0001943181818181818,
28
+ "loss": 1.2132,
29
+ "step": 1000
30
+ },
31
+ {
32
+ "epoch": 2.0,
33
+ "eval_accuracy": 0.8446,
34
+ "eval_loss": 0.5557541251182556,
35
+ "eval_runtime": 24.0685,
36
+ "eval_samples_per_second": 207.74,
37
+ "eval_steps_per_second": 3.282,
38
+ "step": 1408
39
+ },
40
+ {
41
+ "epoch": 2.13,
42
+ "learning_rate": 0.00019147727272727274,
43
+ "loss": 1.0151,
44
+ "step": 1500
45
+ },
46
+ {
47
+ "epoch": 2.84,
48
+ "learning_rate": 0.00018863636363636364,
49
+ "loss": 0.9014,
50
+ "step": 2000
51
+ },
52
+ {
53
+ "epoch": 3.0,
54
+ "eval_accuracy": 0.8508,
55
+ "eval_loss": 0.5289514660835266,
56
+ "eval_runtime": 23.869,
57
+ "eval_samples_per_second": 209.477,
58
+ "eval_steps_per_second": 3.31,
59
+ "step": 2112
60
+ },
61
+ {
62
+ "epoch": 3.55,
63
+ "learning_rate": 0.00018579545454545454,
64
+ "loss": 0.8114,
65
+ "step": 2500
66
+ },
67
+ {
68
+ "epoch": 4.0,
69
+ "eval_accuracy": 0.87,
70
+ "eval_loss": 0.476001501083374,
71
+ "eval_runtime": 24.1421,
72
+ "eval_samples_per_second": 207.107,
73
+ "eval_steps_per_second": 3.272,
74
+ "step": 2816
75
+ },
76
+ {
77
+ "epoch": 4.26,
78
+ "learning_rate": 0.00018295454545454547,
79
+ "loss": 0.7639,
80
+ "step": 3000
81
+ },
82
+ {
83
+ "epoch": 4.97,
84
+ "learning_rate": 0.00018011363636363638,
85
+ "loss": 0.7272,
86
+ "step": 3500
87
+ },
88
+ {
89
+ "epoch": 5.0,
90
+ "eval_accuracy": 0.8648,
91
+ "eval_loss": 0.46676263213157654,
92
+ "eval_runtime": 24.1887,
93
+ "eval_samples_per_second": 206.708,
94
+ "eval_steps_per_second": 3.266,
95
+ "step": 3520
96
+ }
97
+ ],
98
+ "max_steps": 35200,
99
+ "num_train_epochs": 50,
100
+ "total_flos": 1.74510121863168e+19,
101
+ "trial_name": null,
102
+ "trial_params": null
103
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5c14c69a507899df9fb75ddc65552a9d9ec21e94493591278fb81115682933e
3
+ size 3899