momo345 commited on
Commit
d0f9fed
1 Parent(s): 42824b3

End of training

Browse files
README.md ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: microsoft/cvt-13
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - imagefolder
8
+ metrics:
9
+ - accuracy
10
+ - precision
11
+ - recall
12
+ - f1
13
+ model-index:
14
+ - name: cvt-13-finetuned-flower
15
+ results:
16
+ - task:
17
+ name: Image Classification
18
+ type: image-classification
19
+ dataset:
20
+ name: imagefolder
21
+ type: imagefolder
22
+ config: default
23
+ split: train
24
+ args: default
25
+ metrics:
26
+ - name: Accuracy
27
+ type: accuracy
28
+ value: 0.9368421052631579
29
+ - name: Precision
30
+ type: precision
31
+ value: 0.9374630861809764
32
+ - name: Recall
33
+ type: recall
34
+ value: 0.9368421052631579
35
+ - name: F1
36
+ type: f1
37
+ value: 0.9341589949056075
38
+ ---
39
+
40
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
41
+ should probably proofread and complete it, then remove this comment. -->
42
+
43
+ # cvt-13-finetuned-flower
44
+
45
+ This model is a fine-tuned version of [microsoft/cvt-13](https://huggingface.co/microsoft/cvt-13) on the imagefolder dataset.
46
+ It achieves the following results on the evaluation set:
47
+ - Loss: 0.2151
48
+ - Accuracy: 0.9368
49
+ - Precision: 0.9375
50
+ - Recall: 0.9368
51
+ - F1: 0.9342
52
+
53
+ ## Model description
54
+
55
+ More information needed
56
+
57
+ ## Intended uses & limitations
58
+
59
+ More information needed
60
+
61
+ ## Training and evaluation data
62
+
63
+ More information needed
64
+
65
+ ## Training procedure
66
+
67
+ ### Training hyperparameters
68
+
69
+ The following hyperparameters were used during training:
70
+ - learning_rate: 0.005
71
+ - train_batch_size: 64
72
+ - eval_batch_size: 64
73
+ - seed: 42
74
+ - gradient_accumulation_steps: 4
75
+ - total_train_batch_size: 256
76
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
77
+ - lr_scheduler_type: linear
78
+ - lr_scheduler_warmup_ratio: 0.1
79
+ - num_epochs: 20
80
+
81
+ ### Training results
82
+
83
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
84
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
85
+ | 1.0555 | 1.0 | 40 | 0.3933 | 0.8766 | 0.8828 | 0.8766 | 0.8713 |
86
+ | 1.1941 | 2.0 | 80 | 1.0797 | 0.6726 | 0.7515 | 0.6726 | 0.6546 |
87
+ | 1.2286 | 3.0 | 120 | 0.8459 | 0.7347 | 0.7820 | 0.7347 | 0.7343 |
88
+ | 1.209 | 4.0 | 160 | 0.6660 | 0.7880 | 0.8173 | 0.7880 | 0.7833 |
89
+ | 1.1158 | 5.0 | 200 | 0.7348 | 0.7597 | 0.7809 | 0.7597 | 0.7561 |
90
+ | 1.1113 | 6.0 | 240 | 0.6387 | 0.8062 | 0.8164 | 0.8062 | 0.7986 |
91
+ | 1.0332 | 7.0 | 280 | 0.6555 | 0.7887 | 0.8064 | 0.7887 | 0.7831 |
92
+ | 1.0234 | 8.0 | 320 | 0.5776 | 0.8276 | 0.8447 | 0.8276 | 0.8177 |
93
+ | 0.9997 | 9.0 | 360 | 0.5784 | 0.8214 | 0.8421 | 0.8214 | 0.8169 |
94
+ | 0.9421 | 10.0 | 400 | 0.4667 | 0.8486 | 0.8600 | 0.8486 | 0.8453 |
95
+ | 0.9057 | 11.0 | 440 | 0.4508 | 0.8541 | 0.8711 | 0.8541 | 0.8487 |
96
+ | 0.8662 | 12.0 | 480 | 0.3517 | 0.8911 | 0.8938 | 0.8911 | 0.8868 |
97
+ | 0.8341 | 13.0 | 520 | 0.3191 | 0.8976 | 0.9021 | 0.8976 | 0.8945 |
98
+ | 0.757 | 14.0 | 560 | 0.2785 | 0.9183 | 0.9199 | 0.9183 | 0.9144 |
99
+ | 0.7906 | 15.0 | 600 | 0.2698 | 0.9201 | 0.9218 | 0.9201 | 0.9172 |
100
+ | 0.7464 | 16.0 | 640 | 0.2594 | 0.9216 | 0.9232 | 0.9216 | 0.9188 |
101
+ | 0.7335 | 17.0 | 680 | 0.2491 | 0.9263 | 0.9281 | 0.9263 | 0.9240 |
102
+ | 0.7085 | 18.0 | 720 | 0.2396 | 0.9303 | 0.9304 | 0.9303 | 0.9272 |
103
+ | 0.7177 | 19.0 | 760 | 0.2171 | 0.9350 | 0.9355 | 0.9350 | 0.9321 |
104
+ | 0.6735 | 20.0 | 800 | 0.2151 | 0.9368 | 0.9375 | 0.9368 | 0.9342 |
105
+
106
+
107
+ ### Framework versions
108
+
109
+ - Transformers 4.39.3
110
+ - Pytorch 2.0.1
111
+ - Datasets 2.18.0
112
+ - Tokenizers 0.15.2
config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/cvt-13",
3
+ "architectures": [
4
+ "CvtForImageClassification"
5
+ ],
6
+ "attention_drop_rate": [
7
+ 0.0,
8
+ 0.0,
9
+ 0.0
10
+ ],
11
+ "cls_token": [
12
+ false,
13
+ false,
14
+ true
15
+ ],
16
+ "depth": [
17
+ 1,
18
+ 2,
19
+ 10
20
+ ],
21
+ "drop_path_rate": [
22
+ 0.0,
23
+ 0.0,
24
+ 0.1
25
+ ],
26
+ "drop_rate": [
27
+ 0.0,
28
+ 0.0,
29
+ 0.0
30
+ ],
31
+ "embed_dim": [
32
+ 64,
33
+ 192,
34
+ 384
35
+ ],
36
+ "id2label": {
37
+ "0": "astilbe",
38
+ "1": "bellflower",
39
+ "2": "black_eyed_susan",
40
+ "3": "calendula",
41
+ "4": "california_poppy",
42
+ "5": "carnation",
43
+ "6": "common_daisy",
44
+ "7": "coreopsis",
45
+ "8": "daffodil",
46
+ "9": "dandelion",
47
+ "10": "iris",
48
+ "11": "magnolia",
49
+ "12": "rose",
50
+ "13": "sunflower",
51
+ "14": "tulip",
52
+ "15": "water_lily"
53
+ },
54
+ "image_size": 224,
55
+ "initializer_range": 0.02,
56
+ "kernel_qkv": [
57
+ 3,
58
+ 3,
59
+ 3
60
+ ],
61
+ "label2id": {
62
+ "astilbe": 0,
63
+ "bellflower": 1,
64
+ "black_eyed_susan": 2,
65
+ "calendula": 3,
66
+ "california_poppy": 4,
67
+ "carnation": 5,
68
+ "common_daisy": 6,
69
+ "coreopsis": 7,
70
+ "daffodil": 8,
71
+ "dandelion": 9,
72
+ "iris": 10,
73
+ "magnolia": 11,
74
+ "rose": 12,
75
+ "sunflower": 13,
76
+ "tulip": 14,
77
+ "water_lily": 15
78
+ },
79
+ "layer_norm_eps": 1e-12,
80
+ "mlp_ratio": [
81
+ 4.0,
82
+ 4.0,
83
+ 4.0
84
+ ],
85
+ "model_type": "cvt",
86
+ "num_channels": 3,
87
+ "num_heads": [
88
+ 1,
89
+ 3,
90
+ 6
91
+ ],
92
+ "num_stages": 3,
93
+ "padding_kv": [
94
+ 1,
95
+ 1,
96
+ 1
97
+ ],
98
+ "padding_q": [
99
+ 1,
100
+ 1,
101
+ 1
102
+ ],
103
+ "patch_padding": [
104
+ 2,
105
+ 1,
106
+ 1
107
+ ],
108
+ "patch_sizes": [
109
+ 7,
110
+ 3,
111
+ 3
112
+ ],
113
+ "patch_stride": [
114
+ 4,
115
+ 2,
116
+ 2
117
+ ],
118
+ "pos_embed": [
119
+ false,
120
+ false,
121
+ false
122
+ ],
123
+ "problem_type": "single_label_classification",
124
+ "qkv_bias": [
125
+ true,
126
+ true,
127
+ true
128
+ ],
129
+ "qkv_projection_method": [
130
+ "dw_bn",
131
+ "dw_bn",
132
+ "dw_bn"
133
+ ],
134
+ "stride_kv": [
135
+ 2,
136
+ 2,
137
+ 2
138
+ ],
139
+ "stride_q": [
140
+ 1,
141
+ 1,
142
+ 1
143
+ ],
144
+ "torch_dtype": "float32",
145
+ "transformers_version": "4.39.3"
146
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6463538d1b5ac793b57f44a56176b26c9e42f4cff63727bc55e851ebd7467b9
3
+ size 78651152
preprocessor_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "crop_pct",
7
+ "resample",
8
+ "do_rescale",
9
+ "rescale_factor",
10
+ "do_normalize",
11
+ "image_mean",
12
+ "image_std",
13
+ "return_tensors",
14
+ "data_format",
15
+ "input_data_format"
16
+ ],
17
+ "crop_pct": 0.875,
18
+ "do_normalize": true,
19
+ "do_rescale": true,
20
+ "do_resize": true,
21
+ "image_mean": [
22
+ 0.485,
23
+ 0.456,
24
+ 0.406
25
+ ],
26
+ "image_processor_type": "ConvNextImageProcessor",
27
+ "image_std": [
28
+ 0.229,
29
+ 0.224,
30
+ 0.225
31
+ ],
32
+ "resample": 3,
33
+ "rescale_factor": 0.00392156862745098,
34
+ "size": {
35
+ "shortest_edge": 224
36
+ }
37
+ }
runs/Apr09_18-34-22_165-1-65-97/events.out.tfevents.1712687667.165-1-65-97.1392.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56e2662a756f43369614b5bcbb1840ca046da6f66d5f4967c83745bd8c25b2ec
3
+ size 32510
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8562dfda49708332113521b2caa694305aa8b4c2ec1af526cdffe0f73979854c
3
+ size 4475