bdpc commited on
Commit
a0629ad
1 Parent(s): d6ebd0d

Saving best model to hub

Browse files
README.md ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: microsoft/resnet-50
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ model-index:
9
+ - name: resnet101_rvl-cdip-cnn_rvl_cdip-NK1000_kd_CEKD_t2.5_a0.5
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # resnet101_rvl-cdip-cnn_rvl_cdip-NK1000_kd_CEKD_t2.5_a0.5
17
+
18
+ This model is a fine-tuned version of [microsoft/resnet-50](https://huggingface.co/microsoft/resnet-50) on the None dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 0.5837
21
+ - Accuracy: 0.7867
22
+ - Brier Loss: 0.3013
23
+ - Nll: 1.9882
24
+ - F1 Micro: 0.7868
25
+ - F1 Macro: 0.7860
26
+ - Ece: 0.0529
27
+ - Aurc: 0.0581
28
+
29
+ ## Model description
30
+
31
+ More information needed
32
+
33
+ ## Intended uses & limitations
34
+
35
+ More information needed
36
+
37
+ ## Training and evaluation data
38
+
39
+ More information needed
40
+
41
+ ## Training procedure
42
+
43
+ ### Training hyperparameters
44
+
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 0.0001
47
+ - train_batch_size: 64
48
+ - eval_batch_size: 64
49
+ - seed: 42
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - lr_scheduler_warmup_ratio: 0.1
53
+ - num_epochs: 100
54
+
55
+ ### Training results
56
+
57
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Brier Loss | Nll | F1 Micro | F1 Macro | Ece | Aurc |
58
+ |:-------------:|:-----:|:-----:|:---------------:|:--------:|:----------:|:------:|:--------:|:--------:|:------:|:------:|
59
+ | No log | 1.0 | 250 | 4.1958 | 0.1035 | 0.9350 | 9.1004 | 0.1035 | 0.0792 | 0.0472 | 0.9013 |
60
+ | 4.2322 | 2.0 | 500 | 4.0778 | 0.173 | 0.9251 | 6.5742 | 0.173 | 0.1393 | 0.0993 | 0.7501 |
61
+ | 4.2322 | 3.0 | 750 | 3.6484 | 0.339 | 0.8778 | 4.9108 | 0.339 | 0.2957 | 0.2172 | 0.5305 |
62
+ | 3.5256 | 4.0 | 1000 | 2.5967 | 0.4592 | 0.6991 | 3.3640 | 0.4592 | 0.4220 | 0.1274 | 0.3285 |
63
+ | 3.5256 | 5.0 | 1250 | 2.0345 | 0.5417 | 0.6078 | 3.0118 | 0.5417 | 0.5180 | 0.0976 | 0.2447 |
64
+ | 1.9172 | 6.0 | 1500 | 1.4417 | 0.625 | 0.5029 | 2.7890 | 0.625 | 0.6123 | 0.0549 | 0.1623 |
65
+ | 1.9172 | 7.0 | 1750 | 1.3298 | 0.639 | 0.4852 | 2.6110 | 0.639 | 0.6320 | 0.0558 | 0.1501 |
66
+ | 1.1801 | 8.0 | 2000 | 1.1697 | 0.674 | 0.4473 | 2.4787 | 0.674 | 0.6712 | 0.0466 | 0.1283 |
67
+ | 1.1801 | 9.0 | 2250 | 0.9625 | 0.7093 | 0.4020 | 2.3242 | 0.7093 | 0.7085 | 0.0526 | 0.1017 |
68
+ | 0.8029 | 10.0 | 2500 | 0.9477 | 0.7215 | 0.3893 | 2.3193 | 0.7215 | 0.7228 | 0.0515 | 0.0971 |
69
+ | 0.8029 | 11.0 | 2750 | 0.8527 | 0.7375 | 0.3692 | 2.2785 | 0.7375 | 0.7377 | 0.0490 | 0.0870 |
70
+ | 0.5717 | 12.0 | 3000 | 0.7377 | 0.7515 | 0.3470 | 2.1475 | 0.7515 | 0.7529 | 0.0552 | 0.0757 |
71
+ | 0.5717 | 13.0 | 3250 | 0.7309 | 0.7498 | 0.3469 | 2.1250 | 0.7498 | 0.7494 | 0.0589 | 0.0758 |
72
+ | 0.4414 | 14.0 | 3500 | 0.7165 | 0.7558 | 0.3427 | 2.1045 | 0.7558 | 0.7576 | 0.0582 | 0.0721 |
73
+ | 0.4414 | 15.0 | 3750 | 0.6865 | 0.7678 | 0.3319 | 2.0457 | 0.7678 | 0.7688 | 0.0551 | 0.0697 |
74
+ | 0.3691 | 16.0 | 4000 | 0.7002 | 0.7662 | 0.3348 | 2.1280 | 0.7663 | 0.7664 | 0.0567 | 0.0698 |
75
+ | 0.3691 | 17.0 | 4250 | 0.6896 | 0.7628 | 0.3326 | 2.0750 | 0.7628 | 0.7631 | 0.0608 | 0.0691 |
76
+ | 0.3214 | 18.0 | 4500 | 0.6666 | 0.7715 | 0.3258 | 2.0468 | 0.7715 | 0.7707 | 0.0544 | 0.0680 |
77
+ | 0.3214 | 19.0 | 4750 | 0.6735 | 0.7702 | 0.3277 | 2.0544 | 0.7702 | 0.7700 | 0.0571 | 0.0681 |
78
+ | 0.2914 | 20.0 | 5000 | 0.6607 | 0.772 | 0.3241 | 2.0364 | 0.772 | 0.7729 | 0.0525 | 0.0659 |
79
+ | 0.2914 | 21.0 | 5250 | 0.6625 | 0.7688 | 0.3217 | 2.0387 | 0.7688 | 0.7703 | 0.0455 | 0.0664 |
80
+ | 0.2653 | 22.0 | 5500 | 0.6543 | 0.775 | 0.3200 | 2.0560 | 0.775 | 0.7752 | 0.0507 | 0.0647 |
81
+ | 0.2653 | 23.0 | 5750 | 0.6409 | 0.7725 | 0.3188 | 2.0091 | 0.7725 | 0.7733 | 0.0554 | 0.0647 |
82
+ | 0.2482 | 24.0 | 6000 | 0.6452 | 0.7758 | 0.3191 | 2.0256 | 0.7758 | 0.7756 | 0.0502 | 0.0655 |
83
+ | 0.2482 | 25.0 | 6250 | 0.6401 | 0.7742 | 0.3196 | 2.0668 | 0.7742 | 0.7745 | 0.0528 | 0.0648 |
84
+ | 0.2354 | 26.0 | 6500 | 0.6316 | 0.775 | 0.3171 | 2.0150 | 0.775 | 0.7755 | 0.0555 | 0.0634 |
85
+ | 0.2354 | 27.0 | 6750 | 0.6257 | 0.7808 | 0.3147 | 2.0129 | 0.7808 | 0.7808 | 0.0503 | 0.0624 |
86
+ | 0.2229 | 28.0 | 7000 | 0.6343 | 0.7778 | 0.3144 | 2.0910 | 0.7778 | 0.7776 | 0.0510 | 0.0624 |
87
+ | 0.2229 | 29.0 | 7250 | 0.6206 | 0.781 | 0.3115 | 2.0399 | 0.7810 | 0.7798 | 0.0555 | 0.0606 |
88
+ | 0.2147 | 30.0 | 7500 | 0.6262 | 0.777 | 0.3124 | 2.0603 | 0.777 | 0.7772 | 0.0539 | 0.0616 |
89
+ | 0.2147 | 31.0 | 7750 | 0.6265 | 0.7788 | 0.3137 | 2.0833 | 0.7788 | 0.7777 | 0.0532 | 0.0614 |
90
+ | 0.2058 | 32.0 | 8000 | 0.6134 | 0.7815 | 0.3119 | 2.0369 | 0.7815 | 0.7815 | 0.0514 | 0.0615 |
91
+ | 0.2058 | 33.0 | 8250 | 0.6153 | 0.7772 | 0.3133 | 2.0513 | 0.7773 | 0.7772 | 0.0534 | 0.0623 |
92
+ | 0.1994 | 34.0 | 8500 | 0.6143 | 0.7853 | 0.3098 | 2.0188 | 0.7853 | 0.7857 | 0.0508 | 0.0611 |
93
+ | 0.1994 | 35.0 | 8750 | 0.6096 | 0.7827 | 0.3086 | 2.0134 | 0.7828 | 0.7828 | 0.0512 | 0.0606 |
94
+ | 0.1932 | 36.0 | 9000 | 0.6094 | 0.784 | 0.3067 | 2.0151 | 0.7840 | 0.7847 | 0.0471 | 0.0602 |
95
+ | 0.1932 | 37.0 | 9250 | 0.6142 | 0.7833 | 0.3111 | 2.0213 | 0.7833 | 0.7829 | 0.0542 | 0.0608 |
96
+ | 0.1895 | 38.0 | 9500 | 0.6103 | 0.7812 | 0.3094 | 2.0594 | 0.7812 | 0.7799 | 0.0529 | 0.0603 |
97
+ | 0.1895 | 39.0 | 9750 | 0.6059 | 0.781 | 0.3078 | 2.0386 | 0.7810 | 0.7806 | 0.0545 | 0.0607 |
98
+ | 0.1848 | 40.0 | 10000 | 0.6042 | 0.782 | 0.3072 | 2.0133 | 0.782 | 0.7824 | 0.0527 | 0.0603 |
99
+ | 0.1848 | 41.0 | 10250 | 0.5991 | 0.785 | 0.3043 | 2.0124 | 0.785 | 0.7853 | 0.0496 | 0.0594 |
100
+ | 0.1793 | 42.0 | 10500 | 0.6034 | 0.784 | 0.3058 | 2.0607 | 0.7840 | 0.7838 | 0.0490 | 0.0599 |
101
+ | 0.1793 | 43.0 | 10750 | 0.6047 | 0.7827 | 0.3068 | 2.0139 | 0.7828 | 0.7819 | 0.0492 | 0.0595 |
102
+ | 0.1768 | 44.0 | 11000 | 0.5982 | 0.785 | 0.3057 | 2.0303 | 0.785 | 0.7843 | 0.0473 | 0.0596 |
103
+ | 0.1768 | 45.0 | 11250 | 0.6036 | 0.7795 | 0.3087 | 2.0173 | 0.7795 | 0.7788 | 0.0549 | 0.0607 |
104
+ | 0.1743 | 46.0 | 11500 | 0.5974 | 0.785 | 0.3060 | 2.0026 | 0.785 | 0.7839 | 0.0478 | 0.0596 |
105
+ | 0.1743 | 47.0 | 11750 | 0.5996 | 0.782 | 0.3068 | 2.0144 | 0.782 | 0.7825 | 0.0480 | 0.0598 |
106
+ | 0.1707 | 48.0 | 12000 | 0.5958 | 0.7833 | 0.3079 | 2.0344 | 0.7833 | 0.7827 | 0.0500 | 0.0598 |
107
+ | 0.1707 | 49.0 | 12250 | 0.5969 | 0.782 | 0.3060 | 2.0162 | 0.782 | 0.7820 | 0.0482 | 0.0597 |
108
+ | 0.1683 | 50.0 | 12500 | 0.5933 | 0.784 | 0.3043 | 1.9897 | 0.7840 | 0.7836 | 0.0496 | 0.0589 |
109
+ | 0.1683 | 51.0 | 12750 | 0.5935 | 0.7833 | 0.3042 | 2.0142 | 0.7833 | 0.7829 | 0.0501 | 0.0586 |
110
+ | 0.1649 | 52.0 | 13000 | 0.5950 | 0.7847 | 0.3050 | 2.0125 | 0.7847 | 0.7851 | 0.0475 | 0.0591 |
111
+ | 0.1649 | 53.0 | 13250 | 0.5904 | 0.7837 | 0.3020 | 1.9830 | 0.7837 | 0.7837 | 0.0504 | 0.0584 |
112
+ | 0.1636 | 54.0 | 13500 | 0.5926 | 0.785 | 0.3042 | 2.0006 | 0.785 | 0.7845 | 0.0493 | 0.0588 |
113
+ | 0.1636 | 55.0 | 13750 | 0.5885 | 0.7847 | 0.3029 | 2.0025 | 0.7847 | 0.7843 | 0.0505 | 0.0585 |
114
+ | 0.1616 | 56.0 | 14000 | 0.5920 | 0.788 | 0.3041 | 2.0174 | 0.788 | 0.7878 | 0.0520 | 0.0591 |
115
+ | 0.1616 | 57.0 | 14250 | 0.5927 | 0.7863 | 0.3033 | 2.0321 | 0.7863 | 0.7858 | 0.0521 | 0.0588 |
116
+ | 0.1592 | 58.0 | 14500 | 0.5878 | 0.787 | 0.3017 | 1.9751 | 0.787 | 0.7874 | 0.0461 | 0.0584 |
117
+ | 0.1592 | 59.0 | 14750 | 0.5888 | 0.7867 | 0.3030 | 1.9996 | 0.7868 | 0.7864 | 0.0494 | 0.0582 |
118
+ | 0.1585 | 60.0 | 15000 | 0.5929 | 0.786 | 0.3052 | 2.0237 | 0.786 | 0.7857 | 0.0512 | 0.0584 |
119
+ | 0.1585 | 61.0 | 15250 | 0.5894 | 0.7865 | 0.3026 | 1.9895 | 0.7865 | 0.7864 | 0.0548 | 0.0585 |
120
+ | 0.1562 | 62.0 | 15500 | 0.5903 | 0.7873 | 0.3033 | 1.9670 | 0.7873 | 0.7870 | 0.0481 | 0.0584 |
121
+ | 0.1562 | 63.0 | 15750 | 0.5896 | 0.7853 | 0.3023 | 1.9681 | 0.7853 | 0.7850 | 0.0520 | 0.0587 |
122
+ | 0.1548 | 64.0 | 16000 | 0.5903 | 0.7847 | 0.3027 | 1.9865 | 0.7847 | 0.7846 | 0.0506 | 0.0587 |
123
+ | 0.1548 | 65.0 | 16250 | 0.5910 | 0.7853 | 0.3039 | 2.0009 | 0.7853 | 0.7849 | 0.0515 | 0.0593 |
124
+ | 0.1537 | 66.0 | 16500 | 0.5866 | 0.7883 | 0.3012 | 1.9561 | 0.7883 | 0.7881 | 0.0447 | 0.0581 |
125
+ | 0.1537 | 67.0 | 16750 | 0.5858 | 0.7867 | 0.3009 | 1.9868 | 0.7868 | 0.7861 | 0.0486 | 0.0577 |
126
+ | 0.1526 | 68.0 | 17000 | 0.5886 | 0.7867 | 0.3024 | 2.0009 | 0.7868 | 0.7862 | 0.0530 | 0.0587 |
127
+ | 0.1526 | 69.0 | 17250 | 0.5850 | 0.7863 | 0.3010 | 2.0095 | 0.7863 | 0.7860 | 0.0510 | 0.0581 |
128
+ | 0.1508 | 70.0 | 17500 | 0.5867 | 0.7865 | 0.3019 | 2.0304 | 0.7865 | 0.7861 | 0.0525 | 0.0583 |
129
+ | 0.1508 | 71.0 | 17750 | 0.5895 | 0.7857 | 0.3038 | 2.0013 | 0.7857 | 0.7853 | 0.0478 | 0.0586 |
130
+ | 0.15 | 72.0 | 18000 | 0.5894 | 0.7847 | 0.3025 | 2.0051 | 0.7847 | 0.7845 | 0.0500 | 0.0586 |
131
+ | 0.15 | 73.0 | 18250 | 0.5867 | 0.7865 | 0.3022 | 1.9634 | 0.7865 | 0.7860 | 0.0489 | 0.0582 |
132
+ | 0.149 | 74.0 | 18500 | 0.5888 | 0.7857 | 0.3026 | 1.9817 | 0.7857 | 0.7851 | 0.0497 | 0.0584 |
133
+ | 0.149 | 75.0 | 18750 | 0.5823 | 0.7885 | 0.2994 | 1.9873 | 0.7885 | 0.7880 | 0.0476 | 0.0577 |
134
+ | 0.1483 | 76.0 | 19000 | 0.5866 | 0.7853 | 0.3025 | 1.9870 | 0.7853 | 0.7849 | 0.0531 | 0.0583 |
135
+ | 0.1483 | 77.0 | 19250 | 0.5866 | 0.7867 | 0.3013 | 1.9933 | 0.7868 | 0.7862 | 0.0498 | 0.0577 |
136
+ | 0.1478 | 78.0 | 19500 | 0.5844 | 0.787 | 0.3010 | 1.9793 | 0.787 | 0.7868 | 0.0465 | 0.0579 |
137
+ | 0.1478 | 79.0 | 19750 | 0.5850 | 0.7857 | 0.3005 | 1.9856 | 0.7857 | 0.7855 | 0.0489 | 0.0580 |
138
+ | 0.1463 | 80.0 | 20000 | 0.5829 | 0.7893 | 0.2999 | 2.0003 | 0.7893 | 0.7890 | 0.0543 | 0.0578 |
139
+ | 0.1463 | 81.0 | 20250 | 0.5845 | 0.7867 | 0.3011 | 2.0178 | 0.7868 | 0.7864 | 0.0494 | 0.0580 |
140
+ | 0.1457 | 82.0 | 20500 | 0.5878 | 0.7865 | 0.3022 | 2.0108 | 0.7865 | 0.7861 | 0.0507 | 0.0583 |
141
+ | 0.1457 | 83.0 | 20750 | 0.5862 | 0.7865 | 0.3016 | 1.9996 | 0.7865 | 0.7865 | 0.0505 | 0.0585 |
142
+ | 0.1452 | 84.0 | 21000 | 0.5851 | 0.7863 | 0.3011 | 2.0002 | 0.7863 | 0.7859 | 0.0481 | 0.0582 |
143
+ | 0.1452 | 85.0 | 21250 | 0.5850 | 0.787 | 0.3013 | 1.9659 | 0.787 | 0.7867 | 0.0524 | 0.0582 |
144
+ | 0.1449 | 86.0 | 21500 | 0.5878 | 0.7867 | 0.3023 | 1.9837 | 0.7868 | 0.7866 | 0.0526 | 0.0581 |
145
+ | 0.1449 | 87.0 | 21750 | 0.5844 | 0.7873 | 0.3010 | 1.9807 | 0.7873 | 0.7865 | 0.0522 | 0.0577 |
146
+ | 0.1437 | 88.0 | 22000 | 0.5846 | 0.7877 | 0.3012 | 1.9947 | 0.7877 | 0.7869 | 0.0464 | 0.0580 |
147
+ | 0.1437 | 89.0 | 22250 | 0.5859 | 0.787 | 0.3016 | 2.0002 | 0.787 | 0.7867 | 0.0503 | 0.0581 |
148
+ | 0.143 | 90.0 | 22500 | 0.5838 | 0.7865 | 0.3010 | 1.9996 | 0.7865 | 0.7859 | 0.0496 | 0.0576 |
149
+ | 0.143 | 91.0 | 22750 | 0.5843 | 0.7837 | 0.3011 | 1.9683 | 0.7837 | 0.7834 | 0.0501 | 0.0583 |
150
+ | 0.1426 | 92.0 | 23000 | 0.5843 | 0.7873 | 0.3010 | 1.9960 | 0.7873 | 0.7870 | 0.0524 | 0.0578 |
151
+ | 0.1426 | 93.0 | 23250 | 0.5827 | 0.7847 | 0.3005 | 1.9719 | 0.7847 | 0.7844 | 0.0506 | 0.0579 |
152
+ | 0.1428 | 94.0 | 23500 | 0.5831 | 0.7865 | 0.3009 | 1.9781 | 0.7865 | 0.7862 | 0.0517 | 0.0579 |
153
+ | 0.1428 | 95.0 | 23750 | 0.5821 | 0.784 | 0.3001 | 1.9641 | 0.7840 | 0.7838 | 0.0505 | 0.0579 |
154
+ | 0.1424 | 96.0 | 24000 | 0.5850 | 0.7845 | 0.3020 | 1.9667 | 0.7845 | 0.7842 | 0.0526 | 0.0584 |
155
+ | 0.1424 | 97.0 | 24250 | 0.5850 | 0.7847 | 0.3012 | 1.9776 | 0.7847 | 0.7844 | 0.0508 | 0.0579 |
156
+ | 0.142 | 98.0 | 24500 | 0.5845 | 0.7877 | 0.3011 | 1.9745 | 0.7877 | 0.7870 | 0.0491 | 0.0579 |
157
+ | 0.142 | 99.0 | 24750 | 0.5834 | 0.7853 | 0.3010 | 1.9679 | 0.7853 | 0.7852 | 0.0506 | 0.0581 |
158
+ | 0.1416 | 100.0 | 25000 | 0.5837 | 0.7867 | 0.3013 | 1.9882 | 0.7868 | 0.7860 | 0.0529 | 0.0581 |
159
+
160
+
161
+ ### Framework versions
162
+
163
+ - Transformers 4.33.3
164
+ - Pytorch 2.2.0.dev20231002
165
+ - Datasets 2.7.1
166
+ - Tokenizers 0.13.3
config.json ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/resnet-50",
3
+ "architectures": [
4
+ "ResNetForImageClassification"
5
+ ],
6
+ "depths": [
7
+ 3,
8
+ 4,
9
+ 6,
10
+ 3
11
+ ],
12
+ "downsample_in_first_stage": false,
13
+ "embedding_size": 64,
14
+ "hidden_act": "relu",
15
+ "hidden_sizes": [
16
+ 256,
17
+ 512,
18
+ 1024,
19
+ 2048
20
+ ],
21
+ "id2label": {
22
+ "0": "letter",
23
+ "1": "form",
24
+ "2": "email",
25
+ "3": "handwritten",
26
+ "4": "advertisement",
27
+ "5": "scientific_report",
28
+ "6": "scientific_publication",
29
+ "7": "specification",
30
+ "8": "file_folder",
31
+ "9": "news_article",
32
+ "10": "budget",
33
+ "11": "invoice",
34
+ "12": "presentation",
35
+ "13": "questionnaire",
36
+ "14": "resume",
37
+ "15": "memo"
38
+ },
39
+ "label2id": {
40
+ "advertisement": 4,
41
+ "budget": 10,
42
+ "email": 2,
43
+ "file_folder": 8,
44
+ "form": 1,
45
+ "handwritten": 3,
46
+ "invoice": 11,
47
+ "letter": 0,
48
+ "memo": 15,
49
+ "news_article": 9,
50
+ "presentation": 12,
51
+ "questionnaire": 13,
52
+ "resume": 14,
53
+ "scientific_publication": 6,
54
+ "scientific_report": 5,
55
+ "specification": 7
56
+ },
57
+ "layer_type": "bottleneck",
58
+ "model_type": "resnet",
59
+ "num_channels": 3,
60
+ "out_features": [
61
+ "stage4"
62
+ ],
63
+ "out_indices": [
64
+ 4
65
+ ],
66
+ "problem_type": "single_label_classification",
67
+ "stage_names": [
68
+ "stem",
69
+ "stage1",
70
+ "stage2",
71
+ "stage3",
72
+ "stage4"
73
+ ],
74
+ "torch_dtype": "float32",
75
+ "transformers_version": "4.33.3"
76
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:512dc13c878586a02c31f49eefac143bbb678307c69e158ba016ba1e1c4b94a6
3
+ size 94490186
test-logits.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2fc7fc30636a69d042f7c2b3ff3c1bcc5d32e3823da0976365a56d61ce9385c
3
+ size 2310336
test-references.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e7ff7e9109ab12165bc737f7df3f352e97c091a42dbf5ba5121e717df12a42a
3
+ size 34462
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c934bdd4a9d440a95a4a8c49cfe2948d206773c66a3883b3ddcbb9a7ff728ac
3
+ size 4792
validation-logits.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abbecbbfcd9ed098d673244c56f80d1918e1d5a7196408fcf6df6c49258ea885
3
+ size 230932
validation-references.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9d6bda37454a6b9af2a9246dad3ac3aed3f946e7420dd22cdc88f6c0f94c4e0
3
+ size 326