kharato commited on
Commit
6ffaa72
1 Parent(s): 7e398c0

Model save

Browse files
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ base_model: MCG-NJU/videomae-base-finetuned-kinetics
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ model-index:
9
+ - name: videomae-finetuned_41
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # videomae-finetuned_41
17
+
18
+ This model is a fine-tuned version of [MCG-NJU/videomae-base-finetuned-kinetics](https://huggingface.co/MCG-NJU/videomae-base-finetuned-kinetics) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 7.9297
21
+ - Accuracy: 0.3317
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 5e-05
41
+ - train_batch_size: 2
42
+ - eval_batch_size: 2
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_ratio: 0.2
47
+ - training_steps: 61640
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
+ |:-------------:|:-----:|:-----:|:---------------:|:--------:|
53
+ | 0.5175 | 0.05 | 3083 | 4.6977 | 0.3008 |
54
+ | 0.5311 | 1.05 | 6166 | 5.6012 | 0.3081 |
55
+ | 0.5884 | 2.05 | 9249 | 6.2252 | 0.3175 |
56
+ | 0.5206 | 3.05 | 12332 | 6.7917 | 0.3248 |
57
+ | 0.4449 | 4.05 | 15415 | 6.1943 | 0.3008 |
58
+ | 0.3783 | 5.05 | 18498 | 6.8339 | 0.3150 |
59
+ | 0.5032 | 6.05 | 21581 | 6.6566 | 0.3077 |
60
+ | 0.4091 | 7.05 | 24664 | 6.8013 | 0.2972 |
61
+ | 0.4436 | 8.05 | 27747 | 6.8549 | 0.3 |
62
+ | 0.3474 | 9.05 | 30830 | 7.0015 | 0.3268 |
63
+ | 0.2151 | 10.05 | 33913 | 7.7671 | 0.3041 |
64
+ | 0.3597 | 11.05 | 36996 | 7.0724 | 0.3293 |
65
+ | 0.1673 | 12.05 | 40079 | 7.5805 | 0.3248 |
66
+ | 0.114 | 13.05 | 43162 | 7.8196 | 0.3175 |
67
+ | 0.2088 | 14.05 | 46245 | 7.7103 | 0.3272 |
68
+ | 0.1662 | 15.05 | 49328 | 7.7613 | 0.3248 |
69
+ | 0.1961 | 16.05 | 52411 | 7.7730 | 0.3297 |
70
+ | 0.1436 | 17.05 | 55494 | 7.9297 | 0.3317 |
71
+ | 0.1134 | 18.05 | 58577 | 8.0447 | 0.3268 |
72
+ | 0.0634 | 19.05 | 61640 | 7.9717 | 0.3305 |
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.38.1
78
+ - Pytorch 2.0.1+cu118
79
+ - Datasets 2.18.0
80
+ - Tokenizers 0.15.2
config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "MCG-NJU/videomae-base-finetuned-kinetics",
3
+ "architectures": [
4
+ "VideoMAEForVideoClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "decoder_hidden_size": 384,
8
+ "decoder_intermediate_size": 1536,
9
+ "decoder_num_attention_heads": 6,
10
+ "decoder_num_hidden_layers": 4,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.0,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "BAND",
16
+ "1": "BOX",
17
+ "2": "DIRECTION",
18
+ "3": "ELECTRIC_CAR",
19
+ "4": "FACILITY",
20
+ "5": "KANBAN",
21
+ "6": "NEUTRAL",
22
+ "7": "PALLET",
23
+ "8": "PARTS",
24
+ "9": "SHELF",
25
+ "10": "WAGON"
26
+ },
27
+ "image_size": 224,
28
+ "initializer_range": 0.02,
29
+ "intermediate_size": 3072,
30
+ "label2id": {
31
+ "BAND": 0,
32
+ "BOX": 1,
33
+ "DIRECTION": 2,
34
+ "ELECTRIC_CAR": 3,
35
+ "FACILITY": 4,
36
+ "KANBAN": 5,
37
+ "NEUTRAL": 6,
38
+ "PALLET": 7,
39
+ "PARTS": 8,
40
+ "SHELF": 9,
41
+ "WAGON": 10
42
+ },
43
+ "layer_norm_eps": 1e-12,
44
+ "model_type": "videomae",
45
+ "norm_pix_loss": false,
46
+ "num_attention_heads": 12,
47
+ "num_channels": 3,
48
+ "num_frames": 16,
49
+ "num_hidden_layers": 12,
50
+ "patch_size": 16,
51
+ "problem_type": "single_label_classification",
52
+ "qkv_bias": true,
53
+ "torch_dtype": "float32",
54
+ "transformers_version": "4.38.1",
55
+ "tubelet_size": 2,
56
+ "use_mean_pooling": true
57
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:317e80e226e8dec173011e2f7aa2e44fc7a87c18be41c6dcd5cc457f62db8966
3
+ size 344965020
preprocessor_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_normalize": true,
8
+ "do_rescale": true,
9
+ "do_resize": true,
10
+ "image_mean": [
11
+ 0.485,
12
+ 0.456,
13
+ 0.406
14
+ ],
15
+ "image_processor_type": "VideoMAEImageProcessor",
16
+ "image_std": [
17
+ 0.229,
18
+ 0.224,
19
+ 0.225
20
+ ],
21
+ "resample": 2,
22
+ "rescale_factor": 0.00392156862745098,
23
+ "size": {
24
+ "shortest_edge": 224
25
+ }
26
+ }
runs/Aug13_15-37-23_WIN-645TP8FEMLV/events.out.tfevents.1723531046.WIN-645TP8FEMLV.24196.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03c7283c87e93b98ee9478be006da26d7f3e6eceadbe54c013b2351e541058cc
3
+ size 78014
runs/Aug13_15-37-23_WIN-645TP8FEMLV/events.out.tfevents.1723597114.WIN-645TP8FEMLV.24196.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fc5b140319fc39723dac0ee1f43b5e3c4096802b68660e83e4b4c38b671cb04
3
+ size 417
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a84ba33865b66b576ff42c8e2bd04fa99b4e4df755bdb9e590ef978343c0053
3
+ size 4475