End of training
Browse files- README.md +14 -23
- adapter_model.safetensors +1 -1
- training_args.bin +1 -1
README.md
CHANGED
@@ -15,10 +15,10 @@ should probably proofread and complete it, then remove this comment. -->
|
|
15 |
|
16 |
# finetune_colpali_v1_2-german-4bit
|
17 |
|
18 |
-
This model is a fine-tuned version of [vidore/colpaligemma-3b-pt-448-base](https://huggingface.co/vidore/colpaligemma-3b-pt-448-base) on the
|
19 |
It achieves the following results on the evaluation set:
|
20 |
-
- Loss: 0.
|
21 |
-
- Model Preparation Time: 0.
|
22 |
|
23 |
## Model description
|
24 |
|
@@ -46,31 +46,22 @@ The following hyperparameters were used during training:
|
|
46 |
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
47 |
- lr_scheduler_type: linear
|
48 |
- lr_scheduler_warmup_steps: 100
|
49 |
-
- num_epochs:
|
50 |
|
51 |
### Training results
|
52 |
|
53 |
| Training Loss | Epoch | Step | Validation Loss | Model Preparation Time |
|
54 |
|:-------------:|:------:|:----:|:---------------:|:----------------------:|
|
55 |
-
| No log | 0.0533 | 1 | 0.
|
56 |
-
| 1.
|
57 |
-
| 1.
|
58 |
-
| 0.
|
59 |
-
| 0.
|
60 |
-
| 0.
|
61 |
-
| 0.
|
62 |
-
| 0.
|
63 |
-
| 0.
|
64 |
-
| 0.
|
65 |
-
| 0.2019 | 5.3333 | 100 | 0.0390 | 0.0095 |
|
66 |
-
| 0.1092 | 5.8667 | 110 | 0.0337 | 0.0095 |
|
67 |
-
| 0.0884 | 6.4 | 120 | 0.0344 | 0.0095 |
|
68 |
-
| 0.2341 | 6.9333 | 130 | 0.0433 | 0.0095 |
|
69 |
-
| 0.1872 | 7.4667 | 140 | 0.0448 | 0.0095 |
|
70 |
-
| 0.1533 | 8.0 | 150 | 0.0485 | 0.0095 |
|
71 |
-
| 0.1681 | 8.5333 | 160 | 0.0525 | 0.0095 |
|
72 |
-
| 0.2414 | 9.0667 | 170 | 0.0590 | 0.0095 |
|
73 |
-
| 0.1814 | 9.6 | 180 | 0.0594 | 0.0095 |
|
74 |
|
75 |
|
76 |
### Framework versions
|
|
|
15 |
|
16 |
# finetune_colpali_v1_2-german-4bit
|
17 |
|
18 |
+
This model is a fine-tuned version of [vidore/colpaligemma-3b-pt-448-base](https://huggingface.co/vidore/colpaligemma-3b-pt-448-base) on the German_docx dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
+
- Loss: 0.1100
|
21 |
+
- Model Preparation Time: 0.008
|
22 |
|
23 |
## Model description
|
24 |
|
|
|
46 |
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
47 |
- lr_scheduler_type: linear
|
48 |
- lr_scheduler_warmup_steps: 100
|
49 |
+
- num_epochs: 5
|
50 |
|
51 |
### Training results
|
52 |
|
53 |
| Training Loss | Epoch | Step | Validation Loss | Model Preparation Time |
|
54 |
|:-------------:|:------:|:----:|:---------------:|:----------------------:|
|
55 |
+
| No log | 0.0533 | 1 | 0.3717 | 0.008 |
|
56 |
+
| 1.1358 | 0.5333 | 10 | 0.3356 | 0.008 |
|
57 |
+
| 1.2182 | 1.0667 | 20 | 0.2811 | 0.008 |
|
58 |
+
| 0.844 | 1.6 | 30 | 0.2365 | 0.008 |
|
59 |
+
| 0.7722 | 2.1333 | 40 | 0.1990 | 0.008 |
|
60 |
+
| 0.4823 | 2.6667 | 50 | 0.1758 | 0.008 |
|
61 |
+
| 0.46 | 3.2 | 60 | 0.1451 | 0.008 |
|
62 |
+
| 0.1477 | 3.7333 | 70 | 0.1252 | 0.008 |
|
63 |
+
| 0.1764 | 4.2667 | 80 | 0.1258 | 0.008 |
|
64 |
+
| 0.2329 | 4.8 | 90 | 0.1100 | 0.008 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
65 |
|
66 |
|
67 |
### Framework versions
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 157071680
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a7c8c840d5514721e64e756d0e0c35897eda4348d6d8fa2d0a258bf1cb86e47
|
3 |
size 157071680
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5240
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:07741319549368dc90fb30caeb2bf231216e51d35276e55123323b45017e06d4
|
3 |
size 5240
|