FemkeBakker commited on
Commit
151bdf7
1 Parent(s): 6e7fc51

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -13
README.md CHANGED
@@ -7,32 +7,32 @@ tags:
7
  model-index:
8
  - name: AmsterdamDocClassificationMistral200T1Epochs
9
  results: []
 
 
 
 
 
10
  ---
11
 
12
  # AmsterdamDocClassificationMistral200T1Epochs
13
 
14
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the [AmsterdamDocClassification](https://huggingface.co/datasets/FemkeBakker/AmsterdamBalancedFirst200Tokens) dataset.
15
-
16
-
17
-
18
 
19
  It achieves the following results on the evaluation set:
20
  - Loss: 0.7673
21
 
22
- ## Model description
23
-
24
- More information needed
25
-
26
- ## Intended uses & limitations
27
-
28
- More information needed
29
-
30
  ## Training and evaluation data
31
 
32
- More information needed
 
33
 
34
  ## Training procedure
35
 
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
@@ -57,6 +57,7 @@ The following hyperparameters were used during training:
57
  | 0.5285 | 0.7952 | 492 | 0.7687 |
58
  | 0.7677 | 0.9939 | 615 | 0.7673 |
59
 
 
60
 
61
  ### Framework versions
62
 
@@ -64,3 +65,7 @@ The following hyperparameters were used during training:
64
  - Pytorch 2.3.0+cu121
65
  - Datasets 2.19.1
66
  - Tokenizers 0.19.1
 
 
 
 
 
7
  model-index:
8
  - name: AmsterdamDocClassificationMistral200T1Epochs
9
  results: []
10
+ license: eupl-1.1
11
+ datasets:
12
+ - FemkeBakker/AmsterdamBalancedFirst200Tokens
13
+ language:
14
+ - nl
15
  ---
16
 
17
  # AmsterdamDocClassificationMistral200T1Epochs
18
 
19
+ As part of the Assessing Large Language Models for Document Classification project by the Municipality of Amsterdam, we fine-tune Mistral, Llama, and GEITje for document classification.
20
+ The fine-tuning is performed using the [AmsterdamBalancedFirst200Tokens](https://huggingface.co/datasets/FemkeBakker/AmsterdamBalancedFirst200Tokens) dataset, which consists of documents truncated to the first 200 tokens.
21
+ In our research, we evaluate the fine-tuning of these LLMs across one, two, and three epochs.
22
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) and has been fine-tuned for one epoch.
23
 
24
  It achieves the following results on the evaluation set:
25
  - Loss: 0.7673
26
 
 
 
 
 
 
 
 
 
27
  ## Training and evaluation data
28
 
29
+ - The training data consists of 9900 documents and their labels formatted into conversations.
30
+ - The evaluation data consists of 1100 documents and their labels formatted into conversations.
31
 
32
  ## Training procedure
33
 
34
+ See the [GitHub](https://github.com/Amsterdam-Internships/document-classification-using-large-language-models) for specifics about the training and the code.
35
+
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
 
57
  | 0.5285 | 0.7952 | 492 | 0.7687 |
58
  | 0.7677 | 0.9939 | 615 | 0.7673 |
59
 
60
+ Training time: it took in total 44 minutes to fine-tuned the model for one epoch.
61
 
62
  ### Framework versions
63
 
 
65
  - Pytorch 2.3.0+cu121
66
  - Datasets 2.19.1
67
  - Tokenizers 0.19.1
68
+
69
+
70
+ ### Acknowledgements
71
+ This model was trained as part of [insert thesis info] in collaboration with Amsterdam Intelligence for the City of Amsterdam.