gofilipa commited on
Commit
39e6979
1 Parent(s): 3ada7d9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -17
README.md CHANGED
@@ -17,21 +17,11 @@ tags: []
17
 
18
  This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
 
20
- - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
- - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
-
28
- ### Model Sources [optional]
29
-
30
- <!-- Provide the basic links for the model. -->
31
-
32
- - **Repository:** [More Information Needed]
33
- - **Paper [optional]:** [More Information Needed]
34
- - **Demo [optional]:** [More Information Needed]
35
 
36
  ## Uses
37
 
@@ -75,11 +65,15 @@ Use the code below to get started with the model.
75
 
76
  ## Training Details
77
 
 
 
 
 
78
  ### Training Data
79
 
80
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
 
82
- [More Information Needed]
83
 
84
  ### Training Procedure
85
 
@@ -89,7 +83,6 @@ Use the code below to get started with the model.
89
 
90
  [More Information Needed]
91
 
92
-
93
  #### Training Hyperparameters
94
 
95
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
 
17
 
18
  This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
 
20
+ - **Developed by:** Filipa Calado
21
+ - **Model type:** Text Generation
22
+ - **Language(s) (NLP):** English
23
+ - **License:** MIT
24
+ - **Finetuned from model:** Mistral 7-b
 
 
 
 
 
 
 
 
 
 
25
 
26
  ## Uses
27
 
 
65
 
66
  ## Training Details
67
 
68
+ num_train_epochs=5
69
+ learning_rate=2e-4
70
+ weight_decay=0.001
71
+
72
  ### Training Data
73
 
74
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
75
 
76
+ [gender_congress_117-118 ](https://huggingface.co/datasets/gofilipa/gender_congress_117-118)
77
 
78
  ### Training Procedure
79
 
 
83
 
84
  [More Information Needed]
85
 
 
86
  #### Training Hyperparameters
87
 
88
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->