Training in progress, step 10
Browse files- 26_10_23_config_test.csv +1 -0
- 26_10_23_results_real_3.csv +1 -1
- 26_10_23_results_test.csv +1 -0
- README.md +13 -0
- adapter_model.bin +1 -1
- adapter_model.safetensors +2 -2
- config.json +18 -18
- training_args.bin +1 -1
26_10_23_config_test.csv
CHANGED
@@ -10,3 +10,4 @@ run_number,comment,peformed_already,num_train_epochs,max_tokens,temperature,stop
|
|
10 |
9,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
|
11 |
10,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
|
12 |
11,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
|
|
|
|
10 |
9,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
|
11 |
10,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
|
12 |
11,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
|
13 |
+
12,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
|
26_10_23_results_real_3.csv
CHANGED
@@ -5,4 +5,4 @@ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnes
|
|
5 |
4,1525.602524101584,classification_of_valuems set True,17.304638385772705,0,0.3449559795118336,0.3423341666457499,0.3223067218438548,0.3962323417858379,0.3934280156800367,0.3705820791169619,0.4480554100301472,0.4472379094634592,0.4191781582462838,0.1924640322240557,0.1864412220397102,0.1791588870738077,0.2322731554369229,0.2249807377601852,0.2130668974234288,0.275524472174082,0.267281230315844,0.2520826707151871,0.25,0.25,0.25
|
6 |
5,1526.1514943613795,lemmatization set True stemming set True classification_of_valuems set True,17.29841375350952,0,0.2740090800292036,0.3255265414032295,0.2775650458916542,0.3242606066397121,0.382074138505741,0.3267769081728278,0.3772870132781871,0.4458722947774239,0.3791647882107381,0.1589738558026723,0.1948121530040918,0.163589425650448,0.197346401556911,0.2400173323974756,0.2012985648757239,0.2417203208425002,0.291225773302693,0.2448890209857868,0.5,0.5,0.5
|
7 |
6,1530.684509829563,lemmatization set True classification_of_valuems set True,17.247185707092285,0,0.2089967577635823,0.2539863482828974,0.2180593819184433,0.2566736342083835,0.3067268601430931,0.2634860133658033,0.3083760408920908,0.3641939637052689,0.3143688188949664,0.1022732513904408,0.1231935696888519,0.1061505173202719,0.1366654941744046,0.1619876288426343,0.1387847257382091,0.1783473709510714,0.2101453928900783,0.1797982054917218,0.4999999999999998,0.4999999999999998,0.4999999999999998
|
8 |
-
|
|
|
5 |
4,1525.602524101584,classification_of_valuems set True,17.304638385772705,0,0.3449559795118336,0.3423341666457499,0.3223067218438548,0.3962323417858379,0.3934280156800367,0.3705820791169619,0.4480554100301472,0.4472379094634592,0.4191781582462838,0.1924640322240557,0.1864412220397102,0.1791588870738077,0.2322731554369229,0.2249807377601852,0.2130668974234288,0.275524472174082,0.267281230315844,0.2520826707151871,0.25,0.25,0.25
|
6 |
5,1526.1514943613795,lemmatization set True stemming set True classification_of_valuems set True,17.29841375350952,0,0.2740090800292036,0.3255265414032295,0.2775650458916542,0.3242606066397121,0.382074138505741,0.3267769081728278,0.3772870132781871,0.4458722947774239,0.3791647882107381,0.1589738558026723,0.1948121530040918,0.163589425650448,0.197346401556911,0.2400173323974756,0.2012985648757239,0.2417203208425002,0.291225773302693,0.2448890209857868,0.5,0.5,0.5
|
7 |
6,1530.684509829563,lemmatization set True classification_of_valuems set True,17.247185707092285,0,0.2089967577635823,0.2539863482828974,0.2180593819184433,0.2566736342083835,0.3067268601430931,0.2634860133658033,0.3083760408920908,0.3641939637052689,0.3143688188949664,0.1022732513904408,0.1231935696888519,0.1061505173202719,0.1366654941744046,0.1619876288426343,0.1387847257382091,0.1783473709510714,0.2101453928900783,0.1797982054917218,0.4999999999999998,0.4999999999999998,0.4999999999999998
|
8 |
+
15,1537.4830854419854,lemmatization set True stemming set True classification_of_valuems set True,17.170920610427856,0,0.3137446610565907,0.346938031739325,0.3116890258549815,0.36716366175121,0.4078239534610949,0.3658593356405938,0.4216956378639542,0.4672338475399563,0.4174884317080075,0.1868318428239036,0.2090565062942769,0.1872856926409339,0.2333035163819085,0.2570762720696271,0.2307891585347144,0.2805117716175197,0.3026827885812185,0.2738993050878683,0.25,0.25,0.25
|
26_10_23_results_test.csv
CHANGED
@@ -13,3 +13,4 @@ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnes
|
|
13 |
0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
|
14 |
0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
|
15 |
0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
|
|
|
|
13 |
0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
|
14 |
0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
|
15 |
0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
|
16 |
+
0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
|
README.md
CHANGED
@@ -976,6 +976,18 @@ The following `bitsandbytes` quantization config was used during training:
|
|
976 |
- bnb_4bit_use_double_quant: True
|
977 |
- bnb_4bit_compute_dtype: bfloat16
|
978 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
979 |
The following `bitsandbytes` quantization config was used during training:
|
980 |
- quant_method: bitsandbytes
|
981 |
- load_in_8bit: False
|
@@ -1070,5 +1082,6 @@ The following `bitsandbytes` quantization config was used during training:
|
|
1070 |
- PEFT 0.5.0
|
1071 |
- PEFT 0.5.0
|
1072 |
- PEFT 0.5.0
|
|
|
1073 |
|
1074 |
- PEFT 0.5.0
|
|
|
976 |
- bnb_4bit_use_double_quant: True
|
977 |
- bnb_4bit_compute_dtype: bfloat16
|
978 |
|
979 |
+
The following `bitsandbytes` quantization config was used during training:
|
980 |
+
- quant_method: bitsandbytes
|
981 |
+
- load_in_8bit: False
|
982 |
+
- load_in_4bit: True
|
983 |
+
- llm_int8_threshold: 6.0
|
984 |
+
- llm_int8_skip_modules: None
|
985 |
+
- llm_int8_enable_fp32_cpu_offload: False
|
986 |
+
- llm_int8_has_fp16_weight: False
|
987 |
+
- bnb_4bit_quant_type: nf4
|
988 |
+
- bnb_4bit_use_double_quant: True
|
989 |
+
- bnb_4bit_compute_dtype: bfloat16
|
990 |
+
|
991 |
The following `bitsandbytes` quantization config was used during training:
|
992 |
- quant_method: bitsandbytes
|
993 |
- load_in_8bit: False
|
|
|
1082 |
- PEFT 0.5.0
|
1083 |
- PEFT 0.5.0
|
1084 |
- PEFT 0.5.0
|
1085 |
+
- PEFT 0.5.0
|
1086 |
|
1087 |
- PEFT 0.5.0
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 261189453
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b536a1d4926ea279ed7f2999a3e87c440f1d8d077c4438f523a01fb41ff2a737
|
3 |
size 261189453
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f619bac221ad47d5e75d145e26ade5d18998b4360842d8357557853adf65ebdf
|
3 |
+
size 261131840
|
config.json
CHANGED
@@ -1,33 +1,33 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "tiiuae/falcon-
|
3 |
-
"alibi":
|
4 |
"apply_residual_connection_post_layernorm": false,
|
5 |
"architectures": [
|
6 |
"FalconForCausalLM"
|
7 |
],
|
8 |
"attention_dropout": 0.0,
|
9 |
"auto_map": {
|
10 |
-
"AutoConfig": "tiiuae/falcon-
|
11 |
-
"AutoModel": "tiiuae/falcon-
|
12 |
-
"AutoModelForCausalLM": "tiiuae/falcon-
|
13 |
-
"AutoModelForQuestionAnswering": "tiiuae/falcon-
|
14 |
-
"AutoModelForSequenceClassification": "tiiuae/falcon-
|
15 |
-
"AutoModelForTokenClassification": "tiiuae/falcon-
|
16 |
},
|
17 |
-
"bias":
|
18 |
-
"bos_token_id":
|
19 |
-
"eos_token_id":
|
20 |
"hidden_dropout": 0.0,
|
21 |
-
"hidden_size":
|
22 |
"initializer_range": 0.02,
|
23 |
"layer_norm_epsilon": 1e-05,
|
24 |
"model_type": "falcon",
|
25 |
-
"multi_query":
|
26 |
"new_decoder_architecture": false,
|
27 |
-
"num_attention_heads":
|
28 |
-
"num_hidden_layers":
|
29 |
-
"num_kv_heads":
|
30 |
-
"parallel_attn":
|
31 |
"quantization_config": {
|
32 |
"bnb_4bit_compute_dtype": "bfloat16",
|
33 |
"bnb_4bit_quant_type": "nf4",
|
@@ -43,5 +43,5 @@
|
|
43 |
"torch_dtype": "bfloat16",
|
44 |
"transformers_version": "4.34.1",
|
45 |
"use_cache": false,
|
46 |
-
"vocab_size":
|
47 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "tiiuae/falcon-7b",
|
3 |
+
"alibi": false,
|
4 |
"apply_residual_connection_post_layernorm": false,
|
5 |
"architectures": [
|
6 |
"FalconForCausalLM"
|
7 |
],
|
8 |
"attention_dropout": 0.0,
|
9 |
"auto_map": {
|
10 |
+
"AutoConfig": "tiiuae/falcon-7b--configuration_falcon.FalconConfig",
|
11 |
+
"AutoModel": "tiiuae/falcon-7b--modeling_falcon.FalconModel",
|
12 |
+
"AutoModelForCausalLM": "tiiuae/falcon-7b--modeling_falcon.FalconForCausalLM",
|
13 |
+
"AutoModelForQuestionAnswering": "tiiuae/falcon-7b--modeling_falcon.FalconForQuestionAnswering",
|
14 |
+
"AutoModelForSequenceClassification": "tiiuae/falcon-7b--modeling_falcon.FalconForSequenceClassification",
|
15 |
+
"AutoModelForTokenClassification": "tiiuae/falcon-7b--modeling_falcon.FalconForTokenClassification"
|
16 |
},
|
17 |
+
"bias": false,
|
18 |
+
"bos_token_id": 11,
|
19 |
+
"eos_token_id": 11,
|
20 |
"hidden_dropout": 0.0,
|
21 |
+
"hidden_size": 4544,
|
22 |
"initializer_range": 0.02,
|
23 |
"layer_norm_epsilon": 1e-05,
|
24 |
"model_type": "falcon",
|
25 |
+
"multi_query": true,
|
26 |
"new_decoder_architecture": false,
|
27 |
+
"num_attention_heads": 71,
|
28 |
+
"num_hidden_layers": 32,
|
29 |
+
"num_kv_heads": 71,
|
30 |
+
"parallel_attn": true,
|
31 |
"quantization_config": {
|
32 |
"bnb_4bit_compute_dtype": "bfloat16",
|
33 |
"bnb_4bit_quant_type": "nf4",
|
|
|
43 |
"torch_dtype": "bfloat16",
|
44 |
"transformers_version": "4.34.1",
|
45 |
"use_cache": false,
|
46 |
+
"vocab_size": 65024
|
47 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4283
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:17e1d6ea179881703aebc9cf5ee9455199d40dcc06ec43acb88843d785a2245b
|
3 |
size 4283
|