deuswoof commited on
Commit
5363704
1 Parent(s): a894c6f

Training in progress, step 10

Browse files
26_10_23_config_test.csv CHANGED
@@ -1,2 +1,3 @@
1
  run_number,comment,peformed_already,num_train_epochs,max_tokens,temperature,stop_token,classification_of_valuems,stemming,lemmatization
2
- 1,no variations,False,2,100,0.8,False,False,False,False
 
 
1
  run_number,comment,peformed_already,num_train_epochs,max_tokens,temperature,stop_token,classification_of_valuems,stemming,lemmatization
2
+ 1,no variations,True,2,100,0.8,False,False,False,False
3
+ 2,classification_of_valuems set False,False,2,100,0.8,False,False,False,False
26_10_23_results_real_3.csv CHANGED
@@ -5,4 +5,4 @@ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnes
5
  4,1525.602524101584,classification_of_valuems set True,17.304638385772705,0,0.3449559795118336,0.3423341666457499,0.3223067218438548,0.3962323417858379,0.3934280156800367,0.3705820791169619,0.4480554100301472,0.4472379094634592,0.4191781582462838,0.1924640322240557,0.1864412220397102,0.1791588870738077,0.2322731554369229,0.2249807377601852,0.2130668974234288,0.275524472174082,0.267281230315844,0.2520826707151871,0.25,0.25,0.25
6
  5,1526.1514943613795,lemmatization set True stemming set True classification_of_valuems set True,17.29841375350952,0,0.2740090800292036,0.3255265414032295,0.2775650458916542,0.3242606066397121,0.382074138505741,0.3267769081728278,0.3772870132781871,0.4458722947774239,0.3791647882107381,0.1589738558026723,0.1948121530040918,0.163589425650448,0.197346401556911,0.2400173323974756,0.2012985648757239,0.2417203208425002,0.291225773302693,0.2448890209857868,0.5,0.5,0.5
7
  6,1530.684509829563,lemmatization set True classification_of_valuems set True,17.247185707092285,0,0.2089967577635823,0.2539863482828974,0.2180593819184433,0.2566736342083835,0.3067268601430931,0.2634860133658033,0.3083760408920908,0.3641939637052689,0.3143688188949664,0.1022732513904408,0.1231935696888519,0.1061505173202719,0.1366654941744046,0.1619876288426343,0.1387847257382091,0.1783473709510714,0.2101453928900783,0.1797982054917218,0.4999999999999998,0.4999999999999998,0.4999999999999998
8
- 3,1537.4830854419854,lemmatization set True stemming set True classification_of_valuems set True,17.170920610427856,0,0.3137446610565907,0.346938031739325,0.3116890258549815,0.36716366175121,0.4078239534610949,0.3658593356405938,0.4216956378639542,0.4672338475399563,0.4174884317080075,0.1868318428239036,0.2090565062942769,0.1872856926409339,0.2333035163819085,0.2570762720696271,0.2307891585347144,0.2805117716175197,0.3026827885812185,0.2738993050878683,0.25,0.25,0.25
 
5
  4,1525.602524101584,classification_of_valuems set True,17.304638385772705,0,0.3449559795118336,0.3423341666457499,0.3223067218438548,0.3962323417858379,0.3934280156800367,0.3705820791169619,0.4480554100301472,0.4472379094634592,0.4191781582462838,0.1924640322240557,0.1864412220397102,0.1791588870738077,0.2322731554369229,0.2249807377601852,0.2130668974234288,0.275524472174082,0.267281230315844,0.2520826707151871,0.25,0.25,0.25
6
  5,1526.1514943613795,lemmatization set True stemming set True classification_of_valuems set True,17.29841375350952,0,0.2740090800292036,0.3255265414032295,0.2775650458916542,0.3242606066397121,0.382074138505741,0.3267769081728278,0.3772870132781871,0.4458722947774239,0.3791647882107381,0.1589738558026723,0.1948121530040918,0.163589425650448,0.197346401556911,0.2400173323974756,0.2012985648757239,0.2417203208425002,0.291225773302693,0.2448890209857868,0.5,0.5,0.5
7
  6,1530.684509829563,lemmatization set True classification_of_valuems set True,17.247185707092285,0,0.2089967577635823,0.2539863482828974,0.2180593819184433,0.2566736342083835,0.3067268601430931,0.2634860133658033,0.3083760408920908,0.3641939637052689,0.3143688188949664,0.1022732513904408,0.1231935696888519,0.1061505173202719,0.1366654941744046,0.1619876288426343,0.1387847257382091,0.1783473709510714,0.2101453928900783,0.1797982054917218,0.4999999999999998,0.4999999999999998,0.4999999999999998
8
+ 4,1537.4830854419854,lemmatization set True stemming set True classification_of_valuems set True,17.170920610427856,0,0.3137446610565907,0.346938031739325,0.3116890258549815,0.36716366175121,0.4078239534610949,0.3658593356405938,0.4216956378639542,0.4672338475399563,0.4174884317080075,0.1868318428239036,0.2090565062942769,0.1872856926409339,0.2333035163819085,0.2570762720696271,0.2307891585347144,0.2805117716175197,0.3026827885812185,0.2738993050878683,0.25,0.25,0.25
26_10_23_results_test.csv CHANGED
@@ -1,4 +1,5 @@
1
  run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnest,rouge1 low Precision,rouge1 low Recall,rouge1 low F1 Score,rouge1 mid Precision,rouge1 mid Recall,rouge1 mid F1 Score,rouge1 high Precision,rouge1 high Recall,rouge1 high F1 Score,rouge2 low Precision,rouge2 low Recall,rouge2 low F1 Score,rouge2 mid Precision,rouge2 mid Recall,rouge2 mid F1 Score,rouge2 high Precision,rouge2 high Recall,rouge2 high F1 Score,min_cosine_sim_value,max_cosine_sim_value,mean_cosine_sim_value
2
  0,1525.78169293058,0,17.302606344223022,0,0.2208788951005069,0.2975883615881153,0.240614991862749,0.2581963155823283,0.3380846243794819,0.2741873775462263,0.2975950054307982,0.3786356612216308,0.3105477853693633,0.0725478275967877,0.0914455107483677,0.076518173971504,0.0912179928867811,0.1126082896323259,0.0948303028773734,0.1141810282401093,0.1365525195957442,0.1172672123671381,0.25,0.25,0.25
3
  0,1526.9601087665926,0,17.28925323486328,0,0.211648928940661,0.3099076516941764,0.2349623655451098,0.2510485488340126,0.3427790236432527,0.2697717126098345,0.2921795539464257,0.3779250040331974,0.3024802842684405,0.0681760966648744,0.0863483981349806,0.0714061957144911,0.0882483532156721,0.1067406117807395,0.0896180979585989,0.1141695115472147,0.1327745046315235,0.1126574769723024,0.5022742922973714,0.5022742922973714,0.5022742922973714
 
4
  0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
 
1
  run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnest,rouge1 low Precision,rouge1 low Recall,rouge1 low F1 Score,rouge1 mid Precision,rouge1 mid Recall,rouge1 mid F1 Score,rouge1 high Precision,rouge1 high Recall,rouge1 high F1 Score,rouge2 low Precision,rouge2 low Recall,rouge2 low F1 Score,rouge2 mid Precision,rouge2 mid Recall,rouge2 mid F1 Score,rouge2 high Precision,rouge2 high Recall,rouge2 high F1 Score,min_cosine_sim_value,max_cosine_sim_value,mean_cosine_sim_value
2
  0,1525.78169293058,0,17.302606344223022,0,0.2208788951005069,0.2975883615881153,0.240614991862749,0.2581963155823283,0.3380846243794819,0.2741873775462263,0.2975950054307982,0.3786356612216308,0.3105477853693633,0.0725478275967877,0.0914455107483677,0.076518173971504,0.0912179928867811,0.1126082896323259,0.0948303028773734,0.1141810282401093,0.1365525195957442,0.1172672123671381,0.25,0.25,0.25
3
  0,1526.9601087665926,0,17.28925323486328,0,0.211648928940661,0.3099076516941764,0.2349623655451098,0.2510485488340126,0.3427790236432527,0.2697717126098345,0.2921795539464257,0.3779250040331974,0.3024802842684405,0.0681760966648744,0.0863483981349806,0.0714061957144911,0.0882483532156721,0.1067406117807395,0.0896180979585989,0.1141695115472147,0.1327745046315235,0.1126574769723024,0.5022742922973714,0.5022742922973714,0.5022742922973714
4
+ 0,1525.1505741778146,no variations,17.30976629257202,0,0.1944833026292195,0.2608866277179161,0.212506973427488,0.231774299851333,0.2992392908143453,0.2475690501161298,0.2685696129851041,0.3381393563030371,0.2834733235474062,0.0626355093279982,0.0747762221568468,0.0659507681155903,0.0809426639328388,0.095104133215893,0.0838846169340258,0.1020442556026316,0.1176611208472423,0.1049781579164324,0.25,0.25,0.25
5
  0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
README.md CHANGED
@@ -892,6 +892,18 @@ The following `bitsandbytes` quantization config was used during training:
892
  - bnb_4bit_use_double_quant: True
893
  - bnb_4bit_compute_dtype: bfloat16
894
 
 
 
 
 
 
 
 
 
 
 
 
 
895
  The following `bitsandbytes` quantization config was used during training:
896
  - quant_method: bitsandbytes
897
  - load_in_8bit: False
@@ -979,5 +991,6 @@ The following `bitsandbytes` quantization config was used during training:
979
  - PEFT 0.5.0
980
  - PEFT 0.5.0
981
  - PEFT 0.5.0
 
982
 
983
  - PEFT 0.5.0
 
892
  - bnb_4bit_use_double_quant: True
893
  - bnb_4bit_compute_dtype: bfloat16
894
 
895
+ The following `bitsandbytes` quantization config was used during training:
896
+ - quant_method: bitsandbytes
897
+ - load_in_8bit: False
898
+ - load_in_4bit: True
899
+ - llm_int8_threshold: 6.0
900
+ - llm_int8_skip_modules: None
901
+ - llm_int8_enable_fp32_cpu_offload: False
902
+ - llm_int8_has_fp16_weight: False
903
+ - bnb_4bit_quant_type: nf4
904
+ - bnb_4bit_use_double_quant: True
905
+ - bnb_4bit_compute_dtype: bfloat16
906
+
907
  The following `bitsandbytes` quantization config was used during training:
908
  - quant_method: bitsandbytes
909
  - load_in_8bit: False
 
991
  - PEFT 0.5.0
992
  - PEFT 0.5.0
993
  - PEFT 0.5.0
994
+ - PEFT 0.5.0
995
 
996
  - PEFT 0.5.0
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e638a4d5db158c508749bd6c4ec2e85f316d825e1068083feef25f919420958f
3
  size 100733709
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcfe516cb5a52b72c7c32afd3c11ddd090b223a575825c5e44ad753d3db383cf
3
  size 100733709
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:066d0b012cf3f51ec20706e7ec4e23c31e3b6b0c4b12dfc3456cb6d68f013342
3
  size 100690288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96a1d2e8216dd86029687d2b112ff3cedcb3e524a37fc1a116218a361a0e5ed6
3
  size 100690288
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c7cec4233cd9e93442395c1972ccf4fe9a23b1546751ca7e00e90c709a5d126
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c845e51951d5c38f507bd5115b6682f736e02af517bf764c255e44bb05fef53c
3
  size 4283