bobox commited on
Commit
3acf097
1 Parent(s): 2e0c9a0

Training in progress, step 610, checkpoint

Browse files
checkpoint-610/1_AdvancedWeightedPooling/config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "embed_dim": 768,
3
+ "num_heads": 4,
4
+ "dropout": 0.025,
5
+ "bias": true,
6
+ "gate_min": 0.05,
7
+ "gate_max": 0.95,
8
+ "gate_dropout": 0.05,
9
+ "dropout_gate_open": 0.0,
10
+ "dropout_gate_close": 0.0,
11
+ "CLS_self_attn": 0
12
+ }
checkpoint-610/1_AdvancedWeightedPooling/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4556fb69701933e8a2c4d45b80ba1daa3867c8ef11c47e1d5f77fa56f1dffe7a
3
+ size 14201307
checkpoint-610/README.md ADDED
@@ -0,0 +1,1437 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/deberta-v3-small
3
+ library_name: sentence-transformers
4
+ metrics:
5
+ - pearson_cosine
6
+ - spearman_cosine
7
+ - pearson_manhattan
8
+ - spearman_manhattan
9
+ - pearson_euclidean
10
+ - spearman_euclidean
11
+ - pearson_dot
12
+ - spearman_dot
13
+ - pearson_max
14
+ - spearman_max
15
+ - cosine_accuracy
16
+ - cosine_accuracy_threshold
17
+ - cosine_f1
18
+ - cosine_f1_threshold
19
+ - cosine_precision
20
+ - cosine_recall
21
+ - cosine_ap
22
+ - dot_accuracy
23
+ - dot_accuracy_threshold
24
+ - dot_f1
25
+ - dot_f1_threshold
26
+ - dot_precision
27
+ - dot_recall
28
+ - dot_ap
29
+ - manhattan_accuracy
30
+ - manhattan_accuracy_threshold
31
+ - manhattan_f1
32
+ - manhattan_f1_threshold
33
+ - manhattan_precision
34
+ - manhattan_recall
35
+ - manhattan_ap
36
+ - euclidean_accuracy
37
+ - euclidean_accuracy_threshold
38
+ - euclidean_f1
39
+ - euclidean_f1_threshold
40
+ - euclidean_precision
41
+ - euclidean_recall
42
+ - euclidean_ap
43
+ - max_accuracy
44
+ - max_accuracy_threshold
45
+ - max_f1
46
+ - max_f1_threshold
47
+ - max_precision
48
+ - max_recall
49
+ - max_ap
50
+ pipeline_tag: sentence-similarity
51
+ tags:
52
+ - sentence-transformers
53
+ - sentence-similarity
54
+ - feature-extraction
55
+ - generated_from_trainer
56
+ - dataset_size:32500
57
+ - loss:GISTEmbedLoss
58
+ widget:
59
+ - source_sentence: phase changes do not change
60
+ sentences:
61
+ - The major Atlantic slave trading nations, ordered by trade volume, were the Portuguese,
62
+ the British, the Spanish, the French, the Dutch, and the Danish. Several had established
63
+ outposts on the African coast where they purchased slaves from local African leaders.
64
+ - "phase changes do not change mass. Particles have mass, but mass is energy. \n\
65
+ \ phase changes do not change energy"
66
+ - According to the U.S. Census Bureau , the county is a total area of , which has
67
+ land and ( 0.2 % ) is water .
68
+ - source_sentence: what jobs can you get with a bachelor degree in anthropology?
69
+ sentences:
70
+ - To determine the atomic weight of an element, you should add up protons and neutrons.
71
+ - '[''Paleontologist*'', ''Archaeologist*'', ''University Professor*'', ''Market
72
+ Research Analyst*'', ''Primatologist.'', ''Forensic Scientist*'', ''Medical Anthropologist.'',
73
+ ''Museum Technician.'']'
74
+ - The wingspan flies , the moth comes depending on the location from July to August
75
+ .
76
+ - source_sentence: Identify different forms of energy (e.g., light, sound, heat).
77
+ sentences:
78
+ - '`` Irreplaceable '''' '''' remained on the chart for thirty weeks , and was certified
79
+ double-platinum by the Recording Industry Association of America ( RIAA ) , denoting
80
+ sales of two million downloads , and had sold over 3,139,000 paid digital downloads
81
+ in the US as of October 2012 , according to Nielsen SoundScan . '''''
82
+ - On Rotten Tomatoes , the film has a rating of 63 % , based on 87 reviews , with
83
+ an average rating of 5.9/10 .
84
+ - Heat, light, and sound are all different forms of energy.
85
+ - source_sentence: what is so small it can only be seen with an electron microscope?
86
+ sentences:
87
+ - "Viruses are so small that they can be seen only with an electron microscope..\
88
+ \ Where most viruses are DNA, HIV is an RNA virus. \n HIV is so small it can only\
89
+ \ be seen with an electron microscope"
90
+ - The development of modern lasers has opened many doors to both research and applications.
91
+ A laser beam was used to measure the distance from the Earth to the moon. Lasers
92
+ are important components of CD players. As the image above illustrates, lasers
93
+ can provide precise focusing of beams to selectively destroy cancer cells in patients.
94
+ The ability of a laser to focus precisely is due to high-quality crystals that
95
+ help give rise to the laser beam. A variety of techniques are used to manufacture
96
+ pure crystals for use in lasers.
97
+ - Discussion for (a) This value is the net work done on the package. The person
98
+ actually does more work than this, because friction opposes the motion. Friction
99
+ does negative work and removes some of the energy the person expends and converts
100
+ it to thermal energy. The net work equals the sum of the work done by each individual
101
+ force. Strategy and Concept for (b) The forces acting on the package are gravity,
102
+ the normal force, the force of friction, and the applied force. The normal force
103
+ and force of gravity are each perpendicular to the displacement, and therefore
104
+ do no work. Solution for (b) The applied force does work.
105
+ - source_sentence: what aspects of your environment may relate to the epidemic of
106
+ obesity
107
+ sentences:
108
+ - Jan Kromkamp ( born August 17 , 1980 in Makkinga , Netherlands ) is a Dutch footballer
109
+ .
110
+ - When chemicals in solution react, the proper way of writing the chemical formulas
111
+ of the dissolved ionic compounds is in terms of the dissociated ions, not the
112
+ complete ionic formula. A complete ionic equation is a chemical equation in which
113
+ the dissolved ionic compounds are written as separated ions. Solubility rules
114
+ are very useful in determining which ionic compounds are dissolved and which are
115
+ not. For example, when NaCl(aq) reacts with AgNO3(aq) in a double-replacement
116
+ reaction to precipitate AgCl(s) and form NaNO3(aq), the complete ionic equation
117
+ includes NaCl, AgNO3, and NaNO3 written as separated ions:.
118
+ - Genetic changes in human populations occur too slowly to be responsible for the
119
+ obesity epidemic. Nevertheless, the variation in how people respond to the environment
120
+ that promotes physical inactivity and intake of high-calorie foods suggests that
121
+ genes do play a role in the development of obesity.
122
+ model-index:
123
+ - name: SentenceTransformer based on microsoft/deberta-v3-small
124
+ results:
125
+ - task:
126
+ type: semantic-similarity
127
+ name: Semantic Similarity
128
+ dataset:
129
+ name: sts test
130
+ type: sts-test
131
+ metrics:
132
+ - type: pearson_cosine
133
+ value: 0.5377382226514003
134
+ name: Pearson Cosine
135
+ - type: spearman_cosine
136
+ value: 0.5410237309359288
137
+ name: Spearman Cosine
138
+ - type: pearson_manhattan
139
+ value: 0.5464293120330461
140
+ name: Pearson Manhattan
141
+ - type: spearman_manhattan
142
+ value: 0.5401021234588343
143
+ name: Spearman Manhattan
144
+ - type: pearson_euclidean
145
+ value: 0.5469897917607747
146
+ name: Pearson Euclidean
147
+ - type: spearman_euclidean
148
+ value: 0.5409800984560722
149
+ name: Spearman Euclidean
150
+ - type: pearson_dot
151
+ value: 0.5376496659087263
152
+ name: Pearson Dot
153
+ - type: spearman_dot
154
+ value: 0.5408811086658744
155
+ name: Spearman Dot
156
+ - type: pearson_max
157
+ value: 0.5469897917607747
158
+ name: Pearson Max
159
+ - type: spearman_max
160
+ value: 0.5410237309359288
161
+ name: Spearman Max
162
+ - task:
163
+ type: binary-classification
164
+ name: Binary Classification
165
+ dataset:
166
+ name: allNLI dev
167
+ type: allNLI-dev
168
+ metrics:
169
+ - type: cosine_accuracy
170
+ value: 0.68359375
171
+ name: Cosine Accuracy
172
+ - type: cosine_accuracy_threshold
173
+ value: 0.9088386297225952
174
+ name: Cosine Accuracy Threshold
175
+ - type: cosine_f1
176
+ value: 0.5350553505535056
177
+ name: Cosine F1
178
+ - type: cosine_f1_threshold
179
+ value: 0.8140230178833008
180
+ name: Cosine F1 Threshold
181
+ - type: cosine_precision
182
+ value: 0.39295392953929537
183
+ name: Cosine Precision
184
+ - type: cosine_recall
185
+ value: 0.838150289017341
186
+ name: Cosine Recall
187
+ - type: cosine_ap
188
+ value: 0.48873606015680937
189
+ name: Cosine Ap
190
+ - type: dot_accuracy
191
+ value: 0.68359375
192
+ name: Dot Accuracy
193
+ - type: dot_accuracy_threshold
194
+ value: 699.0950927734375
195
+ name: Dot Accuracy Threshold
196
+ - type: dot_f1
197
+ value: 0.5350553505535056
198
+ name: Dot F1
199
+ - type: dot_f1_threshold
200
+ value: 625.3240356445312
201
+ name: Dot F1 Threshold
202
+ - type: dot_precision
203
+ value: 0.39295392953929537
204
+ name: Dot Precision
205
+ - type: dot_recall
206
+ value: 0.838150289017341
207
+ name: Dot Recall
208
+ - type: dot_ap
209
+ value: 0.48885724782911755
210
+ name: Dot Ap
211
+ - type: manhattan_accuracy
212
+ value: 0.68359375
213
+ name: Manhattan Accuracy
214
+ - type: manhattan_accuracy_threshold
215
+ value: 256.45477294921875
216
+ name: Manhattan Accuracy Threshold
217
+ - type: manhattan_f1
218
+ value: 0.5396145610278372
219
+ name: Manhattan F1
220
+ - type: manhattan_f1_threshold
221
+ value: 339.225830078125
222
+ name: Manhattan F1 Threshold
223
+ - type: manhattan_precision
224
+ value: 0.42857142857142855
225
+ name: Manhattan Precision
226
+ - type: manhattan_recall
227
+ value: 0.7283236994219653
228
+ name: Manhattan Recall
229
+ - type: manhattan_ap
230
+ value: 0.4920209563997524
231
+ name: Manhattan Ap
232
+ - type: euclidean_accuracy
233
+ value: 0.68359375
234
+ name: Euclidean Accuracy
235
+ - type: euclidean_accuracy_threshold
236
+ value: 11.834823608398438
237
+ name: Euclidean Accuracy Threshold
238
+ - type: euclidean_f1
239
+ value: 0.5350553505535056
240
+ name: Euclidean F1
241
+ - type: euclidean_f1_threshold
242
+ value: 16.90357780456543
243
+ name: Euclidean F1 Threshold
244
+ - type: euclidean_precision
245
+ value: 0.39295392953929537
246
+ name: Euclidean Precision
247
+ - type: euclidean_recall
248
+ value: 0.838150289017341
249
+ name: Euclidean Recall
250
+ - type: euclidean_ap
251
+ value: 0.4887203371983184
252
+ name: Euclidean Ap
253
+ - type: max_accuracy
254
+ value: 0.68359375
255
+ name: Max Accuracy
256
+ - type: max_accuracy_threshold
257
+ value: 699.0950927734375
258
+ name: Max Accuracy Threshold
259
+ - type: max_f1
260
+ value: 0.5396145610278372
261
+ name: Max F1
262
+ - type: max_f1_threshold
263
+ value: 625.3240356445312
264
+ name: Max F1 Threshold
265
+ - type: max_precision
266
+ value: 0.42857142857142855
267
+ name: Max Precision
268
+ - type: max_recall
269
+ value: 0.838150289017341
270
+ name: Max Recall
271
+ - type: max_ap
272
+ value: 0.4920209563997524
273
+ name: Max Ap
274
+ - task:
275
+ type: binary-classification
276
+ name: Binary Classification
277
+ dataset:
278
+ name: Qnli dev
279
+ type: Qnli-dev
280
+ metrics:
281
+ - type: cosine_accuracy
282
+ value: 0.693359375
283
+ name: Cosine Accuracy
284
+ - type: cosine_accuracy_threshold
285
+ value: 0.8319265842437744
286
+ name: Cosine Accuracy Threshold
287
+ - type: cosine_f1
288
+ value: 0.685337726523888
289
+ name: Cosine F1
290
+ - type: cosine_f1_threshold
291
+ value: 0.74552983045578
292
+ name: Cosine F1 Threshold
293
+ - type: cosine_precision
294
+ value: 0.5606469002695418
295
+ name: Cosine Precision
296
+ - type: cosine_recall
297
+ value: 0.8813559322033898
298
+ name: Cosine Recall
299
+ - type: cosine_ap
300
+ value: 0.6873625888187367
301
+ name: Cosine Ap
302
+ - type: dot_accuracy
303
+ value: 0.693359375
304
+ name: Dot Accuracy
305
+ - type: dot_accuracy_threshold
306
+ value: 639.0776977539062
307
+ name: Dot Accuracy Threshold
308
+ - type: dot_f1
309
+ value: 0.685337726523888
310
+ name: Dot F1
311
+ - type: dot_f1_threshold
312
+ value: 572.7136840820312
313
+ name: Dot F1 Threshold
314
+ - type: dot_precision
315
+ value: 0.5606469002695418
316
+ name: Dot Precision
317
+ - type: dot_recall
318
+ value: 0.8813559322033898
319
+ name: Dot Recall
320
+ - type: dot_ap
321
+ value: 0.6878718449643791
322
+ name: Dot Ap
323
+ - type: manhattan_accuracy
324
+ value: 0.69921875
325
+ name: Manhattan Accuracy
326
+ - type: manhattan_accuracy_threshold
327
+ value: 362.1485900878906
328
+ name: Manhattan Accuracy Threshold
329
+ - type: manhattan_f1
330
+ value: 0.6857142857142857
331
+ name: Manhattan F1
332
+ - type: manhattan_f1_threshold
333
+ value: 430.38519287109375
334
+ name: Manhattan F1 Threshold
335
+ - type: manhattan_precision
336
+ value: 0.5682451253481894
337
+ name: Manhattan Precision
338
+ - type: manhattan_recall
339
+ value: 0.864406779661017
340
+ name: Manhattan Recall
341
+ - type: manhattan_ap
342
+ value: 0.6874910715870401
343
+ name: Manhattan Ap
344
+ - type: euclidean_accuracy
345
+ value: 0.693359375
346
+ name: Euclidean Accuracy
347
+ - type: euclidean_accuracy_threshold
348
+ value: 16.06937026977539
349
+ name: Euclidean Accuracy Threshold
350
+ - type: euclidean_f1
351
+ value: 0.685337726523888
352
+ name: Euclidean F1
353
+ - type: euclidean_f1_threshold
354
+ value: 19.772865295410156
355
+ name: Euclidean F1 Threshold
356
+ - type: euclidean_precision
357
+ value: 0.5606469002695418
358
+ name: Euclidean Precision
359
+ - type: euclidean_recall
360
+ value: 0.8813559322033898
361
+ name: Euclidean Recall
362
+ - type: euclidean_ap
363
+ value: 0.6873686687008952
364
+ name: Euclidean Ap
365
+ - type: max_accuracy
366
+ value: 0.69921875
367
+ name: Max Accuracy
368
+ - type: max_accuracy_threshold
369
+ value: 639.0776977539062
370
+ name: Max Accuracy Threshold
371
+ - type: max_f1
372
+ value: 0.6857142857142857
373
+ name: Max F1
374
+ - type: max_f1_threshold
375
+ value: 572.7136840820312
376
+ name: Max F1 Threshold
377
+ - type: max_precision
378
+ value: 0.5682451253481894
379
+ name: Max Precision
380
+ - type: max_recall
381
+ value: 0.8813559322033898
382
+ name: Max Recall
383
+ - type: max_ap
384
+ value: 0.6878718449643791
385
+ name: Max Ap
386
+ ---
387
+
388
+ # SentenceTransformer based on microsoft/deberta-v3-small
389
+
390
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
391
+
392
+ ## Model Details
393
+
394
+ ### Model Description
395
+ - **Model Type:** Sentence Transformer
396
+ - **Base model:** [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) <!-- at revision a36c739020e01763fe789b4b85e2df55d6180012 -->
397
+ - **Maximum Sequence Length:** 512 tokens
398
+ - **Output Dimensionality:** 768 tokens
399
+ - **Similarity Function:** Cosine Similarity
400
+ <!-- - **Training Dataset:** Unknown -->
401
+ <!-- - **Language:** Unknown -->
402
+ <!-- - **License:** Unknown -->
403
+
404
+ ### Model Sources
405
+
406
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
407
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
408
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
409
+
410
+ ### Full Model Architecture
411
+
412
+ ```
413
+ SentenceTransformer(
414
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: DebertaV2Model
415
+ (1): AdvancedWeightedPooling(
416
+ (alpha_dropout_layer): Dropout(p=0.05, inplace=False)
417
+ (gate_dropout_layer): Dropout(p=0.0, inplace=False)
418
+ (linear_cls_Qpj): Linear(in_features=768, out_features=768, bias=True)
419
+ (linear_attnOut): Linear(in_features=768, out_features=768, bias=True)
420
+ (mha): MultiheadAttention(
421
+ (out_proj): NonDynamicallyQuantizableLinear(in_features=768, out_features=768, bias=True)
422
+ )
423
+ (layernorm_output): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
424
+ (layernorm_weightedPooing): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
425
+ (layernorm_attnOut): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
426
+ )
427
+ )
428
+ ```
429
+
430
+ ## Usage
431
+
432
+ ### Direct Usage (Sentence Transformers)
433
+
434
+ First install the Sentence Transformers library:
435
+
436
+ ```bash
437
+ pip install -U sentence-transformers
438
+ ```
439
+
440
+ Then you can load this model and run inference.
441
+ ```python
442
+ from sentence_transformers import SentenceTransformer
443
+
444
+ # Download from the 🤗 Hub
445
+ model = SentenceTransformer("bobox/DeBERTa3-s-CustomPoolin-toytest4-step1-checkpoints-tmp")
446
+ # Run inference
447
+ sentences = [
448
+ 'what aspects of your environment may relate to the epidemic of obesity',
449
+ 'Genetic changes in human populations occur too slowly to be responsible for the obesity epidemic. Nevertheless, the variation in how people respond to the environment that promotes physical inactivity and intake of high-calorie foods suggests that genes do play a role in the development of obesity.',
450
+ 'When chemicals in solution react, the proper way of writing the chemical formulas of the dissolved ionic compounds is in terms of the dissociated ions, not the complete ionic formula. A complete ionic equation is a chemical equation in which the dissolved ionic compounds are written as separated ions. Solubility rules are very useful in determining which ionic compounds are dissolved and which are not. For example, when NaCl(aq) reacts with AgNO3(aq) in a double-replacement reaction to precipitate AgCl(s) and form NaNO3(aq), the complete ionic equation includes NaCl, AgNO3, and NaNO3 written as separated ions:.',
451
+ ]
452
+ embeddings = model.encode(sentences)
453
+ print(embeddings.shape)
454
+ # [3, 768]
455
+
456
+ # Get the similarity scores for the embeddings
457
+ similarities = model.similarity(embeddings, embeddings)
458
+ print(similarities.shape)
459
+ # [3, 3]
460
+ ```
461
+
462
+ <!--
463
+ ### Direct Usage (Transformers)
464
+
465
+ <details><summary>Click to see the direct usage in Transformers</summary>
466
+
467
+ </details>
468
+ -->
469
+
470
+ <!--
471
+ ### Downstream Usage (Sentence Transformers)
472
+
473
+ You can finetune this model on your own dataset.
474
+
475
+ <details><summary>Click to expand</summary>
476
+
477
+ </details>
478
+ -->
479
+
480
+ <!--
481
+ ### Out-of-Scope Use
482
+
483
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
484
+ -->
485
+
486
+ ## Evaluation
487
+
488
+ ### Metrics
489
+
490
+ #### Semantic Similarity
491
+ * Dataset: `sts-test`
492
+ * Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
493
+
494
+ | Metric | Value |
495
+ |:--------------------|:----------|
496
+ | pearson_cosine | 0.5377 |
497
+ | **spearman_cosine** | **0.541** |
498
+ | pearson_manhattan | 0.5464 |
499
+ | spearman_manhattan | 0.5401 |
500
+ | pearson_euclidean | 0.547 |
501
+ | spearman_euclidean | 0.541 |
502
+ | pearson_dot | 0.5376 |
503
+ | spearman_dot | 0.5409 |
504
+ | pearson_max | 0.547 |
505
+ | spearman_max | 0.541 |
506
+
507
+ #### Binary Classification
508
+ * Dataset: `allNLI-dev`
509
+ * Evaluated with [<code>BinaryClassificationEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator)
510
+
511
+ | Metric | Value |
512
+ |:-----------------------------|:----------|
513
+ | cosine_accuracy | 0.6836 |
514
+ | cosine_accuracy_threshold | 0.9088 |
515
+ | cosine_f1 | 0.5351 |
516
+ | cosine_f1_threshold | 0.814 |
517
+ | cosine_precision | 0.393 |
518
+ | cosine_recall | 0.8382 |
519
+ | cosine_ap | 0.4887 |
520
+ | dot_accuracy | 0.6836 |
521
+ | dot_accuracy_threshold | 699.0951 |
522
+ | dot_f1 | 0.5351 |
523
+ | dot_f1_threshold | 625.324 |
524
+ | dot_precision | 0.393 |
525
+ | dot_recall | 0.8382 |
526
+ | dot_ap | 0.4889 |
527
+ | manhattan_accuracy | 0.6836 |
528
+ | manhattan_accuracy_threshold | 256.4548 |
529
+ | manhattan_f1 | 0.5396 |
530
+ | manhattan_f1_threshold | 339.2258 |
531
+ | manhattan_precision | 0.4286 |
532
+ | manhattan_recall | 0.7283 |
533
+ | manhattan_ap | 0.492 |
534
+ | euclidean_accuracy | 0.6836 |
535
+ | euclidean_accuracy_threshold | 11.8348 |
536
+ | euclidean_f1 | 0.5351 |
537
+ | euclidean_f1_threshold | 16.9036 |
538
+ | euclidean_precision | 0.393 |
539
+ | euclidean_recall | 0.8382 |
540
+ | euclidean_ap | 0.4887 |
541
+ | max_accuracy | 0.6836 |
542
+ | max_accuracy_threshold | 699.0951 |
543
+ | max_f1 | 0.5396 |
544
+ | max_f1_threshold | 625.324 |
545
+ | max_precision | 0.4286 |
546
+ | max_recall | 0.8382 |
547
+ | **max_ap** | **0.492** |
548
+
549
+ #### Binary Classification
550
+ * Dataset: `Qnli-dev`
551
+ * Evaluated with [<code>BinaryClassificationEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator)
552
+
553
+ | Metric | Value |
554
+ |:-----------------------------|:-----------|
555
+ | cosine_accuracy | 0.6934 |
556
+ | cosine_accuracy_threshold | 0.8319 |
557
+ | cosine_f1 | 0.6853 |
558
+ | cosine_f1_threshold | 0.7455 |
559
+ | cosine_precision | 0.5606 |
560
+ | cosine_recall | 0.8814 |
561
+ | cosine_ap | 0.6874 |
562
+ | dot_accuracy | 0.6934 |
563
+ | dot_accuracy_threshold | 639.0777 |
564
+ | dot_f1 | 0.6853 |
565
+ | dot_f1_threshold | 572.7137 |
566
+ | dot_precision | 0.5606 |
567
+ | dot_recall | 0.8814 |
568
+ | dot_ap | 0.6879 |
569
+ | manhattan_accuracy | 0.6992 |
570
+ | manhattan_accuracy_threshold | 362.1486 |
571
+ | manhattan_f1 | 0.6857 |
572
+ | manhattan_f1_threshold | 430.3852 |
573
+ | manhattan_precision | 0.5682 |
574
+ | manhattan_recall | 0.8644 |
575
+ | manhattan_ap | 0.6875 |
576
+ | euclidean_accuracy | 0.6934 |
577
+ | euclidean_accuracy_threshold | 16.0694 |
578
+ | euclidean_f1 | 0.6853 |
579
+ | euclidean_f1_threshold | 19.7729 |
580
+ | euclidean_precision | 0.5606 |
581
+ | euclidean_recall | 0.8814 |
582
+ | euclidean_ap | 0.6874 |
583
+ | max_accuracy | 0.6992 |
584
+ | max_accuracy_threshold | 639.0777 |
585
+ | max_f1 | 0.6857 |
586
+ | max_f1_threshold | 572.7137 |
587
+ | max_precision | 0.5682 |
588
+ | max_recall | 0.8814 |
589
+ | **max_ap** | **0.6879** |
590
+
591
+ <!--
592
+ ## Bias, Risks and Limitations
593
+
594
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
595
+ -->
596
+
597
+ <!--
598
+ ### Recommendations
599
+
600
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
601
+ -->
602
+
603
+ ## Training Details
604
+
605
+ ### Training Dataset
606
+
607
+ #### Unnamed Dataset
608
+
609
+
610
+ * Size: 32,500 training samples
611
+ * Columns: <code>sentence1</code> and <code>sentence2</code>
612
+ * Approximate statistics based on the first 1000 samples:
613
+ | | sentence1 | sentence2 |
614
+ |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
615
+ | type | string | string |
616
+ | details | <ul><li>min: 4 tokens</li><li>mean: 29.39 tokens</li><li>max: 323 tokens</li></ul> | <ul><li>min: 2 tokens</li><li>mean: 54.42 tokens</li><li>max: 423 tokens</li></ul> |
617
+ * Samples:
618
+ | sentence1 | sentence2 |
619
+ |:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
620
+ | <code>In which London road is Harrod’s department store?</code> | <code>Harrods, Brompton Road, London | Shopping/Department Stores in London | LondonTown.com Opening Times Britain's most famous store and possibly the most famous store in the world, Harrods features on many tourist 'must-see' lists - and with good reason. Its humble beginnings date back to 1849, when Charles Henry Harrod opened a small East End grocer and tea merchant business that emphasised impeccable service over value. Today, it occupies a vast seven floor site in London's fashionable Knightsbridge and boasts a phenomenal range of products from pianos and cooking pans to fashion and perfumery. The luxurious Urban Retreat can be found on the sixth floor while newer departments include Superbrands, with 17 boutiques from top international designers, and Salon du Parfums, housing some of the most exceptional and exclusive perfumes in the world. The Food Hall is ostentatious to the core and mouth-wateringly exotic, and the store as a whole is well served with 27 restaurants. At Christmas time the Brompton Road windows are transformed into a magical winter wonderland and Father Christmas takes up residence at the enchanting Christmas Grotto. The summer and winter sales are calendar events in the shopping year, and although both sales are extremely crowded there are some great bargains on offer. �</code> |
621
+ | <code>e.&#9;in solids the atoms are closely locked in position and can only vibrate, in liquids the atoms and molecules are more loosely connected and can collide with and move past one another, while in gases the atoms or molecules are free to move independently, colliding frequently.</code> | <code>Within a substance, atoms that collide frequently and move independently of one another are most likely in a gas</code> |
622
+ | <code>Joe Cole was unable to join West Bromwich Albion .</code> | <code>On 16th October Joe Cole took a long hard look at himself realising that he would never get the opportunity to join West Bromwich Albion and joined Coventry City instead.</code> |
623
+ * Loss: [<code>GISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#gistembedloss) with these parameters:
624
+ ```json
625
+ {'guide': SentenceTransformer(
626
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
627
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
628
+ (2): Normalize()
629
+ ), 'temperature': 0.025}
630
+ ```
631
+
632
+ ### Training Hyperparameters
633
+ #### Non-Default Hyperparameters
634
+
635
+ - `eval_strategy`: steps
636
+ - `per_device_train_batch_size`: 32
637
+ - `per_device_eval_batch_size`: 256
638
+ - `lr_scheduler_type`: cosine_with_min_lr
639
+ - `lr_scheduler_kwargs`: {'num_cycles': 0.5, 'min_lr': 3.3333333333333337e-06}
640
+ - `warmup_ratio`: 0.33
641
+ - `save_safetensors`: False
642
+ - `fp16`: True
643
+ - `push_to_hub`: True
644
+ - `hub_model_id`: bobox/DeBERTa3-s-CustomPoolin-toytest4-step1-checkpoints-tmp
645
+ - `hub_strategy`: all_checkpoints
646
+ - `batch_sampler`: no_duplicates
647
+
648
+ #### All Hyperparameters
649
+ <details><summary>Click to expand</summary>
650
+
651
+ - `overwrite_output_dir`: False
652
+ - `do_predict`: False
653
+ - `eval_strategy`: steps
654
+ - `prediction_loss_only`: True
655
+ - `per_device_train_batch_size`: 32
656
+ - `per_device_eval_batch_size`: 256
657
+ - `per_gpu_train_batch_size`: None
658
+ - `per_gpu_eval_batch_size`: None
659
+ - `gradient_accumulation_steps`: 1
660
+ - `eval_accumulation_steps`: None
661
+ - `torch_empty_cache_steps`: None
662
+ - `learning_rate`: 5e-05
663
+ - `weight_decay`: 0.0
664
+ - `adam_beta1`: 0.9
665
+ - `adam_beta2`: 0.999
666
+ - `adam_epsilon`: 1e-08
667
+ - `max_grad_norm`: 1.0
668
+ - `num_train_epochs`: 3
669
+ - `max_steps`: -1
670
+ - `lr_scheduler_type`: cosine_with_min_lr
671
+ - `lr_scheduler_kwargs`: {'num_cycles': 0.5, 'min_lr': 3.3333333333333337e-06}
672
+ - `warmup_ratio`: 0.33
673
+ - `warmup_steps`: 0
674
+ - `log_level`: passive
675
+ - `log_level_replica`: warning
676
+ - `log_on_each_node`: True
677
+ - `logging_nan_inf_filter`: True
678
+ - `save_safetensors`: False
679
+ - `save_on_each_node`: False
680
+ - `save_only_model`: False
681
+ - `restore_callback_states_from_checkpoint`: False
682
+ - `no_cuda`: False
683
+ - `use_cpu`: False
684
+ - `use_mps_device`: False
685
+ - `seed`: 42
686
+ - `data_seed`: None
687
+ - `jit_mode_eval`: False
688
+ - `use_ipex`: False
689
+ - `bf16`: False
690
+ - `fp16`: True
691
+ - `fp16_opt_level`: O1
692
+ - `half_precision_backend`: auto
693
+ - `bf16_full_eval`: False
694
+ - `fp16_full_eval`: False
695
+ - `tf32`: None
696
+ - `local_rank`: 0
697
+ - `ddp_backend`: None
698
+ - `tpu_num_cores`: None
699
+ - `tpu_metrics_debug`: False
700
+ - `debug`: []
701
+ - `dataloader_drop_last`: False
702
+ - `dataloader_num_workers`: 0
703
+ - `dataloader_prefetch_factor`: None
704
+ - `past_index`: -1
705
+ - `disable_tqdm`: False
706
+ - `remove_unused_columns`: True
707
+ - `label_names`: None
708
+ - `load_best_model_at_end`: False
709
+ - `ignore_data_skip`: False
710
+ - `fsdp`: []
711
+ - `fsdp_min_num_params`: 0
712
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
713
+ - `fsdp_transformer_layer_cls_to_wrap`: None
714
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
715
+ - `deepspeed`: None
716
+ - `label_smoothing_factor`: 0.0
717
+ - `optim`: adamw_torch
718
+ - `optim_args`: None
719
+ - `adafactor`: False
720
+ - `group_by_length`: False
721
+ - `length_column_name`: length
722
+ - `ddp_find_unused_parameters`: None
723
+ - `ddp_bucket_cap_mb`: None
724
+ - `ddp_broadcast_buffers`: False
725
+ - `dataloader_pin_memory`: True
726
+ - `dataloader_persistent_workers`: False
727
+ - `skip_memory_metrics`: True
728
+ - `use_legacy_prediction_loop`: False
729
+ - `push_to_hub`: True
730
+ - `resume_from_checkpoint`: None
731
+ - `hub_model_id`: bobox/DeBERTa3-s-CustomPoolin-toytest4-step1-checkpoints-tmp
732
+ - `hub_strategy`: all_checkpoints
733
+ - `hub_private_repo`: False
734
+ - `hub_always_push`: False
735
+ - `gradient_checkpointing`: False
736
+ - `gradient_checkpointing_kwargs`: None
737
+ - `include_inputs_for_metrics`: False
738
+ - `eval_do_concat_batches`: True
739
+ - `fp16_backend`: auto
740
+ - `push_to_hub_model_id`: None
741
+ - `push_to_hub_organization`: None
742
+ - `mp_parameters`:
743
+ - `auto_find_batch_size`: False
744
+ - `full_determinism`: False
745
+ - `torchdynamo`: None
746
+ - `ray_scope`: last
747
+ - `ddp_timeout`: 1800
748
+ - `torch_compile`: False
749
+ - `torch_compile_backend`: None
750
+ - `torch_compile_mode`: None
751
+ - `dispatch_batches`: None
752
+ - `split_batches`: None
753
+ - `include_tokens_per_second`: False
754
+ - `include_num_input_tokens_seen`: False
755
+ - `neftune_noise_alpha`: None
756
+ - `optim_target_modules`: None
757
+ - `batch_eval_metrics`: False
758
+ - `eval_on_start`: False
759
+ - `eval_use_gather_object`: False
760
+ - `batch_sampler`: no_duplicates
761
+ - `multi_dataset_batch_sampler`: proportional
762
+
763
+ </details>
764
+
765
+ ### Training Logs
766
+ <details><summary>Click to expand</summary>
767
+
768
+ | Epoch | Step | Training Loss | sts-test_spearman_cosine | allNLI-dev_max_ap | Qnli-dev_max_ap |
769
+ |:------:|:----:|:-------------:|:------------------------:|:-----------------:|:---------------:|
770
+ | 0.0010 | 1 | 6.0688 | - | - | - |
771
+ | 0.0020 | 2 | 7.5576 | - | - | - |
772
+ | 0.0030 | 3 | 4.6849 | - | - | - |
773
+ | 0.0039 | 4 | 5.4503 | - | - | - |
774
+ | 0.0049 | 5 | 5.6057 | - | - | - |
775
+ | 0.0059 | 6 | 6.3049 | - | - | - |
776
+ | 0.0069 | 7 | 6.8336 | - | - | - |
777
+ | 0.0079 | 8 | 5.0777 | - | - | - |
778
+ | 0.0089 | 9 | 4.8358 | - | - | - |
779
+ | 0.0098 | 10 | 4.641 | - | - | - |
780
+ | 0.0108 | 11 | 4.828 | - | - | - |
781
+ | 0.0118 | 12 | 5.2269 | - | - | - |
782
+ | 0.0128 | 13 | 5.6772 | - | - | - |
783
+ | 0.0138 | 14 | 5.1422 | - | - | - |
784
+ | 0.0148 | 15 | 6.2469 | - | - | - |
785
+ | 0.0157 | 16 | 4.6802 | - | - | - |
786
+ | 0.0167 | 17 | 4.5492 | - | - | - |
787
+ | 0.0177 | 18 | 4.8062 | - | - | - |
788
+ | 0.0187 | 19 | 7.5141 | - | - | - |
789
+ | 0.0197 | 20 | 5.5202 | - | - | - |
790
+ | 0.0207 | 21 | 6.5025 | - | - | - |
791
+ | 0.0217 | 22 | 7.318 | - | - | - |
792
+ | 0.0226 | 23 | 4.6458 | - | - | - |
793
+ | 0.0236 | 24 | 4.6191 | - | - | - |
794
+ | 0.0246 | 25 | 4.3159 | - | - | - |
795
+ | 0.0256 | 26 | 6.3677 | - | - | - |
796
+ | 0.0266 | 27 | 5.6052 | - | - | - |
797
+ | 0.0276 | 28 | 4.196 | - | - | - |
798
+ | 0.0285 | 29 | 4.4802 | - | - | - |
799
+ | 0.0295 | 30 | 4.9193 | - | - | - |
800
+ | 0.0305 | 31 | 4.0996 | - | - | - |
801
+ | 0.0315 | 32 | 5.6307 | - | - | - |
802
+ | 0.0325 | 33 | 4.5745 | - | - | - |
803
+ | 0.0335 | 34 | 4.4514 | - | - | - |
804
+ | 0.0344 | 35 | 4.0617 | - | - | - |
805
+ | 0.0354 | 36 | 5.0298 | - | - | - |
806
+ | 0.0364 | 37 | 3.9815 | - | - | - |
807
+ | 0.0374 | 38 | 4.0871 | - | - | - |
808
+ | 0.0384 | 39 | 4.2378 | - | - | - |
809
+ | 0.0394 | 40 | 3.8226 | - | - | - |
810
+ | 0.0404 | 41 | 4.3519 | - | - | - |
811
+ | 0.0413 | 42 | 3.6345 | - | - | - |
812
+ | 0.0423 | 43 | 5.0829 | - | - | - |
813
+ | 0.0433 | 44 | 4.6701 | - | - | - |
814
+ | 0.0443 | 45 | 4.1371 | - | - | - |
815
+ | 0.0453 | 46 | 4.2418 | - | - | - |
816
+ | 0.0463 | 47 | 4.4766 | - | - | - |
817
+ | 0.0472 | 48 | 4.4797 | - | - | - |
818
+ | 0.0482 | 49 | 3.8471 | - | - | - |
819
+ | 0.0492 | 50 | 4.3194 | - | - | - |
820
+ | 0.0502 | 51 | 3.9426 | - | - | - |
821
+ | 0.0512 | 52 | 3.5333 | - | - | - |
822
+ | 0.0522 | 53 | 4.2426 | - | - | - |
823
+ | 0.0531 | 54 | 3.9816 | - | - | - |
824
+ | 0.0541 | 55 | 3.663 | - | - | - |
825
+ | 0.0551 | 56 | 3.9057 | - | - | - |
826
+ | 0.0561 | 57 | 4.0345 | - | - | - |
827
+ | 0.0571 | 58 | 3.5233 | - | - | - |
828
+ | 0.0581 | 59 | 3.7999 | - | - | - |
829
+ | 0.0591 | 60 | 3.1885 | - | - | - |
830
+ | 0.0600 | 61 | 3.6013 | - | - | - |
831
+ | 0.0610 | 62 | 3.392 | - | - | - |
832
+ | 0.0620 | 63 | 3.3814 | - | - | - |
833
+ | 0.0630 | 64 | 4.0428 | - | - | - |
834
+ | 0.0640 | 65 | 3.7825 | - | - | - |
835
+ | 0.0650 | 66 | 3.4181 | - | - | - |
836
+ | 0.0659 | 67 | 3.7793 | - | - | - |
837
+ | 0.0669 | 68 | 3.8344 | - | - | - |
838
+ | 0.0679 | 69 | 3.2165 | - | - | - |
839
+ | 0.0689 | 70 | 3.3811 | - | - | - |
840
+ | 0.0699 | 71 | 3.5984 | - | - | - |
841
+ | 0.0709 | 72 | 3.8583 | - | - | - |
842
+ | 0.0719 | 73 | 3.296 | - | - | - |
843
+ | 0.0728 | 74 | 2.7661 | - | - | - |
844
+ | 0.0738 | 75 | 2.9805 | - | - | - |
845
+ | 0.0748 | 76 | 2.566 | - | - | - |
846
+ | 0.0758 | 77 | 3.258 | - | - | - |
847
+ | 0.0768 | 78 | 3.3804 | - | - | - |
848
+ | 0.0778 | 79 | 2.8828 | - | - | - |
849
+ | 0.0787 | 80 | 3.1077 | - | - | - |
850
+ | 0.0797 | 81 | 2.9441 | - | - | - |
851
+ | 0.0807 | 82 | 2.9465 | - | - | - |
852
+ | 0.0817 | 83 | 2.7088 | - | - | - |
853
+ | 0.0827 | 84 | 2.9215 | - | - | - |
854
+ | 0.0837 | 85 | 3.4698 | - | - | - |
855
+ | 0.0846 | 86 | 2.2414 | - | - | - |
856
+ | 0.0856 | 87 | 3.1601 | - | - | - |
857
+ | 0.0866 | 88 | 2.7714 | - | - | - |
858
+ | 0.0876 | 89 | 3.0311 | - | - | - |
859
+ | 0.0886 | 90 | 3.0336 | - | - | - |
860
+ | 0.0896 | 91 | 1.9358 | - | - | - |
861
+ | 0.0906 | 92 | 2.6031 | - | - | - |
862
+ | 0.0915 | 93 | 2.7515 | - | - | - |
863
+ | 0.0925 | 94 | 2.8496 | - | - | - |
864
+ | 0.0935 | 95 | 1.8015 | - | - | - |
865
+ | 0.0945 | 96 | 2.8138 | - | - | - |
866
+ | 0.0955 | 97 | 2.0597 | - | - | - |
867
+ | 0.0965 | 98 | 2.1053 | - | - | - |
868
+ | 0.0974 | 99 | 2.6785 | - | - | - |
869
+ | 0.0984 | 100 | 2.588 | - | - | - |
870
+ | 0.0994 | 101 | 2.0099 | - | - | - |
871
+ | 0.1004 | 102 | 2.7947 | - | - | - |
872
+ | 0.1014 | 103 | 2.3274 | - | - | - |
873
+ | 0.1024 | 104 | 2.2545 | - | - | - |
874
+ | 0.1033 | 105 | 2.4575 | - | - | - |
875
+ | 0.1043 | 106 | 2.4413 | - | - | - |
876
+ | 0.1053 | 107 | 2.3185 | - | - | - |
877
+ | 0.1063 | 108 | 2.1577 | - | - | - |
878
+ | 0.1073 | 109 | 2.1278 | - | - | - |
879
+ | 0.1083 | 110 | 2.0967 | - | - | - |
880
+ | 0.1093 | 111 | 2.6142 | - | - | - |
881
+ | 0.1102 | 112 | 1.8553 | - | - | - |
882
+ | 0.1112 | 113 | 2.1523 | - | - | - |
883
+ | 0.1122 | 114 | 2.1726 | - | - | - |
884
+ | 0.1132 | 115 | 1.8564 | - | - | - |
885
+ | 0.1142 | 116 | 1.8413 | - | - | - |
886
+ | 0.1152 | 117 | 2.0441 | - | - | - |
887
+ | 0.1161 | 118 | 2.2159 | - | - | - |
888
+ | 0.1171 | 119 | 2.6779 | - | - | - |
889
+ | 0.1181 | 120 | 2.2976 | - | - | - |
890
+ | 0.1191 | 121 | 1.9407 | - | - | - |
891
+ | 0.1201 | 122 | 1.9019 | - | - | - |
892
+ | 0.1211 | 123 | 2.2149 | - | - | - |
893
+ | 0.1220 | 124 | 1.6823 | - | - | - |
894
+ | 0.1230 | 125 | 1.8402 | - | - | - |
895
+ | 0.1240 | 126 | 1.6914 | - | - | - |
896
+ | 0.125 | 127 | 2.1626 | - | - | - |
897
+ | 0.1260 | 128 | 1.6414 | - | - | - |
898
+ | 0.1270 | 129 | 2.2043 | - | - | - |
899
+ | 0.1280 | 130 | 1.9987 | - | - | - |
900
+ | 0.1289 | 131 | 1.8868 | - | - | - |
901
+ | 0.1299 | 132 | 1.8262 | - | - | - |
902
+ | 0.1309 | 133 | 2.0404 | - | - | - |
903
+ | 0.1319 | 134 | 1.9134 | - | - | - |
904
+ | 0.1329 | 135 | 2.3725 | - | - | - |
905
+ | 0.1339 | 136 | 1.4127 | - | - | - |
906
+ | 0.1348 | 137 | 1.6876 | - | - | - |
907
+ | 0.1358 | 138 | 1.8376 | - | - | - |
908
+ | 0.1368 | 139 | 1.6992 | - | - | - |
909
+ | 0.1378 | 140 | 1.5032 | - | - | - |
910
+ | 0.1388 | 141 | 2.0334 | - | - | - |
911
+ | 0.1398 | 142 | 2.3581 | - | - | - |
912
+ | 0.1407 | 143 | 1.4236 | - | - | - |
913
+ | 0.1417 | 144 | 2.202 | - | - | - |
914
+ | 0.1427 | 145 | 1.7654 | - | - | - |
915
+ | 0.1437 | 146 | 1.5748 | - | - | - |
916
+ | 0.1447 | 147 | 1.7996 | - | - | - |
917
+ | 0.1457 | 148 | 1.7517 | - | - | - |
918
+ | 0.1467 | 149 | 1.8933 | - | - | - |
919
+ | 0.1476 | 150 | 1.2836 | - | - | - |
920
+ | 0.1486 | 151 | 1.7145 | - | - | - |
921
+ | 0.1496 | 152 | 1.6499 | - | - | - |
922
+ | 0.1506 | 153 | 1.8273 | 0.4057 | 0.4389 | 0.6725 |
923
+ | 0.1516 | 154 | 2.2859 | - | - | - |
924
+ | 0.1526 | 155 | 1.0833 | - | - | - |
925
+ | 0.1535 | 156 | 1.6829 | - | - | - |
926
+ | 0.1545 | 157 | 2.1464 | - | - | - |
927
+ | 0.1555 | 158 | 1.745 | - | - | - |
928
+ | 0.1565 | 159 | 1.7319 | - | - | - |
929
+ | 0.1575 | 160 | 1.6968 | - | - | - |
930
+ | 0.1585 | 161 | 1.7401 | - | - | - |
931
+ | 0.1594 | 162 | 1.729 | - | - | - |
932
+ | 0.1604 | 163 | 2.0782 | - | - | - |
933
+ | 0.1614 | 164 | 2.6545 | - | - | - |
934
+ | 0.1624 | 165 | 1.4045 | - | - | - |
935
+ | 0.1634 | 166 | 1.2937 | - | - | - |
936
+ | 0.1644 | 167 | 1.1171 | - | - | - |
937
+ | 0.1654 | 168 | 1.3537 | - | - | - |
938
+ | 0.1663 | 169 | 1.7028 | - | - | - |
939
+ | 0.1673 | 170 | 1.4143 | - | - | - |
940
+ | 0.1683 | 171 | 1.8648 | - | - | - |
941
+ | 0.1693 | 172 | 1.6768 | - | - | - |
942
+ | 0.1703 | 173 | 1.9528 | - | - | - |
943
+ | 0.1713 | 174 | 1.1718 | - | - | - |
944
+ | 0.1722 | 175 | 1.8176 | - | - | - |
945
+ | 0.1732 | 176 | 0.8439 | - | - | - |
946
+ | 0.1742 | 177 | 1.5092 | - | - | - |
947
+ | 0.1752 | 178 | 1.1947 | - | - | - |
948
+ | 0.1762 | 179 | 1.6395 | - | - | - |
949
+ | 0.1772 | 180 | 1.4394 | - | - | - |
950
+ | 0.1781 | 181 | 1.7548 | - | - | - |
951
+ | 0.1791 | 182 | 1.1181 | - | - | - |
952
+ | 0.1801 | 183 | 1.0271 | - | - | - |
953
+ | 0.1811 | 184 | 2.3108 | - | - | - |
954
+ | 0.1821 | 185 | 2.1242 | - | - | - |
955
+ | 0.1831 | 186 | 1.9822 | - | - | - |
956
+ | 0.1841 | 187 | 2.3605 | - | - | - |
957
+ | 0.1850 | 188 | 1.5251 | - | - | - |
958
+ | 0.1860 | 189 | 1.2351 | - | - | - |
959
+ | 0.1870 | 190 | 1.5859 | - | - | - |
960
+ | 0.1880 | 191 | 1.8056 | - | - | - |
961
+ | 0.1890 | 192 | 1.349 | - | - | - |
962
+ | 0.1900 | 193 | 0.893 | - | - | - |
963
+ | 0.1909 | 194 | 1.5122 | - | - | - |
964
+ | 0.1919 | 195 | 1.3875 | - | - | - |
965
+ | 0.1929 | 196 | 1.29 | - | - | - |
966
+ | 0.1939 | 197 | 2.2931 | - | - | - |
967
+ | 0.1949 | 198 | 1.2663 | - | - | - |
968
+ | 0.1959 | 199 | 1.9712 | - | - | - |
969
+ | 0.1969 | 200 | 2.3307 | - | - | - |
970
+ | 0.1978 | 201 | 1.6544 | - | - | - |
971
+ | 0.1988 | 202 | 1.638 | - | - | - |
972
+ | 0.1998 | 203 | 1.3412 | - | - | - |
973
+ | 0.2008 | 204 | 1.4454 | - | - | - |
974
+ | 0.2018 | 205 | 1.5437 | - | - | - |
975
+ | 0.2028 | 206 | 1.4921 | - | - | - |
976
+ | 0.2037 | 207 | 1.4298 | - | - | - |
977
+ | 0.2047 | 208 | 1.6174 | - | - | - |
978
+ | 0.2057 | 209 | 1.4137 | - | - | - |
979
+ | 0.2067 | 210 | 1.5652 | - | - | - |
980
+ | 0.2077 | 211 | 1.1631 | - | - | - |
981
+ | 0.2087 | 212 | 1.2351 | - | - | - |
982
+ | 0.2096 | 213 | 1.7537 | - | - | - |
983
+ | 0.2106 | 214 | 1.3186 | - | - | - |
984
+ | 0.2116 | 215 | 1.2258 | - | - | - |
985
+ | 0.2126 | 216 | 0.7695 | - | - | - |
986
+ | 0.2136 | 217 | 1.2775 | - | - | - |
987
+ | 0.2146 | 218 | 1.6795 | - | - | - |
988
+ | 0.2156 | 219 | 1.2862 | - | - | - |
989
+ | 0.2165 | 220 | 1.1723 | - | - | - |
990
+ | 0.2175 | 221 | 1.3322 | - | - | - |
991
+ | 0.2185 | 222 | 1.7564 | - | - | - |
992
+ | 0.2195 | 223 | 1.1071 | - | - | - |
993
+ | 0.2205 | 224 | 1.2011 | - | - | - |
994
+ | 0.2215 | 225 | 1.2303 | - | - | - |
995
+ | 0.2224 | 226 | 1.212 | - | - | - |
996
+ | 0.2234 | 227 | 1.0117 | - | - | - |
997
+ | 0.2244 | 228 | 1.1907 | - | - | - |
998
+ | 0.2254 | 229 | 2.1293 | - | - | - |
999
+ | 0.2264 | 230 | 1.3063 | - | - | - |
1000
+ | 0.2274 | 231 | 1.2841 | - | - | - |
1001
+ | 0.2283 | 232 | 1.3778 | - | - | - |
1002
+ | 0.2293 | 233 | 1.2242 | - | - | - |
1003
+ | 0.2303 | 234 | 0.9227 | - | - | - |
1004
+ | 0.2313 | 235 | 1.2221 | - | - | - |
1005
+ | 0.2323 | 236 | 2.1041 | - | - | - |
1006
+ | 0.2333 | 237 | 1.3341 | - | - | - |
1007
+ | 0.2343 | 238 | 1.0876 | - | - | - |
1008
+ | 0.2352 | 239 | 1.3328 | - | - | - |
1009
+ | 0.2362 | 240 | 1.2958 | - | - | - |
1010
+ | 0.2372 | 241 | 1.1522 | - | - | - |
1011
+ | 0.2382 | 242 | 1.7942 | - | - | - |
1012
+ | 0.2392 | 243 | 1.1325 | - | - | - |
1013
+ | 0.2402 | 244 | 1.6466 | - | - | - |
1014
+ | 0.2411 | 245 | 1.4608 | - | - | - |
1015
+ | 0.2421 | 246 | 0.6375 | - | - | - |
1016
+ | 0.2431 | 247 | 2.0177 | - | - | - |
1017
+ | 0.2441 | 248 | 1.2069 | - | - | - |
1018
+ | 0.2451 | 249 | 0.7639 | - | - | - |
1019
+ | 0.2461 | 250 | 1.3465 | - | - | - |
1020
+ | 0.2470 | 251 | 1.064 | - | - | - |
1021
+ | 0.2480 | 252 | 1.3757 | - | - | - |
1022
+ | 0.2490 | 253 | 1.612 | - | - | - |
1023
+ | 0.25 | 254 | 0.7917 | - | - | - |
1024
+ | 0.2510 | 255 | 1.5515 | - | - | - |
1025
+ | 0.2520 | 256 | 0.799 | - | - | - |
1026
+ | 0.2530 | 257 | 0.9882 | - | - | - |
1027
+ | 0.2539 | 258 | 1.1814 | - | - | - |
1028
+ | 0.2549 | 259 | 0.6394 | - | - | - |
1029
+ | 0.2559 | 260 | 1.4756 | - | - | - |
1030
+ | 0.2569 | 261 | 0.5338 | - | - | - |
1031
+ | 0.2579 | 262 | 0.9779 | - | - | - |
1032
+ | 0.2589 | 263 | 1.5307 | - | - | - |
1033
+ | 0.2598 | 264 | 1.1213 | - | - | - |
1034
+ | 0.2608 | 265 | 0.9482 | - | - | - |
1035
+ | 0.2618 | 266 | 0.9599 | - | - | - |
1036
+ | 0.2628 | 267 | 1.4455 | - | - | - |
1037
+ | 0.2638 | 268 | 1.6496 | - | - | - |
1038
+ | 0.2648 | 269 | 0.7402 | - | - | - |
1039
+ | 0.2657 | 270 | 0.7835 | - | - | - |
1040
+ | 0.2667 | 271 | 0.7821 | - | - | - |
1041
+ | 0.2677 | 272 | 1.5422 | - | - | - |
1042
+ | 0.2687 | 273 | 1.0995 | - | - | - |
1043
+ | 0.2697 | 274 | 1.378 | - | - | - |
1044
+ | 0.2707 | 275 | 1.3562 | - | - | - |
1045
+ | 0.2717 | 276 | 0.7376 | - | - | - |
1046
+ | 0.2726 | 277 | 1.1678 | - | - | - |
1047
+ | 0.2736 | 278 | 1.2989 | - | - | - |
1048
+ | 0.2746 | 279 | 1.9559 | - | - | - |
1049
+ | 0.2756 | 280 | 1.1237 | - | - | - |
1050
+ | 0.2766 | 281 | 0.952 | - | - | - |
1051
+ | 0.2776 | 282 | 1.6629 | - | - | - |
1052
+ | 0.2785 | 283 | 1.871 | - | - | - |
1053
+ | 0.2795 | 284 | 1.5946 | - | - | - |
1054
+ | 0.2805 | 285 | 1.4456 | - | - | - |
1055
+ | 0.2815 | 286 | 1.4085 | - | - | - |
1056
+ | 0.2825 | 287 | 1.1394 | - | - | - |
1057
+ | 0.2835 | 288 | 1.0315 | - | - | - |
1058
+ | 0.2844 | 289 | 1.488 | - | - | - |
1059
+ | 0.2854 | 290 | 1.4006 | - | - | - |
1060
+ | 0.2864 | 291 | 0.9237 | - | - | - |
1061
+ | 0.2874 | 292 | 1.163 | - | - | - |
1062
+ | 0.2884 | 293 | 1.7037 | - | - | - |
1063
+ | 0.2894 | 294 | 0.8715 | - | - | - |
1064
+ | 0.2904 | 295 | 1.2101 | - | - | - |
1065
+ | 0.2913 | 296 | 1.1179 | - | - | - |
1066
+ | 0.2923 | 297 | 1.3986 | - | - | - |
1067
+ | 0.2933 | 298 | 1.7068 | - | - | - |
1068
+ | 0.2943 | 299 | 0.8695 | - | - | - |
1069
+ | 0.2953 | 300 | 1.3778 | - | - | - |
1070
+ | 0.2963 | 301 | 1.2834 | - | - | - |
1071
+ | 0.2972 | 302 | 0.8123 | - | - | - |
1072
+ | 0.2982 | 303 | 1.6521 | - | - | - |
1073
+ | 0.2992 | 304 | 1.1064 | - | - | - |
1074
+ | 0.3002 | 305 | 0.9578 | - | - | - |
1075
+ | 0.3012 | 306 | 0.9254 | 0.4888 | 0.4789 | 0.7040 |
1076
+ | 0.3022 | 307 | 0.7541 | - | - | - |
1077
+ | 0.3031 | 308 | 0.7324 | - | - | - |
1078
+ | 0.3041 | 309 | 0.5974 | - | - | - |
1079
+ | 0.3051 | 310 | 1.1481 | - | - | - |
1080
+ | 0.3061 | 311 | 1.6179 | - | - | - |
1081
+ | 0.3071 | 312 | 1.4641 | - | - | - |
1082
+ | 0.3081 | 313 | 1.7185 | - | - | - |
1083
+ | 0.3091 | 314 | 0.9328 | - | - | - |
1084
+ | 0.3100 | 315 | 0.742 | - | - | - |
1085
+ | 0.3110 | 316 | 1.4173 | - | - | - |
1086
+ | 0.3120 | 317 | 0.7267 | - | - | - |
1087
+ | 0.3130 | 318 | 0.9494 | - | - | - |
1088
+ | 0.3140 | 319 | 1.5111 | - | - | - |
1089
+ | 0.3150 | 320 | 1.6949 | - | - | - |
1090
+ | 0.3159 | 321 | 1.7562 | - | - | - |
1091
+ | 0.3169 | 322 | 1.2532 | - | - | - |
1092
+ | 0.3179 | 323 | 1.1086 | - | - | - |
1093
+ | 0.3189 | 324 | 0.7377 | - | - | - |
1094
+ | 0.3199 | 325 | 1.085 | - | - | - |
1095
+ | 0.3209 | 326 | 0.7767 | - | - | - |
1096
+ | 0.3219 | 327 | 1.4441 | - | - | - |
1097
+ | 0.3228 | 328 | 0.8146 | - | - | - |
1098
+ | 0.3238 | 329 | 0.7403 | - | - | - |
1099
+ | 0.3248 | 330 | 0.8476 | - | - | - |
1100
+ | 0.3258 | 331 | 0.7323 | - | - | - |
1101
+ | 0.3268 | 332 | 1.2241 | - | - | - |
1102
+ | 0.3278 | 333 | 1.5065 | - | - | - |
1103
+ | 0.3287 | 334 | 0.5259 | - | - | - |
1104
+ | 0.3297 | 335 | 1.3103 | - | - | - |
1105
+ | 0.3307 | 336 | 0.8655 | - | - | - |
1106
+ | 0.3317 | 337 | 0.7575 | - | - | - |
1107
+ | 0.3327 | 338 | 1.968 | - | - | - |
1108
+ | 0.3337 | 339 | 1.317 | - | - | - |
1109
+ | 0.3346 | 340 | 1.1972 | - | - | - |
1110
+ | 0.3356 | 341 | 1.6323 | - | - | - |
1111
+ | 0.3366 | 342 | 1.0469 | - | - | - |
1112
+ | 0.3376 | 343 | 1.3349 | - | - | - |
1113
+ | 0.3386 | 344 | 0.9544 | - | - | - |
1114
+ | 0.3396 | 345 | 1.1894 | - | - | - |
1115
+ | 0.3406 | 346 | 0.7717 | - | - | - |
1116
+ | 0.3415 | 347 | 1.2563 | - | - | - |
1117
+ | 0.3425 | 348 | 1.2437 | - | - | - |
1118
+ | 0.3435 | 349 | 0.7806 | - | - | - |
1119
+ | 0.3445 | 350 | 0.8303 | - | - | - |
1120
+ | 0.3455 | 351 | 1.0926 | - | - | - |
1121
+ | 0.3465 | 352 | 0.6654 | - | - | - |
1122
+ | 0.3474 | 353 | 1.1087 | - | - | - |
1123
+ | 0.3484 | 354 | 1.1525 | - | - | - |
1124
+ | 0.3494 | 355 | 1.1127 | - | - | - |
1125
+ | 0.3504 | 356 | 1.4267 | - | - | - |
1126
+ | 0.3514 | 357 | 0.6148 | - | - | - |
1127
+ | 0.3524 | 358 | 1.0123 | - | - | - |
1128
+ | 0.3533 | 359 | 1.9682 | - | - | - |
1129
+ | 0.3543 | 360 | 0.8487 | - | - | - |
1130
+ | 0.3553 | 361 | 1.0412 | - | - | - |
1131
+ | 0.3563 | 362 | 1.0902 | - | - | - |
1132
+ | 0.3573 | 363 | 0.9606 | - | - | - |
1133
+ | 0.3583 | 364 | 0.9206 | - | - | - |
1134
+ | 0.3593 | 365 | 1.4727 | - | - | - |
1135
+ | 0.3602 | 366 | 0.9379 | - | - | - |
1136
+ | 0.3612 | 367 | 0.8387 | - | - | - |
1137
+ | 0.3622 | 368 | 0.9692 | - | - | - |
1138
+ | 0.3632 | 369 | 1.6298 | - | - | - |
1139
+ | 0.3642 | 370 | 1.0882 | - | - | - |
1140
+ | 0.3652 | 371 | 1.1558 | - | - | - |
1141
+ | 0.3661 | 372 | 0.9546 | - | - | - |
1142
+ | 0.3671 | 373 | 1.0124 | - | - | - |
1143
+ | 0.3681 | 374 | 1.3916 | - | - | - |
1144
+ | 0.3691 | 375 | 0.527 | - | - | - |
1145
+ | 0.3701 | 376 | 0.6387 | - | - | - |
1146
+ | 0.3711 | 377 | 1.1445 | - | - | - |
1147
+ | 0.3720 | 378 | 1.3309 | - | - | - |
1148
+ | 0.3730 | 379 | 1.5888 | - | - | - |
1149
+ | 0.3740 | 380 | 1.4422 | - | - | - |
1150
+ | 0.375 | 381 | 1.7044 | - | - | - |
1151
+ | 0.3760 | 382 | 0.7913 | - | - | - |
1152
+ | 0.3770 | 383 | 1.3241 | - | - | - |
1153
+ | 0.3780 | 384 | 0.6473 | - | - | - |
1154
+ | 0.3789 | 385 | 1.221 | - | - | - |
1155
+ | 0.3799 | 386 | 0.7773 | - | - | - |
1156
+ | 0.3809 | 387 | 1.054 | - | - | - |
1157
+ | 0.3819 | 388 | 0.9862 | - | - | - |
1158
+ | 0.3829 | 389 | 0.9684 | - | - | - |
1159
+ | 0.3839 | 390 | 1.3244 | - | - | - |
1160
+ | 0.3848 | 391 | 1.1787 | - | - | - |
1161
+ | 0.3858 | 392 | 1.4698 | - | - | - |
1162
+ | 0.3868 | 393 | 1.0961 | - | - | - |
1163
+ | 0.3878 | 394 | 1.1364 | - | - | - |
1164
+ | 0.3888 | 395 | 0.9368 | - | - | - |
1165
+ | 0.3898 | 396 | 1.1731 | - | - | - |
1166
+ | 0.3907 | 397 | 0.8686 | - | - | - |
1167
+ | 0.3917 | 398 | 0.7481 | - | - | - |
1168
+ | 0.3927 | 399 | 0.7261 | - | - | - |
1169
+ | 0.3937 | 400 | 1.2062 | - | - | - |
1170
+ | 0.3947 | 401 | 0.7462 | - | - | - |
1171
+ | 0.3957 | 402 | 1.0318 | - | - | - |
1172
+ | 0.3967 | 403 | 1.105 | - | - | - |
1173
+ | 0.3976 | 404 | 1.009 | - | - | - |
1174
+ | 0.3986 | 405 | 0.5941 | - | - | - |
1175
+ | 0.3996 | 406 | 1.7972 | - | - | - |
1176
+ | 0.4006 | 407 | 1.0544 | - | - | - |
1177
+ | 0.4016 | 408 | 1.3912 | - | - | - |
1178
+ | 0.4026 | 409 | 0.8305 | - | - | - |
1179
+ | 0.4035 | 410 | 0.8688 | - | - | - |
1180
+ | 0.4045 | 411 | 1.0069 | - | - | - |
1181
+ | 0.4055 | 412 | 1.3141 | - | - | - |
1182
+ | 0.4065 | 413 | 1.1042 | - | - | - |
1183
+ | 0.4075 | 414 | 1.1011 | - | - | - |
1184
+ | 0.4085 | 415 | 1.1192 | - | - | - |
1185
+ | 0.4094 | 416 | 1.5957 | - | - | - |
1186
+ | 0.4104 | 417 | 1.164 | - | - | - |
1187
+ | 0.4114 | 418 | 0.6425 | - | - | - |
1188
+ | 0.4124 | 419 | 0.6068 | - | - | - |
1189
+ | 0.4134 | 420 | 0.9275 | - | - | - |
1190
+ | 0.4144 | 421 | 0.8836 | - | - | - |
1191
+ | 0.4154 | 422 | 1.2115 | - | - | - |
1192
+ | 0.4163 | 423 | 0.8367 | - | - | - |
1193
+ | 0.4173 | 424 | 1.0595 | - | - | - |
1194
+ | 0.4183 | 425 | 0.826 | - | - | - |
1195
+ | 0.4193 | 426 | 0.707 | - | - | - |
1196
+ | 0.4203 | 427 | 0.6235 | - | - | - |
1197
+ | 0.4213 | 428 | 0.7719 | - | - | - |
1198
+ | 0.4222 | 429 | 1.0862 | - | - | - |
1199
+ | 0.4232 | 430 | 0.9311 | - | - | - |
1200
+ | 0.4242 | 431 | 1.2339 | - | - | - |
1201
+ | 0.4252 | 432 | 0.9891 | - | - | - |
1202
+ | 0.4262 | 433 | 1.8443 | - | - | - |
1203
+ | 0.4272 | 434 | 1.1799 | - | - | - |
1204
+ | 0.4281 | 435 | 0.759 | - | - | - |
1205
+ | 0.4291 | 436 | 1.1002 | - | - | - |
1206
+ | 0.4301 | 437 | 0.9141 | - | - | - |
1207
+ | 0.4311 | 438 | 0.5467 | - | - | - |
1208
+ | 0.4321 | 439 | 0.7476 | - | - | - |
1209
+ | 0.4331 | 440 | 1.14 | - | - | - |
1210
+ | 0.4341 | 441 | 1.1504 | - | - | - |
1211
+ | 0.4350 | 442 | 1.26 | - | - | - |
1212
+ | 0.4360 | 443 | 1.0311 | - | - | - |
1213
+ | 0.4370 | 444 | 1.0646 | - | - | - |
1214
+ | 0.4380 | 445 | 0.8687 | - | - | - |
1215
+ | 0.4390 | 446 | 0.6839 | - | - | - |
1216
+ | 0.4400 | 447 | 1.1376 | - | - | - |
1217
+ | 0.4409 | 448 | 0.9759 | - | - | - |
1218
+ | 0.4419 | 449 | 0.7971 | - | - | - |
1219
+ | 0.4429 | 450 | 0.9708 | - | - | - |
1220
+ | 0.4439 | 451 | 0.8217 | - | - | - |
1221
+ | 0.4449 | 452 | 1.3728 | - | - | - |
1222
+ | 0.4459 | 453 | 0.9119 | - | - | - |
1223
+ | 0.4469 | 454 | 1.012 | - | - | - |
1224
+ | 0.4478 | 455 | 1.3738 | - | - | - |
1225
+ | 0.4488 | 456 | 0.8219 | - | - | - |
1226
+ | 0.4498 | 457 | 1.2558 | - | - | - |
1227
+ | 0.4508 | 458 | 0.6247 | - | - | - |
1228
+ | 0.4518 | 459 | 0.7295 | 0.5410 | 0.4920 | 0.6879 |
1229
+ | 0.4528 | 460 | 0.8154 | - | - | - |
1230
+ | 0.4537 | 461 | 1.1392 | - | - | - |
1231
+ | 0.4547 | 462 | 0.8618 | - | - | - |
1232
+ | 0.4557 | 463 | 0.9669 | - | - | - |
1233
+ | 0.4567 | 464 | 0.8804 | - | - | - |
1234
+ | 0.4577 | 465 | 0.8479 | - | - | - |
1235
+ | 0.4587 | 466 | 0.6296 | - | - | - |
1236
+ | 0.4596 | 467 | 0.8449 | - | - | - |
1237
+ | 0.4606 | 468 | 0.9772 | - | - | - |
1238
+ | 0.4616 | 469 | 0.6424 | - | - | - |
1239
+ | 0.4626 | 470 | 0.9169 | - | - | - |
1240
+ | 0.4636 | 471 | 0.7599 | - | - | - |
1241
+ | 0.4646 | 472 | 0.8943 | - | - | - |
1242
+ | 0.4656 | 473 | 0.9475 | - | - | - |
1243
+ | 0.4665 | 474 | 1.4518 | - | - | - |
1244
+ | 0.4675 | 475 | 1.274 | - | - | - |
1245
+ | 0.4685 | 476 | 0.7306 | - | - | - |
1246
+ | 0.4695 | 477 | 0.9238 | - | - | - |
1247
+ | 0.4705 | 478 | 0.6593 | - | - | - |
1248
+ | 0.4715 | 479 | 1.0183 | - | - | - |
1249
+ | 0.4724 | 480 | 1.2577 | - | - | - |
1250
+ | 0.4734 | 481 | 0.8738 | - | - | - |
1251
+ | 0.4744 | 482 | 1.1416 | - | - | - |
1252
+ | 0.4754 | 483 | 0.7135 | - | - | - |
1253
+ | 0.4764 | 484 | 1.2587 | - | - | - |
1254
+ | 0.4774 | 485 | 0.8823 | - | - | - |
1255
+ | 0.4783 | 486 | 0.8423 | - | - | - |
1256
+ | 0.4793 | 487 | 0.7704 | - | - | - |
1257
+ | 0.4803 | 488 | 0.7049 | - | - | - |
1258
+ | 0.4813 | 489 | 1.1893 | - | - | - |
1259
+ | 0.4823 | 490 | 1.3985 | - | - | - |
1260
+ | 0.4833 | 491 | 1.3567 | - | - | - |
1261
+ | 0.4843 | 492 | 1.2573 | - | - | - |
1262
+ | 0.4852 | 493 | 0.7671 | - | - | - |
1263
+ | 0.4862 | 494 | 0.5425 | - | - | - |
1264
+ | 0.4872 | 495 | 0.9372 | - | - | - |
1265
+ | 0.4882 | 496 | 0.799 | - | - | - |
1266
+ | 0.4892 | 497 | 0.9548 | - | - | - |
1267
+ | 0.4902 | 498 | 1.0855 | - | - | - |
1268
+ | 0.4911 | 499 | 1.0465 | - | - | - |
1269
+ | 0.4921 | 500 | 1.1004 | - | - | - |
1270
+ | 0.4931 | 501 | 0.6392 | - | - | - |
1271
+ | 0.4941 | 502 | 0.7102 | - | - | - |
1272
+ | 0.4951 | 503 | 1.3242 | - | - | - |
1273
+ | 0.4961 | 504 | 0.6861 | - | - | - |
1274
+ | 0.4970 | 505 | 0.9291 | - | - | - |
1275
+ | 0.4980 | 506 | 0.8592 | - | - | - |
1276
+ | 0.4990 | 507 | 0.9462 | - | - | - |
1277
+ | 0.5 | 508 | 1.0167 | - | - | - |
1278
+ | 0.5010 | 509 | 1.0118 | - | - | - |
1279
+ | 0.5020 | 510 | 0.6741 | - | - | - |
1280
+ | 0.5030 | 511 | 1.4578 | - | - | - |
1281
+ | 0.5039 | 512 | 1.2959 | - | - | - |
1282
+ | 0.5049 | 513 | 0.8533 | - | - | - |
1283
+ | 0.5059 | 514 | 0.6685 | - | - | - |
1284
+ | 0.5069 | 515 | 1.1556 | - | - | - |
1285
+ | 0.5079 | 516 | 0.8177 | - | - | - |
1286
+ | 0.5089 | 517 | 0.6296 | - | - | - |
1287
+ | 0.5098 | 518 | 0.8407 | - | - | - |
1288
+ | 0.5108 | 519 | 0.6987 | - | - | - |
1289
+ | 0.5118 | 520 | 0.9888 | - | - | - |
1290
+ | 0.5128 | 521 | 0.8938 | - | - | - |
1291
+ | 0.5138 | 522 | 0.582 | - | - | - |
1292
+ | 0.5148 | 523 | 0.6596 | - | - | - |
1293
+ | 0.5157 | 524 | 0.6029 | - | - | - |
1294
+ | 0.5167 | 525 | 0.9806 | - | - | - |
1295
+ | 0.5177 | 526 | 0.9463 | - | - | - |
1296
+ | 0.5187 | 527 | 0.7088 | - | - | - |
1297
+ | 0.5197 | 528 | 0.7525 | - | - | - |
1298
+ | 0.5207 | 529 | 0.7625 | - | - | - |
1299
+ | 0.5217 | 530 | 0.8271 | - | - | - |
1300
+ | 0.5226 | 531 | 0.6129 | - | - | - |
1301
+ | 0.5236 | 532 | 1.1563 | - | - | - |
1302
+ | 0.5246 | 533 | 0.8131 | - | - | - |
1303
+ | 0.5256 | 534 | 0.5363 | - | - | - |
1304
+ | 0.5266 | 535 | 0.8819 | - | - | - |
1305
+ | 0.5276 | 536 | 0.9772 | - | - | - |
1306
+ | 0.5285 | 537 | 1.2102 | - | - | - |
1307
+ | 0.5295 | 538 | 1.1234 | - | - | - |
1308
+ | 0.5305 | 539 | 1.1857 | - | - | - |
1309
+ | 0.5315 | 540 | 0.7873 | - | - | - |
1310
+ | 0.5325 | 541 | 0.5034 | - | - | - |
1311
+ | 0.5335 | 542 | 1.3305 | - | - | - |
1312
+ | 0.5344 | 543 | 1.1727 | - | - | - |
1313
+ | 0.5354 | 544 | 1.2825 | - | - | - |
1314
+ | 0.5364 | 545 | 1.0446 | - | - | - |
1315
+ | 0.5374 | 546 | 0.9838 | - | - | - |
1316
+ | 0.5384 | 547 | 1.2194 | - | - | - |
1317
+ | 0.5394 | 548 | 0.7709 | - | - | - |
1318
+ | 0.5404 | 549 | 0.748 | - | - | - |
1319
+ | 0.5413 | 550 | 1.0948 | - | - | - |
1320
+ | 0.5423 | 551 | 0.915 | - | - | - |
1321
+ | 0.5433 | 552 | 1.537 | - | - | - |
1322
+ | 0.5443 | 553 | 0.3239 | - | - | - |
1323
+ | 0.5453 | 554 | 0.9592 | - | - | - |
1324
+ | 0.5463 | 555 | 0.7737 | - | - | - |
1325
+ | 0.5472 | 556 | 0.613 | - | - | - |
1326
+ | 0.5482 | 557 | 1.3646 | - | - | - |
1327
+ | 0.5492 | 558 | 0.6659 | - | - | - |
1328
+ | 0.5502 | 559 | 0.5207 | - | - | - |
1329
+ | 0.5512 | 560 | 0.9467 | - | - | - |
1330
+ | 0.5522 | 561 | 0.5692 | - | - | - |
1331
+ | 0.5531 | 562 | 1.5855 | - | - | - |
1332
+ | 0.5541 | 563 | 0.8855 | - | - | - |
1333
+ | 0.5551 | 564 | 1.1829 | - | - | - |
1334
+ | 0.5561 | 565 | 0.978 | - | - | - |
1335
+ | 0.5571 | 566 | 1.1818 | - | - | - |
1336
+ | 0.5581 | 567 | 0.701 | - | - | - |
1337
+ | 0.5591 | 568 | 1.0226 | - | - | - |
1338
+ | 0.5600 | 569 | 0.5937 | - | - | - |
1339
+ | 0.5610 | 570 | 0.8095 | - | - | - |
1340
+ | 0.5620 | 571 | 1.174 | - | - | - |
1341
+ | 0.5630 | 572 | 0.96 | - | - | - |
1342
+ | 0.5640 | 573 | 0.8339 | - | - | - |
1343
+ | 0.5650 | 574 | 0.717 | - | - | - |
1344
+ | 0.5659 | 575 | 0.5938 | - | - | - |
1345
+ | 0.5669 | 576 | 0.6501 | - | - | - |
1346
+ | 0.5679 | 577 | 0.7003 | - | - | - |
1347
+ | 0.5689 | 578 | 0.5525 | - | - | - |
1348
+ | 0.5699 | 579 | 0.7003 | - | - | - |
1349
+ | 0.5709 | 580 | 1.059 | - | - | - |
1350
+ | 0.5719 | 581 | 0.8625 | - | - | - |
1351
+ | 0.5728 | 582 | 0.5862 | - | - | - |
1352
+ | 0.5738 | 583 | 0.9162 | - | - | - |
1353
+ | 0.5748 | 584 | 0.926 | - | - | - |
1354
+ | 0.5758 | 585 | 1.2729 | - | - | - |
1355
+ | 0.5768 | 586 | 0.8935 | - | - | - |
1356
+ | 0.5778 | 587 | 0.541 | - | - | - |
1357
+ | 0.5787 | 588 | 1.1455 | - | - | - |
1358
+ | 0.5797 | 589 | 0.7306 | - | - | - |
1359
+ | 0.5807 | 590 | 0.9088 | - | - | - |
1360
+ | 0.5817 | 591 | 0.9166 | - | - | - |
1361
+ | 0.5827 | 592 | 0.8679 | - | - | - |
1362
+ | 0.5837 | 593 | 0.9329 | - | - | - |
1363
+ | 0.5846 | 594 | 1.1201 | - | - | - |
1364
+ | 0.5856 | 595 | 0.6418 | - | - | - |
1365
+ | 0.5866 | 596 | 1.145 | - | - | - |
1366
+ | 0.5876 | 597 | 1.4041 | - | - | - |
1367
+ | 0.5886 | 598 | 0.6954 | - | - | - |
1368
+ | 0.5896 | 599 | 0.4567 | - | - | - |
1369
+ | 0.5906 | 600 | 1.1305 | - | - | - |
1370
+ | 0.5915 | 601 | 0.8077 | - | - | - |
1371
+ | 0.5925 | 602 | 0.6143 | - | - | - |
1372
+ | 0.5935 | 603 | 1.3139 | - | - | - |
1373
+ | 0.5945 | 604 | 0.7694 | - | - | - |
1374
+ | 0.5955 | 605 | 0.9622 | - | - | - |
1375
+ | 0.5965 | 606 | 0.91 | - | - | - |
1376
+ | 0.5974 | 607 | 1.3125 | - | - | - |
1377
+ | 0.5984 | 608 | 1.0153 | - | - | - |
1378
+ | 0.5994 | 609 | 0.8468 | - | - | - |
1379
+ | 0.6004 | 610 | 1.1026 | - | - | - |
1380
+
1381
+ </details>
1382
+
1383
+ ### Framework Versions
1384
+ - Python: 3.10.12
1385
+ - Sentence Transformers: 3.2.1
1386
+ - Transformers: 4.44.2
1387
+ - PyTorch: 2.5.0+cu121
1388
+ - Accelerate: 0.34.2
1389
+ - Datasets: 3.0.2
1390
+ - Tokenizers: 0.19.1
1391
+
1392
+ ## Citation
1393
+
1394
+ ### BibTeX
1395
+
1396
+ #### Sentence Transformers
1397
+ ```bibtex
1398
+ @inproceedings{reimers-2019-sentence-bert,
1399
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
1400
+ author = "Reimers, Nils and Gurevych, Iryna",
1401
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
1402
+ month = "11",
1403
+ year = "2019",
1404
+ publisher = "Association for Computational Linguistics",
1405
+ url = "https://arxiv.org/abs/1908.10084",
1406
+ }
1407
+ ```
1408
+
1409
+ #### GISTEmbedLoss
1410
+ ```bibtex
1411
+ @misc{solatorio2024gistembed,
1412
+ title={GISTEmbed: Guided In-sample Selection of Training Negatives for Text Embedding Fine-tuning},
1413
+ author={Aivin V. Solatorio},
1414
+ year={2024},
1415
+ eprint={2402.16829},
1416
+ archivePrefix={arXiv},
1417
+ primaryClass={cs.LG}
1418
+ }
1419
+ ```
1420
+
1421
+ <!--
1422
+ ## Glossary
1423
+
1424
+ *Clearly define terms in order to be accessible across audiences.*
1425
+ -->
1426
+
1427
+ <!--
1428
+ ## Model Card Authors
1429
+
1430
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
1431
+ -->
1432
+
1433
+ <!--
1434
+ ## Model Card Contact
1435
+
1436
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
1437
+ -->
checkpoint-610/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
checkpoint-610/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-small",
3
+ "architectures": [
4
+ "DebertaV2Model"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 3072,
12
+ "layer_norm_eps": 1e-07,
13
+ "max_position_embeddings": 512,
14
+ "max_relative_positions": -1,
15
+ "model_type": "deberta-v2",
16
+ "norm_rel_ebd": "layer_norm",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 0,
20
+ "pooler_dropout": 0,
21
+ "pooler_hidden_act": "gelu",
22
+ "pooler_hidden_size": 768,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "position_buckets": 256,
29
+ "relative_attention": true,
30
+ "share_att_key": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.44.2",
33
+ "type_vocab_size": 0,
34
+ "vocab_size": 128100
35
+ }
checkpoint-610/config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.2.1",
4
+ "transformers": "4.44.2",
5
+ "pytorch": "2.5.0+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": null
10
+ }
checkpoint-610/modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_AdvancedWeightedPooling",
12
+ "type": "__main__.AdvancedWeightedPooling"
13
+ }
14
+ ]
checkpoint-610/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a36efb5235cf631c73bbdad07cca82e45a38e85d1d0e33160551b82d98e3a6b
3
+ size 141824506
checkpoint-610/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14974de0ed293b7621f8bbc7573d4f2123ddbba17549a8bc8047df81bd0d88fe
3
+ size 565251810
checkpoint-610/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b551dc922e9f7e61468803b40e31caffaae37664db85c108ae9332f7b1d8565
3
+ size 14244
checkpoint-610/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70e792ec668d2855c1b7c3e1075993ffab4ca74d45eceda5f2be1465752f3c31
3
+ size 1256
checkpoint-610/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
checkpoint-610/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
checkpoint-610/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
checkpoint-610/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-610/tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
+ }
checkpoint-610/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-610/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4e5a2c389c69a314b44e2abcb7834dfc6e25823a2070d8ca3efd3fc97499c9b
3
+ size 5688