bobox commited on
Commit
a16da11
1 Parent(s): e4bfcc0

Cached Gist embedd loss

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,1337 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: bobox/DeBERTa-small-ST-v1-test-step3
3
+ datasets: []
4
+ language: []
5
+ library_name: sentence-transformers
6
+ metrics:
7
+ - pearson_cosine
8
+ - spearman_cosine
9
+ - pearson_manhattan
10
+ - spearman_manhattan
11
+ - pearson_euclidean
12
+ - spearman_euclidean
13
+ - pearson_dot
14
+ - spearman_dot
15
+ - pearson_max
16
+ - spearman_max
17
+ pipeline_tag: sentence-similarity
18
+ tags:
19
+ - sentence-transformers
20
+ - sentence-similarity
21
+ - feature-extraction
22
+ - generated_from_trainer
23
+ - dataset_size:260034
24
+ - loss:CachedGISTEmbedLoss
25
+ widget:
26
+ - source_sentence: who used to present one man and his dog
27
+ sentences:
28
+ - One Man and His Dog One Man and His Dog is a BBC television series in the United
29
+ Kingdom featuring sheepdog trials, originally presented by Phil Drabble, with
30
+ commentary by Eric Halsall and, later, by Ray Ollerenshaw. It was first aired
31
+ on 17 February 1976 and continues today (since 2013) as a special annual edition
32
+ of Countryfile. In 1994, Robin Page replaced Drabble as the main presenter. Gus
33
+ Dermody took over as commentator until 2012.
34
+ - 'animal adjectives [was: ratto, Ratte, raton] - Google Groups animal adjectives
35
+ [was: ratto, Ratte, raton] Showing 1-9 of 9 messages While trying find the pronunciation
36
+ of the word "munger", I encountered the nearby word    murine [MYOO-ryn] = relating
37
+ to mice or rats    [from Latin _murinus_, which derives from _mus_,    mouse,
38
+ whose genetive form is _muris_] So if you need an adjective to refer to lab rodents
39
+ like _ratto_ or _mausu_, "murine" it is. (I would never have discovered this except
40
+ in an alphabetically arranged dictionary.) There are a lot of animal adjectives
41
+ of this type, such as ovine (sheep), equine (horse), bovine (bull, cow, calf),
42
+ aquiline (eagle), murine (rats and mice).   But what is needed is a way to lookup
43
+ an animal and find what the proper adjective is.  For example, is there an adjective
44
+ form for "goat"? for "seal"? for "elephant"? for "whale"? for "walrus"? By the
45
+ way, I never did find out how "munger" is pronounced; the answer is not found
46
+ in'
47
+ - A boat is docked and filled with bicycles next to a grassy area on a body of water.
48
+ - source_sentence: There were 29 Muslims fatalities in the Cave of the Patriarchs
49
+ massacre .
50
+ sentences:
51
+ - 'Urban Dictionary: Dog and Bone Dog and Bone Cockney rhyming slang for phone -
52
+ the telephone. ''''Pick up the dog and bone now'''' by Brendan April 05, 2003
53
+ Create a mug The Urban Dictionary Mug One side has the word, one side has the
54
+ definition. Microwave and dishwasher safe. Lotsa space for your liquids. Buy the
55
+ t-shirt The Urban Dictionary T-Shirt Smooth, soft, slim fit American Apparel shirt.
56
+ Custom printed. 100% fine jersey cotton, except for heather grey (90% cotton).
57
+ ^Same as above except can be shortened further to ''Dogs'' or just ''dog'' Get
58
+ on the dogs and give us a bell when your ready. by Phaze October 14, 2004'
59
+ - RAF College Cranwell - Local Area Information RAF College Cranwell Local Area
60
+ Information Local Area Information RAF College Cranwell is situated in the North
61
+ Kesteven District Council area in the heart of rural Lincolnshire, 5 miles from
62
+ Sleaford and 14 miles from the City of Lincoln, surrounded by bustling market
63
+ towns, picturesque villages and landscapes steeped in aviation history. Lincolnshire
64
+ is currently home to several operational RAF airfields and was a key location
65
+ during WWII for bomber stations. Museums, memorials, former airfields, heritage
66
+ and visitor centres bear witness to the bravery of the men and women of this time.
67
+ The ancient City of Lincoln dates back at least to Roman times and boasts a spectacular
68
+ Cathedral and Castle area, whilst Sleaford is the home to the National Centre
69
+ for Craft & Design. Please click on the Logo to access website
70
+ - 29 Muslims were killed and more than 100 others wounded . [ Settlers remember
71
+ gunman Goldstein ; Hebron riots continue ] .
72
+ - source_sentence: What requires energy for growth?
73
+ sentences:
74
+ - "an organism requires energy for growth. Fish Fish are the ultimate aquatic organism.\
75
+ \ \n a fish require energy for growth"
76
+ - In August , after the end of the war in June 1902 , Higgins Southampton left the
77
+ `` SSBavarian '' and returned to Cape Town the following month .
78
+ - Rhinestone Cowboy "Rhinestone Cowboy" is a song written by Larry Weiss and most
79
+ famously recorded by American country music singer Glen Campbell. The song enjoyed
80
+ huge popularity with both country and pop audiences when it was released in 1975.
81
+ - source_sentence: Burning wood is used to produce what type of energy?
82
+ sentences:
83
+ - Shawnee Trails Council was formed from the merger of the Four Rivers Council and
84
+ the Audubon Council .
85
+ - A Mercedes parked next to a parking meter on a street.
86
+ - "burning wood is used to produce heat. Heat is kinetic energy. \n burning wood\
87
+ \ is used to produce kinetic energy."
88
+ - source_sentence: As of March , more than 413,000 cases have been confirmed in more
89
+ than 190 countries with more than 107,000 recoveries .
90
+ sentences:
91
+ - As of 24 March , more than 414,000 cases of COVID-19 have been reported in more
92
+ than 190 countries and territories , resulting in more than 18,500 deaths and
93
+ more than 108,000 recoveries .
94
+ - 'Pope Francis makes first visit as head of state to Italy\''s president - YouTube
95
+ Pope Francis makes first visit as head of state to Italy\''s president Want to
96
+ watch this again later? Sign in to add this video to a playlist. Need to report
97
+ the video? Sign in to report inappropriate content. The interactive transcript
98
+ could not be loaded. Loading... Rating is available when the video has been rented.
99
+ This feature is not available right now. Please try again later. Published on
100
+ Nov 14, 2013 Pope Francis stepped out of the Vatican, several hundred feet into
101
+ the heart of Rome, to meet with Italian President Giorgio Napolitano, and the
102
+ country\''s Council of Ministers. . --------------------- Suscríbete al canal:
103
+ http://smarturl.it/RomeReports Visita nuestra web: http://www.romereports.com/
104
+ ROME REPORTS, www.romereports.com, is an independent international TV News Agency
105
+ based in Rome covering the activity of the Pope, the life of the Vatican and current
106
+ social, cultural and religious debates. Reporting on the Catholic Church requires
107
+ proximity to the source, in-depth knowledge of the Institution, and a high standard
108
+ of creativity and technical excellence. As few broadcasters have a permanent correspondent
109
+ in Rome, ROME REPORTS is geared to inform the public and meet the needs of television
110
+ broadcasting companies around the world through daily news packages, weekly newsprograms
111
+ and documentaries. ---------------------'
112
+ - German shepherds and retrievers are commonly used, but the Belgian Malinois has
113
+ proven to be one of the most outstanding working dogs used in military service.
114
+ Around 85 percent of military working dogs are purchased in Germany or the Netherlands,
115
+ where they have been breeding dogs for military purposes for hundreds of years.
116
+ In addition, the Air Force Security Forces Center, Army Veterinary Corps and the
117
+ 341st Training Squadron combine efforts to raise their own dogs; nearly 15 percent
118
+ of all military working dogs are now bred here.
119
+ model-index:
120
+ - name: SentenceTransformer based on bobox/DeBERTa-small-ST-v1-test-step3
121
+ results:
122
+ - task:
123
+ type: semantic-similarity
124
+ name: Semantic Similarity
125
+ dataset:
126
+ name: sts test
127
+ type: sts-test
128
+ metrics:
129
+ - type: pearson_cosine
130
+ value: 0.875643593885091
131
+ name: Pearson Cosine
132
+ - type: spearman_cosine
133
+ value: 0.9063415240472948
134
+ name: Spearman Cosine
135
+ - type: pearson_manhattan
136
+ value: 0.9077403211524888
137
+ name: Pearson Manhattan
138
+ - type: spearman_manhattan
139
+ value: 0.9055112293832712
140
+ name: Spearman Manhattan
141
+ - type: pearson_euclidean
142
+ value: 0.9077080621981075
143
+ name: Pearson Euclidean
144
+ - type: spearman_euclidean
145
+ value: 0.9061498543947556
146
+ name: Spearman Euclidean
147
+ - type: pearson_dot
148
+ value: 0.8591462310934479
149
+ name: Pearson Dot
150
+ - type: spearman_dot
151
+ value: 0.8674279304506193
152
+ name: Spearman Dot
153
+ - type: pearson_max
154
+ value: 0.9077403211524888
155
+ name: Pearson Max
156
+ - type: spearman_max
157
+ value: 0.9063415240472948
158
+ name: Spearman Max
159
+ ---
160
+
161
+ # SentenceTransformer based on bobox/DeBERTa-small-ST-v1-test-step3
162
+
163
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [bobox/DeBERTa-small-ST-v1-test-step3](https://huggingface.co/bobox/DeBERTa-small-ST-v1-test-step3) on the bobox/enhanced_nli-50_k dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
164
+
165
+ ## Model Details
166
+
167
+ ### Model Description
168
+ - **Model Type:** Sentence Transformer
169
+ - **Base model:** [bobox/DeBERTa-small-ST-v1-test-step3](https://huggingface.co/bobox/DeBERTa-small-ST-v1-test-step3) <!-- at revision df9aaa75fe0c2791e5ed35ff33de1689d9a5f5ff -->
170
+ - **Maximum Sequence Length:** 512 tokens
171
+ - **Output Dimensionality:** 768 tokens
172
+ - **Similarity Function:** Cosine Similarity
173
+ - **Training Dataset:**
174
+ - bobox/enhanced_nli-50_k
175
+ <!-- - **Language:** Unknown -->
176
+ <!-- - **License:** Unknown -->
177
+
178
+ ### Model Sources
179
+
180
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
181
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
182
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
183
+
184
+ ### Full Model Architecture
185
+
186
+ ```
187
+ SentenceTransformer(
188
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: DebertaV2Model
189
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
190
+ )
191
+ ```
192
+
193
+ ## Usage
194
+
195
+ ### Direct Usage (Sentence Transformers)
196
+
197
+ First install the Sentence Transformers library:
198
+
199
+ ```bash
200
+ pip install -U sentence-transformers
201
+ ```
202
+
203
+ Then you can load this model and run inference.
204
+ ```python
205
+ from sentence_transformers import SentenceTransformer
206
+
207
+ # Download from the 🤗 Hub
208
+ model = SentenceTransformer("bobox/DeBERTa-small-ST-v1-test-UnifiedDatasets-Ft2")
209
+ # Run inference
210
+ sentences = [
211
+ 'As of March , more than 413,000 cases have been confirmed in more than 190 countries with more than 107,000 recoveries .',
212
+ 'As of 24 March , more than 414,000 cases of COVID-19 have been reported in more than 190 countries and territories , resulting in more than 18,500 deaths and more than 108,000 recoveries .',
213
+ 'German shepherds and retrievers are commonly used, but the Belgian Malinois has proven to be one of the most outstanding working dogs used in military service. Around 85 percent of military working dogs are purchased in Germany or the Netherlands, where they have been breeding dogs for military purposes for hundreds of years. In addition, the Air Force Security Forces Center, Army Veterinary Corps and the 341st Training Squadron combine efforts to raise their own dogs; nearly 15 percent of all military working dogs are now bred here.',
214
+ ]
215
+ embeddings = model.encode(sentences)
216
+ print(embeddings.shape)
217
+ # [3, 768]
218
+
219
+ # Get the similarity scores for the embeddings
220
+ similarities = model.similarity(embeddings, embeddings)
221
+ print(similarities.shape)
222
+ # [3, 3]
223
+ ```
224
+
225
+ <!--
226
+ ### Direct Usage (Transformers)
227
+
228
+ <details><summary>Click to see the direct usage in Transformers</summary>
229
+
230
+ </details>
231
+ -->
232
+
233
+ <!--
234
+ ### Downstream Usage (Sentence Transformers)
235
+
236
+ You can finetune this model on your own dataset.
237
+
238
+ <details><summary>Click to expand</summary>
239
+
240
+ </details>
241
+ -->
242
+
243
+ <!--
244
+ ### Out-of-Scope Use
245
+
246
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
247
+ -->
248
+
249
+ ## Evaluation
250
+
251
+ ### Metrics
252
+
253
+ #### Semantic Similarity
254
+ * Dataset: `sts-test`
255
+ * Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
256
+
257
+ | Metric | Value |
258
+ |:--------------------|:-----------|
259
+ | pearson_cosine | 0.8756 |
260
+ | **spearman_cosine** | **0.9063** |
261
+ | pearson_manhattan | 0.9077 |
262
+ | spearman_manhattan | 0.9055 |
263
+ | pearson_euclidean | 0.9077 |
264
+ | spearman_euclidean | 0.9061 |
265
+ | pearson_dot | 0.8591 |
266
+ | spearman_dot | 0.8674 |
267
+ | pearson_max | 0.9077 |
268
+ | spearman_max | 0.9063 |
269
+
270
+ <!--
271
+ ## Bias, Risks and Limitations
272
+
273
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
274
+ -->
275
+
276
+ <!--
277
+ ### Recommendations
278
+
279
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
280
+ -->
281
+
282
+ ## Training Details
283
+
284
+ ### Training Dataset
285
+
286
+ #### bobox/enhanced_nli-50_k
287
+
288
+ * Dataset: bobox/enhanced_nli-50_k
289
+ * Size: 260,034 training samples
290
+ * Columns: <code>sentence1</code> and <code>sentence2</code>
291
+ * Approximate statistics based on the first 1000 samples:
292
+ | | sentence1 | sentence2 |
293
+ |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
294
+ | type | string | string |
295
+ | details | <ul><li>min: 4 tokens</li><li>mean: 39.12 tokens</li><li>max: 344 tokens</li></ul> | <ul><li>min: 2 tokens</li><li>mean: 60.17 tokens</li><li>max: 442 tokens</li></ul> |
296
+ * Samples:
297
+ | sentence1 | sentence2 |
298
+ |:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
299
+ | <code>Temple Meads Railway Station is in which English city?</code> | <code>Bristol Temple Meads station roof to be replaced - BBC News BBC News Bristol Temple Meads station roof to be replaced 17 October 2013 Image caption Bristol Temple Meads was designed by Isambard Kingdom Brunel Image caption It will cost Network Rail £15m to replace the station's roof Image caption A pact has been signed to redevelop the station over the next 25 years The entire roof on Bristol Temple Meads railway station is to be replaced. Network Rail says it has secured £15m to carry out maintenance of the roof and install new lighting and cables. The announcement was made as a pact was signed to "significantly transform" the station over the next 25 years. Network Rail, Bristol City Council, the West of England Local Enterprise Partnership, Homes and Communities Agency and English Heritage are supporting the plan. Each has signed the 25-year memorandum of understanding to redevelop the station. Patrick Hallgate, of Network Rail Western, said: "Our plans for Bristol will see the railway significantly transformed by the end of the decade, with more seats, better connections and more frequent services." The railway station was designed by Isambard Kingdom Brunel and opened in 1840.</code> |
300
+ | <code>Where do most of the digestion reactions occur?</code> | <code>Most of the digestion reactions occur in the small intestine.</code> |
301
+ | <code>Sacko, 22, joined Sporting from French top-flight side Bordeaux in 2014, but has so far been limited to playing for the Portuguese club's B team.<br>The former France Under-20 player joined Ligue 2 side Sochaux on loan in February and scored twice in 14 games.<br>He is Leeds' third signing of the transfer window, following the arrivals of Marcus Antonsson and Kyle Bartley.<br>Find all the latest football transfers on our dedicated page.</code> | <code>Leeds have signed Sporting Lisbon forward Hadi Sacko on a season-long loan with a view to a permanent deal.</code> |
302
+ * Loss: [<code>CachedGISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss) with these parameters:
303
+ ```json
304
+ {'guide': SentenceTransformer(
305
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
306
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
307
+ (2): Normalize()
308
+ ), 'temperature': 0.025}
309
+ ```
310
+
311
+ ### Evaluation Dataset
312
+
313
+ #### bobox/enhanced_nli-50_k
314
+
315
+ * Dataset: bobox/enhanced_nli-50_k
316
+ * Size: 1,506 evaluation samples
317
+ * Columns: <code>sentence1</code> and <code>sentence2</code>
318
+ * Approximate statistics based on the first 1000 samples:
319
+ | | sentence1 | sentence2 |
320
+ |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
321
+ | type | string | string |
322
+ | details | <ul><li>min: 3 tokens</li><li>mean: 31.16 tokens</li><li>max: 340 tokens</li></ul> | <ul><li>min: 2 tokens</li><li>mean: 62.3 tokens</li><li>max: 455 tokens</li></ul> |
323
+ * Samples:
324
+ | sentence1 | sentence2 |
325
+ |:----------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
326
+ | <code>Interestingly, snakes use their forked tongues to smell.</code> | <code>Snakes use their tongue to smell things.</code> |
327
+ | <code>A voltaic cell generates an electric current through a reaction known as a(n) spontaneous redox.</code> | <code>A voltaic cell uses what type of reaction to generate an electric current</code> |
328
+ | <code>As of March 22 , there were more than 321,000 cases with over 13,600 deaths and more than 96,000 recoveries reported worldwide .</code> | <code>As of 22 March , more than 321,000 cases of COVID-19 have been reported in over 180 countries and territories , resulting in more than 13,600 deaths and 96,000 recoveries .</code> |
329
+ * Loss: [<code>CachedGISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss) with these parameters:
330
+ ```json
331
+ {'guide': SentenceTransformer(
332
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
333
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
334
+ (2): Normalize()
335
+ ), 'temperature': 0.025}
336
+ ```
337
+
338
+ ### Training Hyperparameters
339
+ #### Non-Default Hyperparameters
340
+
341
+ - `eval_strategy`: steps
342
+ - `per_device_train_batch_size`: 320
343
+ - `per_device_eval_batch_size`: 128
344
+ - `learning_rate`: 2e-05
345
+ - `weight_decay`: 0.0001
346
+ - `num_train_epochs`: 1
347
+ - `lr_scheduler_type`: cosine_with_restarts
348
+ - `lr_scheduler_kwargs`: {'num_cycles': 3}
349
+ - `warmup_ratio`: 0.25
350
+ - `save_safetensors`: False
351
+ - `fp16`: True
352
+ - `push_to_hub`: True
353
+ - `hub_model_id`: bobox/DeBERTa-small-ST-v1-test-UnifiedDatasets-Ft2-checkpoints-tmp
354
+ - `hub_strategy`: all_checkpoints
355
+ - `batch_sampler`: no_duplicates
356
+
357
+ #### All Hyperparameters
358
+ <details><summary>Click to expand</summary>
359
+
360
+ - `overwrite_output_dir`: False
361
+ - `do_predict`: False
362
+ - `eval_strategy`: steps
363
+ - `prediction_loss_only`: True
364
+ - `per_device_train_batch_size`: 320
365
+ - `per_device_eval_batch_size`: 128
366
+ - `per_gpu_train_batch_size`: None
367
+ - `per_gpu_eval_batch_size`: None
368
+ - `gradient_accumulation_steps`: 1
369
+ - `eval_accumulation_steps`: None
370
+ - `torch_empty_cache_steps`: None
371
+ - `learning_rate`: 2e-05
372
+ - `weight_decay`: 0.0001
373
+ - `adam_beta1`: 0.9
374
+ - `adam_beta2`: 0.999
375
+ - `adam_epsilon`: 1e-08
376
+ - `max_grad_norm`: 1.0
377
+ - `num_train_epochs`: 1
378
+ - `max_steps`: -1
379
+ - `lr_scheduler_type`: cosine_with_restarts
380
+ - `lr_scheduler_kwargs`: {'num_cycles': 3}
381
+ - `warmup_ratio`: 0.25
382
+ - `warmup_steps`: 0
383
+ - `log_level`: passive
384
+ - `log_level_replica`: warning
385
+ - `log_on_each_node`: True
386
+ - `logging_nan_inf_filter`: True
387
+ - `save_safetensors`: False
388
+ - `save_on_each_node`: False
389
+ - `save_only_model`: False
390
+ - `restore_callback_states_from_checkpoint`: False
391
+ - `no_cuda`: False
392
+ - `use_cpu`: False
393
+ - `use_mps_device`: False
394
+ - `seed`: 42
395
+ - `data_seed`: None
396
+ - `jit_mode_eval`: False
397
+ - `use_ipex`: False
398
+ - `bf16`: False
399
+ - `fp16`: True
400
+ - `fp16_opt_level`: O1
401
+ - `half_precision_backend`: auto
402
+ - `bf16_full_eval`: False
403
+ - `fp16_full_eval`: False
404
+ - `tf32`: None
405
+ - `local_rank`: 0
406
+ - `ddp_backend`: None
407
+ - `tpu_num_cores`: None
408
+ - `tpu_metrics_debug`: False
409
+ - `debug`: []
410
+ - `dataloader_drop_last`: False
411
+ - `dataloader_num_workers`: 0
412
+ - `dataloader_prefetch_factor`: None
413
+ - `past_index`: -1
414
+ - `disable_tqdm`: False
415
+ - `remove_unused_columns`: True
416
+ - `label_names`: None
417
+ - `load_best_model_at_end`: False
418
+ - `ignore_data_skip`: False
419
+ - `fsdp`: []
420
+ - `fsdp_min_num_params`: 0
421
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
422
+ - `fsdp_transformer_layer_cls_to_wrap`: None
423
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
424
+ - `deepspeed`: None
425
+ - `label_smoothing_factor`: 0.0
426
+ - `optim`: adamw_torch
427
+ - `optim_args`: None
428
+ - `adafactor`: False
429
+ - `group_by_length`: False
430
+ - `length_column_name`: length
431
+ - `ddp_find_unused_parameters`: None
432
+ - `ddp_bucket_cap_mb`: None
433
+ - `ddp_broadcast_buffers`: False
434
+ - `dataloader_pin_memory`: True
435
+ - `dataloader_persistent_workers`: False
436
+ - `skip_memory_metrics`: True
437
+ - `use_legacy_prediction_loop`: False
438
+ - `push_to_hub`: True
439
+ - `resume_from_checkpoint`: None
440
+ - `hub_model_id`: bobox/DeBERTa-small-ST-v1-test-UnifiedDatasets-Ft2-checkpoints-tmp
441
+ - `hub_strategy`: all_checkpoints
442
+ - `hub_private_repo`: False
443
+ - `hub_always_push`: False
444
+ - `gradient_checkpointing`: False
445
+ - `gradient_checkpointing_kwargs`: None
446
+ - `include_inputs_for_metrics`: False
447
+ - `eval_do_concat_batches`: True
448
+ - `fp16_backend`: auto
449
+ - `push_to_hub_model_id`: None
450
+ - `push_to_hub_organization`: None
451
+ - `mp_parameters`:
452
+ - `auto_find_batch_size`: False
453
+ - `full_determinism`: False
454
+ - `torchdynamo`: None
455
+ - `ray_scope`: last
456
+ - `ddp_timeout`: 1800
457
+ - `torch_compile`: False
458
+ - `torch_compile_backend`: None
459
+ - `torch_compile_mode`: None
460
+ - `dispatch_batches`: None
461
+ - `split_batches`: None
462
+ - `include_tokens_per_second`: False
463
+ - `include_num_input_tokens_seen`: False
464
+ - `neftune_noise_alpha`: None
465
+ - `optim_target_modules`: None
466
+ - `batch_eval_metrics`: False
467
+ - `eval_on_start`: False
468
+ - `eval_use_gather_object`: False
469
+ - `batch_sampler`: no_duplicates
470
+ - `multi_dataset_batch_sampler`: proportional
471
+
472
+ </details>
473
+
474
+ ### Training Logs
475
+ <details><summary>Click to expand</summary>
476
+
477
+ | Epoch | Step | Training Loss | loss | sts-test_spearman_cosine |
478
+ |:------:|:----:|:-------------:|:------:|:------------------------:|
479
+ | 0.0012 | 1 | 0.3208 | - | - |
480
+ | 0.0025 | 2 | 0.1703 | - | - |
481
+ | 0.0037 | 3 | 0.3362 | - | - |
482
+ | 0.0049 | 4 | 0.3346 | - | - |
483
+ | 0.0062 | 5 | 0.2484 | - | - |
484
+ | 0.0074 | 6 | 0.2249 | - | - |
485
+ | 0.0086 | 7 | 0.2724 | - | - |
486
+ | 0.0098 | 8 | 0.251 | - | - |
487
+ | 0.0111 | 9 | 0.2413 | - | - |
488
+ | 0.0123 | 10 | 0.382 | - | - |
489
+ | 0.0135 | 11 | 0.2695 | - | - |
490
+ | 0.0148 | 12 | 0.2392 | - | - |
491
+ | 0.0160 | 13 | 0.3603 | - | - |
492
+ | 0.0172 | 14 | 0.3282 | - | - |
493
+ | 0.0185 | 15 | 0.2878 | - | - |
494
+ | 0.0197 | 16 | 0.3046 | - | - |
495
+ | 0.0209 | 17 | 0.3946 | - | - |
496
+ | 0.0221 | 18 | 0.2038 | - | - |
497
+ | 0.0234 | 19 | 0.3542 | - | - |
498
+ | 0.0246 | 20 | 0.2369 | - | - |
499
+ | 0.0258 | 21 | 0.1967 | 0.1451 | 0.9081 |
500
+ | 0.0271 | 22 | 0.2368 | - | - |
501
+ | 0.0283 | 23 | 0.263 | - | - |
502
+ | 0.0295 | 24 | 0.3595 | - | - |
503
+ | 0.0308 | 25 | 0.3073 | - | - |
504
+ | 0.0320 | 26 | 0.2232 | - | - |
505
+ | 0.0332 | 27 | 0.1822 | - | - |
506
+ | 0.0344 | 28 | 0.251 | - | - |
507
+ | 0.0357 | 29 | 0.2677 | - | - |
508
+ | 0.0369 | 30 | 0.3252 | - | - |
509
+ | 0.0381 | 31 | 0.2058 | - | - |
510
+ | 0.0394 | 32 | 0.3083 | - | - |
511
+ | 0.0406 | 33 | 0.2109 | - | - |
512
+ | 0.0418 | 34 | 0.2751 | - | - |
513
+ | 0.0431 | 35 | 0.2269 | - | - |
514
+ | 0.0443 | 36 | 0.2333 | - | - |
515
+ | 0.0455 | 37 | 0.2747 | - | - |
516
+ | 0.0467 | 38 | 0.1285 | - | - |
517
+ | 0.0480 | 39 | 0.3659 | - | - |
518
+ | 0.0492 | 40 | 0.3991 | - | - |
519
+ | 0.0504 | 41 | 0.2647 | - | - |
520
+ | 0.0517 | 42 | 0.3627 | 0.1373 | 0.9084 |
521
+ | 0.0529 | 43 | 0.2026 | - | - |
522
+ | 0.0541 | 44 | 0.1923 | - | - |
523
+ | 0.0554 | 45 | 0.2369 | - | - |
524
+ | 0.0566 | 46 | 0.2268 | - | - |
525
+ | 0.0578 | 47 | 0.2975 | - | - |
526
+ | 0.0590 | 48 | 0.1922 | - | - |
527
+ | 0.0603 | 49 | 0.1906 | - | - |
528
+ | 0.0615 | 50 | 0.2379 | - | - |
529
+ | 0.0627 | 51 | 0.3796 | - | - |
530
+ | 0.0640 | 52 | 0.1821 | - | - |
531
+ | 0.0652 | 53 | 0.1257 | - | - |
532
+ | 0.0664 | 54 | 0.2368 | - | - |
533
+ | 0.0677 | 55 | 0.294 | - | - |
534
+ | 0.0689 | 56 | 0.2594 | - | - |
535
+ | 0.0701 | 57 | 0.2972 | - | - |
536
+ | 0.0713 | 58 | 0.2297 | - | - |
537
+ | 0.0726 | 59 | 0.1487 | - | - |
538
+ | 0.0738 | 60 | 0.182 | - | - |
539
+ | 0.0750 | 61 | 0.2516 | - | - |
540
+ | 0.0763 | 62 | 0.2809 | - | - |
541
+ | 0.0775 | 63 | 0.1371 | 0.1308 | 0.9068 |
542
+ | 0.0787 | 64 | 0.2149 | - | - |
543
+ | 0.0800 | 65 | 0.1806 | - | - |
544
+ | 0.0812 | 66 | 0.1458 | - | - |
545
+ | 0.0824 | 67 | 0.249 | - | - |
546
+ | 0.0836 | 68 | 0.2787 | - | - |
547
+ | 0.0849 | 69 | 0.288 | - | - |
548
+ | 0.0861 | 70 | 0.1461 | - | - |
549
+ | 0.0873 | 71 | 0.2304 | - | - |
550
+ | 0.0886 | 72 | 0.3505 | - | - |
551
+ | 0.0898 | 73 | 0.2227 | - | - |
552
+ | 0.0910 | 74 | 0.1746 | - | - |
553
+ | 0.0923 | 75 | 0.1484 | - | - |
554
+ | 0.0935 | 76 | 0.1346 | - | - |
555
+ | 0.0947 | 77 | 0.2112 | - | - |
556
+ | 0.0959 | 78 | 0.3138 | - | - |
557
+ | 0.0972 | 79 | 0.2675 | - | - |
558
+ | 0.0984 | 80 | 0.2849 | - | - |
559
+ | 0.0996 | 81 | 0.1719 | - | - |
560
+ | 0.1009 | 82 | 0.2749 | - | - |
561
+ | 0.1021 | 83 | 0.3097 | - | - |
562
+ | 0.1033 | 84 | 0.2068 | 0.1260 | 0.9045 |
563
+ | 0.1046 | 85 | 0.22 | - | - |
564
+ | 0.1058 | 86 | 0.2977 | - | - |
565
+ | 0.1070 | 87 | 0.209 | - | - |
566
+ | 0.1082 | 88 | 0.2215 | - | - |
567
+ | 0.1095 | 89 | 0.1948 | - | - |
568
+ | 0.1107 | 90 | 0.2084 | - | - |
569
+ | 0.1119 | 91 | 0.1823 | - | - |
570
+ | 0.1132 | 92 | 0.255 | - | - |
571
+ | 0.1144 | 93 | 0.2675 | - | - |
572
+ | 0.1156 | 94 | 0.18 | - | - |
573
+ | 0.1169 | 95 | 0.2891 | - | - |
574
+ | 0.1181 | 96 | 0.253 | - | - |
575
+ | 0.1193 | 97 | 0.3481 | - | - |
576
+ | 0.1205 | 98 | 0.1688 | - | - |
577
+ | 0.1218 | 99 | 0.1808 | - | - |
578
+ | 0.1230 | 100 | 0.2821 | - | - |
579
+ | 0.1242 | 101 | 0.1856 | - | - |
580
+ | 0.1255 | 102 | 0.1441 | - | - |
581
+ | 0.1267 | 103 | 0.226 | - | - |
582
+ | 0.1279 | 104 | 0.1662 | - | - |
583
+ | 0.1292 | 105 | 0.2043 | 0.1187 | 0.9051 |
584
+ | 0.1304 | 106 | 0.3907 | - | - |
585
+ | 0.1316 | 107 | 0.1332 | - | - |
586
+ | 0.1328 | 108 | 0.2243 | - | - |
587
+ | 0.1341 | 109 | 0.162 | - | - |
588
+ | 0.1353 | 110 | 0.1481 | - | - |
589
+ | 0.1365 | 111 | 0.2163 | - | - |
590
+ | 0.1378 | 112 | 0.24 | - | - |
591
+ | 0.1390 | 113 | 0.1406 | - | - |
592
+ | 0.1402 | 114 | 0.1522 | - | - |
593
+ | 0.1415 | 115 | 0.2593 | - | - |
594
+ | 0.1427 | 116 | 0.2426 | - | - |
595
+ | 0.1439 | 117 | 0.1781 | - | - |
596
+ | 0.1451 | 118 | 0.264 | - | - |
597
+ | 0.1464 | 119 | 0.1944 | - | - |
598
+ | 0.1476 | 120 | 0.1341 | - | - |
599
+ | 0.1488 | 121 | 0.155 | - | - |
600
+ | 0.1501 | 122 | 0.2052 | - | - |
601
+ | 0.1513 | 123 | 0.2023 | - | - |
602
+ | 0.1525 | 124 | 0.1519 | - | - |
603
+ | 0.1538 | 125 | 0.2118 | - | - |
604
+ | 0.1550 | 126 | 0.2489 | 0.1147 | 0.9058 |
605
+ | 0.1562 | 127 | 0.1988 | - | - |
606
+ | 0.1574 | 128 | 0.1541 | - | - |
607
+ | 0.1587 | 129 | 0.1819 | - | - |
608
+ | 0.1599 | 130 | 0.1582 | - | - |
609
+ | 0.1611 | 131 | 0.2866 | - | - |
610
+ | 0.1624 | 132 | 0.2766 | - | - |
611
+ | 0.1636 | 133 | 0.1299 | - | - |
612
+ | 0.1648 | 134 | 0.2558 | - | - |
613
+ | 0.1661 | 135 | 0.1687 | - | - |
614
+ | 0.1673 | 136 | 0.173 | - | - |
615
+ | 0.1685 | 137 | 0.2276 | - | - |
616
+ | 0.1697 | 138 | 0.2174 | - | - |
617
+ | 0.1710 | 139 | 0.2666 | - | - |
618
+ | 0.1722 | 140 | 0.1524 | - | - |
619
+ | 0.1734 | 141 | 0.1179 | - | - |
620
+ | 0.1747 | 142 | 0.2475 | - | - |
621
+ | 0.1759 | 143 | 0.2662 | - | - |
622
+ | 0.1771 | 144 | 0.1596 | - | - |
623
+ | 0.1784 | 145 | 0.2331 | - | - |
624
+ | 0.1796 | 146 | 0.2905 | - | - |
625
+ | 0.1808 | 147 | 0.1342 | 0.1088 | 0.9051 |
626
+ | 0.1820 | 148 | 0.0839 | - | - |
627
+ | 0.1833 | 149 | 0.2055 | - | - |
628
+ | 0.1845 | 150 | 0.2196 | - | - |
629
+ | 0.1857 | 151 | 0.2283 | - | - |
630
+ | 0.1870 | 152 | 0.2105 | - | - |
631
+ | 0.1882 | 153 | 0.1534 | - | - |
632
+ | 0.1894 | 154 | 0.1954 | - | - |
633
+ | 0.1907 | 155 | 0.1332 | - | - |
634
+ | 0.1919 | 156 | 0.19 | - | - |
635
+ | 0.1931 | 157 | 0.1878 | - | - |
636
+ | 0.1943 | 158 | 0.1518 | - | - |
637
+ | 0.1956 | 159 | 0.1906 | - | - |
638
+ | 0.1968 | 160 | 0.155 | - | - |
639
+ | 0.1980 | 161 | 0.1519 | - | - |
640
+ | 0.1993 | 162 | 0.1726 | - | - |
641
+ | 0.2005 | 163 | 0.1618 | - | - |
642
+ | 0.2017 | 164 | 0.2767 | - | - |
643
+ | 0.2030 | 165 | 0.1996 | - | - |
644
+ | 0.2042 | 166 | 0.1907 | - | - |
645
+ | 0.2054 | 167 | 0.1928 | - | - |
646
+ | 0.2066 | 168 | 0.1507 | 0.1082 | 0.9045 |
647
+ | 0.2079 | 169 | 0.1637 | - | - |
648
+ | 0.2091 | 170 | 0.1687 | - | - |
649
+ | 0.2103 | 171 | 0.2181 | - | - |
650
+ | 0.2116 | 172 | 0.1496 | - | - |
651
+ | 0.2128 | 173 | 0.1749 | - | - |
652
+ | 0.2140 | 174 | 0.2374 | - | - |
653
+ | 0.2153 | 175 | 0.2122 | - | - |
654
+ | 0.2165 | 176 | 0.1617 | - | - |
655
+ | 0.2177 | 177 | 0.168 | - | - |
656
+ | 0.2189 | 178 | 0.263 | - | - |
657
+ | 0.2202 | 179 | 0.1328 | - | - |
658
+ | 0.2214 | 180 | 0.3157 | - | - |
659
+ | 0.2226 | 181 | 0.2164 | - | - |
660
+ | 0.2239 | 182 | 0.1255 | - | - |
661
+ | 0.2251 | 183 | 0.2863 | - | - |
662
+ | 0.2263 | 184 | 0.155 | - | - |
663
+ | 0.2276 | 185 | 0.1271 | - | - |
664
+ | 0.2288 | 186 | 0.216 | - | - |
665
+ | 0.2300 | 187 | 0.205 | - | - |
666
+ | 0.2312 | 188 | 0.1575 | - | - |
667
+ | 0.2325 | 189 | 0.1939 | 0.1057 | 0.9046 |
668
+ | 0.2337 | 190 | 0.2209 | - | - |
669
+ | 0.2349 | 191 | 0.153 | - | - |
670
+ | 0.2362 | 192 | 0.2187 | - | - |
671
+ | 0.2374 | 193 | 0.1593 | - | - |
672
+ | 0.2386 | 194 | 0.173 | - | - |
673
+ | 0.2399 | 195 | 0.2377 | - | - |
674
+ | 0.2411 | 196 | 0.2281 | - | - |
675
+ | 0.2423 | 197 | 0.2651 | - | - |
676
+ | 0.2435 | 198 | 0.118 | - | - |
677
+ | 0.2448 | 199 | 0.1728 | - | - |
678
+ | 0.2460 | 200 | 0.2299 | - | - |
679
+ | 0.2472 | 201 | 0.2342 | - | - |
680
+ | 0.2485 | 202 | 0.2413 | - | - |
681
+ | 0.2497 | 203 | 0.168 | - | - |
682
+ | 0.2509 | 204 | 0.1474 | - | - |
683
+ | 0.2522 | 205 | 0.1102 | - | - |
684
+ | 0.2534 | 206 | 0.2326 | - | - |
685
+ | 0.2546 | 207 | 0.1787 | - | - |
686
+ | 0.2558 | 208 | 0.1423 | - | - |
687
+ | 0.2571 | 209 | 0.2069 | - | - |
688
+ | 0.2583 | 210 | 0.136 | 0.1040 | 0.9056 |
689
+ | 0.2595 | 211 | 0.2407 | - | - |
690
+ | 0.2608 | 212 | 0.212 | - | - |
691
+ | 0.2620 | 213 | 0.1361 | - | - |
692
+ | 0.2632 | 214 | 0.2356 | - | - |
693
+ | 0.2645 | 215 | 0.1059 | - | - |
694
+ | 0.2657 | 216 | 0.2501 | - | - |
695
+ | 0.2669 | 217 | 0.1817 | - | - |
696
+ | 0.2681 | 218 | 0.2022 | - | - |
697
+ | 0.2694 | 219 | 0.2235 | - | - |
698
+ | 0.2706 | 220 | 0.2437 | - | - |
699
+ | 0.2718 | 221 | 0.1859 | - | - |
700
+ | 0.2731 | 222 | 0.2167 | - | - |
701
+ | 0.2743 | 223 | 0.1495 | - | - |
702
+ | 0.2755 | 224 | 0.2876 | - | - |
703
+ | 0.2768 | 225 | 0.1842 | - | - |
704
+ | 0.2780 | 226 | 0.144 | - | - |
705
+ | 0.2792 | 227 | 0.1571 | - | - |
706
+ | 0.2804 | 228 | 0.209 | - | - |
707
+ | 0.2817 | 229 | 0.2075 | - | - |
708
+ | 0.2829 | 230 | 0.1722 | - | - |
709
+ | 0.2841 | 231 | 0.1464 | 0.1039 | 0.9087 |
710
+ | 0.2854 | 232 | 0.2675 | - | - |
711
+ | 0.2866 | 233 | 0.2585 | - | - |
712
+ | 0.2878 | 234 | 0.134 | - | - |
713
+ | 0.2891 | 235 | 0.1765 | - | - |
714
+ | 0.2903 | 236 | 0.1826 | - | - |
715
+ | 0.2915 | 237 | 0.222 | - | - |
716
+ | 0.2927 | 238 | 0.134 | - | - |
717
+ | 0.2940 | 239 | 0.1902 | - | - |
718
+ | 0.2952 | 240 | 0.2461 | - | - |
719
+ | 0.2964 | 241 | 0.3094 | - | - |
720
+ | 0.2977 | 242 | 0.2252 | - | - |
721
+ | 0.2989 | 243 | 0.2466 | - | - |
722
+ | 0.3001 | 244 | 0.139 | - | - |
723
+ | 0.3014 | 245 | 0.154 | - | - |
724
+ | 0.3026 | 246 | 0.1979 | - | - |
725
+ | 0.3038 | 247 | 0.1121 | - | - |
726
+ | 0.3050 | 248 | 0.1361 | - | - |
727
+ | 0.3063 | 249 | 0.2492 | - | - |
728
+ | 0.3075 | 250 | 0.1903 | - | - |
729
+ | 0.3087 | 251 | 0.2333 | - | - |
730
+ | 0.3100 | 252 | 0.1805 | 0.1030 | 0.9099 |
731
+ | 0.3112 | 253 | 0.1929 | - | - |
732
+ | 0.3124 | 254 | 0.1424 | - | - |
733
+ | 0.3137 | 255 | 0.2318 | - | - |
734
+ | 0.3149 | 256 | 0.1524 | - | - |
735
+ | 0.3161 | 257 | 0.2195 | - | - |
736
+ | 0.3173 | 258 | 0.1338 | - | - |
737
+ | 0.3186 | 259 | 0.2543 | - | - |
738
+ | 0.3198 | 260 | 0.202 | - | - |
739
+ | 0.3210 | 261 | 0.1489 | - | - |
740
+ | 0.3223 | 262 | 0.1937 | - | - |
741
+ | 0.3235 | 263 | 0.2334 | - | - |
742
+ | 0.3247 | 264 | 0.1942 | - | - |
743
+ | 0.3260 | 265 | 0.2013 | - | - |
744
+ | 0.3272 | 266 | 0.2954 | - | - |
745
+ | 0.3284 | 267 | 0.188 | - | - |
746
+ | 0.3296 | 268 | 0.1688 | - | - |
747
+ | 0.3309 | 269 | 0.1415 | - | - |
748
+ | 0.3321 | 270 | 0.2249 | - | - |
749
+ | 0.3333 | 271 | 0.2606 | - | - |
750
+ | 0.3346 | 272 | 0.2559 | - | - |
751
+ | 0.3358 | 273 | 0.2673 | 0.1039 | 0.9078 |
752
+ | 0.3370 | 274 | 0.1618 | - | - |
753
+ | 0.3383 | 275 | 0.2602 | - | - |
754
+ | 0.3395 | 276 | 0.2339 | - | - |
755
+ | 0.3407 | 277 | 0.1843 | - | - |
756
+ | 0.3419 | 278 | 0.133 | - | - |
757
+ | 0.3432 | 279 | 0.2345 | - | - |
758
+ | 0.3444 | 280 | 0.2808 | - | - |
759
+ | 0.3456 | 281 | 0.1044 | - | - |
760
+ | 0.3469 | 282 | 0.1622 | - | - |
761
+ | 0.3481 | 283 | 0.1303 | - | - |
762
+ | 0.3493 | 284 | 0.1453 | - | - |
763
+ | 0.3506 | 285 | 0.237 | - | - |
764
+ | 0.3518 | 286 | 0.1726 | - | - |
765
+ | 0.3530 | 287 | 0.2195 | - | - |
766
+ | 0.3542 | 288 | 0.3016 | - | - |
767
+ | 0.3555 | 289 | 0.1626 | - | - |
768
+ | 0.3567 | 290 | 0.1902 | - | - |
769
+ | 0.3579 | 291 | 0.1387 | - | - |
770
+ | 0.3592 | 292 | 0.1047 | - | - |
771
+ | 0.3604 | 293 | 0.1954 | - | - |
772
+ | 0.3616 | 294 | 0.2089 | 0.1029 | 0.9083 |
773
+ | 0.3629 | 295 | 0.1485 | - | - |
774
+ | 0.3641 | 296 | 0.1724 | - | - |
775
+ | 0.3653 | 297 | 0.2017 | - | - |
776
+ | 0.3665 | 298 | 0.1591 | - | - |
777
+ | 0.3678 | 299 | 0.2396 | - | - |
778
+ | 0.3690 | 300 | 0.1395 | - | - |
779
+ | 0.3702 | 301 | 0.1806 | - | - |
780
+ | 0.3715 | 302 | 0.1882 | - | - |
781
+ | 0.3727 | 303 | 0.1188 | - | - |
782
+ | 0.3739 | 304 | 0.1564 | - | - |
783
+ | 0.3752 | 305 | 0.313 | - | - |
784
+ | 0.3764 | 306 | 0.1455 | - | - |
785
+ | 0.3776 | 307 | 0.1535 | - | - |
786
+ | 0.3788 | 308 | 0.099 | - | - |
787
+ | 0.3801 | 309 | 0.1733 | - | - |
788
+ | 0.3813 | 310 | 0.1891 | - | - |
789
+ | 0.3825 | 311 | 0.2128 | - | - |
790
+ | 0.3838 | 312 | 0.2042 | - | - |
791
+ | 0.3850 | 313 | 0.203 | - | - |
792
+ | 0.3862 | 314 | 0.2249 | - | - |
793
+ | 0.3875 | 315 | 0.1597 | 0.1014 | 0.9074 |
794
+ | 0.3887 | 316 | 0.1358 | - | - |
795
+ | 0.3899 | 317 | 0.207 | - | - |
796
+ | 0.3911 | 318 | 0.193 | - | - |
797
+ | 0.3924 | 319 | 0.1141 | - | - |
798
+ | 0.3936 | 320 | 0.2835 | - | - |
799
+ | 0.3948 | 321 | 0.2589 | - | - |
800
+ | 0.3961 | 322 | 0.088 | - | - |
801
+ | 0.3973 | 323 | 0.1675 | - | - |
802
+ | 0.3985 | 324 | 0.1525 | - | - |
803
+ | 0.3998 | 325 | 0.1401 | - | - |
804
+ | 0.4010 | 326 | 0.2109 | - | - |
805
+ | 0.4022 | 327 | 0.1382 | - | - |
806
+ | 0.4034 | 328 | 0.1724 | - | - |
807
+ | 0.4047 | 329 | 0.1668 | - | - |
808
+ | 0.4059 | 330 | 0.1606 | - | - |
809
+ | 0.4071 | 331 | 0.2102 | - | - |
810
+ | 0.4084 | 332 | 0.1737 | - | - |
811
+ | 0.4096 | 333 | 0.1641 | - | - |
812
+ | 0.4108 | 334 | 0.1984 | - | - |
813
+ | 0.4121 | 335 | 0.1395 | - | - |
814
+ | 0.4133 | 336 | 0.1236 | 0.1008 | 0.9066 |
815
+ | 0.4145 | 337 | 0.1405 | - | - |
816
+ | 0.4157 | 338 | 0.1461 | - | - |
817
+ | 0.4170 | 339 | 0.1151 | - | - |
818
+ | 0.4182 | 340 | 0.1282 | - | - |
819
+ | 0.4194 | 341 | 0.2155 | - | - |
820
+ | 0.4207 | 342 | 0.1344 | - | - |
821
+ | 0.4219 | 343 | 0.1854 | - | - |
822
+ | 0.4231 | 344 | 0.1766 | - | - |
823
+ | 0.4244 | 345 | 0.122 | - | - |
824
+ | 0.4256 | 346 | 0.142 | - | - |
825
+ | 0.4268 | 347 | 0.1434 | - | - |
826
+ | 0.4280 | 348 | 0.1687 | - | - |
827
+ | 0.4293 | 349 | 0.1751 | - | - |
828
+ | 0.4305 | 350 | 0.1253 | - | - |
829
+ | 0.4317 | 351 | 0.1387 | - | - |
830
+ | 0.4330 | 352 | 0.181 | - | - |
831
+ | 0.4342 | 353 | 0.101 | - | - |
832
+ | 0.4354 | 354 | 0.1552 | - | - |
833
+ | 0.4367 | 355 | 0.2676 | - | - |
834
+ | 0.4379 | 356 | 0.1638 | - | - |
835
+ | 0.4391 | 357 | 0.19 | 0.1008 | 0.9072 |
836
+ | 0.4403 | 358 | 0.1152 | - | - |
837
+ | 0.4416 | 359 | 0.1639 | - | - |
838
+ | 0.4428 | 360 | 0.1624 | - | - |
839
+ | 0.4440 | 361 | 0.203 | - | - |
840
+ | 0.4453 | 362 | 0.1856 | - | - |
841
+ | 0.4465 | 363 | 0.1978 | - | - |
842
+ | 0.4477 | 364 | 0.1457 | - | - |
843
+ | 0.4490 | 365 | 0.176 | - | - |
844
+ | 0.4502 | 366 | 0.1742 | - | - |
845
+ | 0.4514 | 367 | 0.1599 | - | - |
846
+ | 0.4526 | 368 | 0.2085 | - | - |
847
+ | 0.4539 | 369 | 0.2255 | - | - |
848
+ | 0.4551 | 370 | 0.1941 | - | - |
849
+ | 0.4563 | 371 | 0.0769 | - | - |
850
+ | 0.4576 | 372 | 0.2031 | - | - |
851
+ | 0.4588 | 373 | 0.2151 | - | - |
852
+ | 0.4600 | 374 | 0.2115 | - | - |
853
+ | 0.4613 | 375 | 0.1241 | - | - |
854
+ | 0.4625 | 376 | 0.1693 | - | - |
855
+ | 0.4637 | 377 | 0.2086 | - | - |
856
+ | 0.4649 | 378 | 0.1661 | 0.1004 | 0.9074 |
857
+ | 0.4662 | 379 | 0.1508 | - | - |
858
+ | 0.4674 | 380 | 0.1802 | - | - |
859
+ | 0.4686 | 381 | 0.1005 | - | - |
860
+ | 0.4699 | 382 | 0.1948 | - | - |
861
+ | 0.4711 | 383 | 0.1618 | - | - |
862
+ | 0.4723 | 384 | 0.216 | - | - |
863
+ | 0.4736 | 385 | 0.132 | - | - |
864
+ | 0.4748 | 386 | 0.2461 | - | - |
865
+ | 0.4760 | 387 | 0.1825 | - | - |
866
+ | 0.4772 | 388 | 0.1912 | - | - |
867
+ | 0.4785 | 389 | 0.1706 | - | - |
868
+ | 0.4797 | 390 | 0.2599 | - | - |
869
+ | 0.4809 | 391 | 0.1837 | - | - |
870
+ | 0.4822 | 392 | 0.23 | - | - |
871
+ | 0.4834 | 393 | 0.1523 | - | - |
872
+ | 0.4846 | 394 | 0.1105 | - | - |
873
+ | 0.4859 | 395 | 0.1478 | - | - |
874
+ | 0.4871 | 396 | 0.2184 | - | - |
875
+ | 0.4883 | 397 | 0.1977 | - | - |
876
+ | 0.4895 | 398 | 0.1607 | - | - |
877
+ | 0.4908 | 399 | 0.2183 | 0.1002 | 0.9077 |
878
+ | 0.4920 | 400 | 0.1155 | - | - |
879
+ | 0.4932 | 401 | 0.2395 | - | - |
880
+ | 0.4945 | 402 | 0.1194 | - | - |
881
+ | 0.4957 | 403 | 0.1567 | - | - |
882
+ | 0.4969 | 404 | 0.1037 | - | - |
883
+ | 0.4982 | 405 | 0.2713 | - | - |
884
+ | 0.4994 | 406 | 0.1742 | - | - |
885
+ | 0.5006 | 407 | 0.221 | - | - |
886
+ | 0.5018 | 408 | 0.1412 | - | - |
887
+ | 0.5031 | 409 | 0.1482 | - | - |
888
+ | 0.5043 | 410 | 0.1347 | - | - |
889
+ | 0.5055 | 411 | 0.2345 | - | - |
890
+ | 0.5068 | 412 | 0.1231 | - | - |
891
+ | 0.5080 | 413 | 0.1418 | - | - |
892
+ | 0.5092 | 414 | 0.152 | - | - |
893
+ | 0.5105 | 415 | 0.1878 | - | - |
894
+ | 0.5117 | 416 | 0.1683 | - | - |
895
+ | 0.5129 | 417 | 0.1501 | - | - |
896
+ | 0.5141 | 418 | 0.2589 | - | - |
897
+ | 0.5154 | 419 | 0.1924 | - | - |
898
+ | 0.5166 | 420 | 0.1166 | 0.0979 | 0.9078 |
899
+ | 0.5178 | 421 | 0.1509 | - | - |
900
+ | 0.5191 | 422 | 0.1457 | - | - |
901
+ | 0.5203 | 423 | 0.2244 | - | - |
902
+ | 0.5215 | 424 | 0.1837 | - | - |
903
+ | 0.5228 | 425 | 0.2649 | - | - |
904
+ | 0.5240 | 426 | 0.1295 | - | - |
905
+ | 0.5252 | 427 | 0.1776 | - | - |
906
+ | 0.5264 | 428 | 0.1949 | - | - |
907
+ | 0.5277 | 429 | 0.1262 | - | - |
908
+ | 0.5289 | 430 | 0.1502 | - | - |
909
+ | 0.5301 | 431 | 0.1927 | - | - |
910
+ | 0.5314 | 432 | 0.2161 | - | - |
911
+ | 0.5326 | 433 | 0.2082 | - | - |
912
+ | 0.5338 | 434 | 0.2171 | - | - |
913
+ | 0.5351 | 435 | 0.209 | - | - |
914
+ | 0.5363 | 436 | 0.1841 | - | - |
915
+ | 0.5375 | 437 | 0.1522 | - | - |
916
+ | 0.5387 | 438 | 0.1644 | - | - |
917
+ | 0.5400 | 439 | 0.1784 | - | - |
918
+ | 0.5412 | 440 | 0.2041 | - | - |
919
+ | 0.5424 | 441 | 0.1564 | 0.0968 | 0.9058 |
920
+ | 0.5437 | 442 | 0.2151 | - | - |
921
+ | 0.5449 | 443 | 0.1797 | - | - |
922
+ | 0.5461 | 444 | 0.1652 | - | - |
923
+ | 0.5474 | 445 | 0.1561 | - | - |
924
+ | 0.5486 | 446 | 0.1063 | - | - |
925
+ | 0.5498 | 447 | 0.1584 | - | - |
926
+ | 0.5510 | 448 | 0.2396 | - | - |
927
+ | 0.5523 | 449 | 0.1952 | - | - |
928
+ | 0.5535 | 450 | 0.1598 | - | - |
929
+ | 0.5547 | 451 | 0.2093 | - | - |
930
+ | 0.5560 | 452 | 0.1585 | - | - |
931
+ | 0.5572 | 453 | 0.2311 | - | - |
932
+ | 0.5584 | 454 | 0.1048 | - | - |
933
+ | 0.5597 | 455 | 0.1571 | - | - |
934
+ | 0.5609 | 456 | 0.1915 | - | - |
935
+ | 0.5621 | 457 | 0.1625 | - | - |
936
+ | 0.5633 | 458 | 0.1613 | - | - |
937
+ | 0.5646 | 459 | 0.1845 | - | - |
938
+ | 0.5658 | 460 | 0.2134 | - | - |
939
+ | 0.5670 | 461 | 0.2059 | - | - |
940
+ | 0.5683 | 462 | 0.1974 | 0.0947 | 0.9067 |
941
+ | 0.5695 | 463 | 0.1624 | - | - |
942
+ | 0.5707 | 464 | 0.2005 | - | - |
943
+ | 0.5720 | 465 | 0.1407 | - | - |
944
+ | 0.5732 | 466 | 0.1175 | - | - |
945
+ | 0.5744 | 467 | 0.1888 | - | - |
946
+ | 0.5756 | 468 | 0.1423 | - | - |
947
+ | 0.5769 | 469 | 0.1195 | - | - |
948
+ | 0.5781 | 470 | 0.1525 | - | - |
949
+ | 0.5793 | 471 | 0.2155 | - | - |
950
+ | 0.5806 | 472 | 0.2048 | - | - |
951
+ | 0.5818 | 473 | 0.2386 | - | - |
952
+ | 0.5830 | 474 | 0.162 | - | - |
953
+ | 0.5843 | 475 | 0.1735 | - | - |
954
+ | 0.5855 | 476 | 0.2067 | - | - |
955
+ | 0.5867 | 477 | 0.1395 | - | - |
956
+ | 0.5879 | 478 | 0.1482 | - | - |
957
+ | 0.5892 | 479 | 0.2399 | - | - |
958
+ | 0.5904 | 480 | 0.1849 | - | - |
959
+ | 0.5916 | 481 | 0.139 | - | - |
960
+ | 0.5929 | 482 | 0.2089 | - | - |
961
+ | 0.5941 | 483 | 0.2066 | 0.0934 | 0.9072 |
962
+ | 0.5953 | 484 | 0.2293 | - | - |
963
+ | 0.5966 | 485 | 0.1919 | - | - |
964
+ | 0.5978 | 486 | 0.1168 | - | - |
965
+ | 0.5990 | 487 | 0.2057 | - | - |
966
+ | 0.6002 | 488 | 0.1866 | - | - |
967
+ | 0.6015 | 489 | 0.2277 | - | - |
968
+ | 0.6027 | 490 | 0.1527 | - | - |
969
+ | 0.6039 | 491 | 0.275 | - | - |
970
+ | 0.6052 | 492 | 0.1212 | - | - |
971
+ | 0.6064 | 493 | 0.1384 | - | - |
972
+ | 0.6076 | 494 | 0.1611 | - | - |
973
+ | 0.6089 | 495 | 0.145 | - | - |
974
+ | 0.6101 | 496 | 0.1996 | - | - |
975
+ | 0.6113 | 497 | 0.3 | - | - |
976
+ | 0.6125 | 498 | 0.1117 | - | - |
977
+ | 0.6138 | 499 | 0.1905 | - | - |
978
+ | 0.6150 | 500 | 0.2221 | - | - |
979
+ | 0.6162 | 501 | 0.1749 | - | - |
980
+ | 0.6175 | 502 | 0.1533 | - | - |
981
+ | 0.6187 | 503 | 0.2268 | - | - |
982
+ | 0.6199 | 504 | 0.1879 | 0.0936 | 0.9066 |
983
+ | 0.6212 | 505 | 0.2956 | - | - |
984
+ | 0.6224 | 506 | 0.1566 | - | - |
985
+ | 0.6236 | 507 | 0.1612 | - | - |
986
+ | 0.6248 | 508 | 0.2312 | - | - |
987
+ | 0.6261 | 509 | 0.181 | - | - |
988
+ | 0.6273 | 510 | 0.235 | - | - |
989
+ | 0.6285 | 511 | 0.1376 | - | - |
990
+ | 0.6298 | 512 | 0.1066 | - | - |
991
+ | 0.6310 | 513 | 0.2235 | - | - |
992
+ | 0.6322 | 514 | 0.2549 | - | - |
993
+ | 0.6335 | 515 | 0.2676 | - | - |
994
+ | 0.6347 | 516 | 0.1652 | - | - |
995
+ | 0.6359 | 517 | 0.1573 | - | - |
996
+ | 0.6371 | 518 | 0.2106 | - | - |
997
+ | 0.6384 | 519 | 0.151 | - | - |
998
+ | 0.6396 | 520 | 0.1491 | - | - |
999
+ | 0.6408 | 521 | 0.2612 | - | - |
1000
+ | 0.6421 | 522 | 0.1287 | - | - |
1001
+ | 0.6433 | 523 | 0.2084 | - | - |
1002
+ | 0.6445 | 524 | 0.1545 | - | - |
1003
+ | 0.6458 | 525 | 0.1946 | 0.0931 | 0.9061 |
1004
+ | 0.6470 | 526 | 0.1684 | - | - |
1005
+ | 0.6482 | 527 | 0.1974 | - | - |
1006
+ | 0.6494 | 528 | 0.2448 | - | - |
1007
+ | 0.6507 | 529 | 0.2255 | - | - |
1008
+ | 0.6519 | 530 | 0.2157 | - | - |
1009
+ | 0.6531 | 531 | 0.1948 | - | - |
1010
+ | 0.6544 | 532 | 0.1418 | - | - |
1011
+ | 0.6556 | 533 | 0.1683 | - | - |
1012
+ | 0.6568 | 534 | 0.193 | - | - |
1013
+ | 0.6581 | 535 | 0.2341 | - | - |
1014
+ | 0.6593 | 536 | 0.131 | - | - |
1015
+ | 0.6605 | 537 | 0.1733 | - | - |
1016
+ | 0.6617 | 538 | 0.1489 | - | - |
1017
+ | 0.6630 | 539 | 0.1918 | - | - |
1018
+ | 0.6642 | 540 | 0.1953 | - | - |
1019
+ | 0.6654 | 541 | 0.1421 | - | - |
1020
+ | 0.6667 | 542 | 0.2214 | - | - |
1021
+ | 0.6679 | 543 | 0.2152 | - | - |
1022
+ | 0.6691 | 544 | 0.209 | - | - |
1023
+ | 0.6704 | 545 | 0.1735 | - | - |
1024
+ | 0.6716 | 546 | 0.2048 | 0.0918 | 0.9060 |
1025
+ | 0.6728 | 547 | 0.1721 | - | - |
1026
+ | 0.6740 | 548 | 0.1838 | - | - |
1027
+ | 0.6753 | 549 | 0.1614 | - | - |
1028
+ | 0.6765 | 550 | 0.1999 | - | - |
1029
+ | 0.6777 | 551 | 0.0984 | - | - |
1030
+ | 0.6790 | 552 | 0.1351 | - | - |
1031
+ | 0.6802 | 553 | 0.1886 | - | - |
1032
+ | 0.6814 | 554 | 0.1148 | - | - |
1033
+ | 0.6827 | 555 | 0.1766 | - | - |
1034
+ | 0.6839 | 556 | 0.19 | - | - |
1035
+ | 0.6851 | 557 | 0.2082 | - | - |
1036
+ | 0.6863 | 558 | 0.222 | - | - |
1037
+ | 0.6876 | 559 | 0.2032 | - | - |
1038
+ | 0.6888 | 560 | 0.1854 | - | - |
1039
+ | 0.6900 | 561 | 0.1473 | - | - |
1040
+ | 0.6913 | 562 | 0.2003 | - | - |
1041
+ | 0.6925 | 563 | 0.1223 | - | - |
1042
+ | 0.6937 | 564 | 0.2319 | - | - |
1043
+ | 0.6950 | 565 | 0.0761 | - | - |
1044
+ | 0.6962 | 566 | 0.2835 | - | - |
1045
+ | 0.6974 | 567 | 0.2331 | 0.0920 | 0.9061 |
1046
+ | 0.6986 | 568 | 0.1698 | - | - |
1047
+ | 0.6999 | 569 | 0.203 | - | - |
1048
+ | 0.7011 | 570 | 0.2344 | - | - |
1049
+ | 0.7023 | 571 | 0.1823 | - | - |
1050
+ | 0.7036 | 572 | 0.2043 | - | - |
1051
+ | 0.7048 | 573 | 0.1881 | - | - |
1052
+ | 0.7060 | 574 | 0.1599 | - | - |
1053
+ | 0.7073 | 575 | 0.0829 | - | - |
1054
+ | 0.7085 | 576 | 0.1816 | - | - |
1055
+ | 0.7097 | 577 | 0.1801 | - | - |
1056
+ | 0.7109 | 578 | 0.1707 | - | - |
1057
+ | 0.7122 | 579 | 0.2306 | - | - |
1058
+ | 0.7134 | 580 | 0.1503 | - | - |
1059
+ | 0.7146 | 581 | 0.1779 | - | - |
1060
+ | 0.7159 | 582 | 0.1422 | - | - |
1061
+ | 0.7171 | 583 | 0.1358 | - | - |
1062
+ | 0.7183 | 584 | 0.0978 | - | - |
1063
+ | 0.7196 | 585 | 0.1713 | - | - |
1064
+ | 0.7208 | 586 | 0.1771 | - | - |
1065
+ | 0.7220 | 587 | 0.1241 | - | - |
1066
+ | 0.7232 | 588 | 0.1267 | 0.0918 | 0.9064 |
1067
+ | 0.7245 | 589 | 0.1126 | - | - |
1068
+ | 0.7257 | 590 | 0.0858 | - | - |
1069
+ | 0.7269 | 591 | 0.1335 | - | - |
1070
+ | 0.7282 | 592 | 0.1958 | - | - |
1071
+ | 0.7294 | 593 | 0.1448 | - | - |
1072
+ | 0.7306 | 594 | 0.2679 | - | - |
1073
+ | 0.7319 | 595 | 0.153 | - | - |
1074
+ | 0.7331 | 596 | 0.1523 | - | - |
1075
+ | 0.7343 | 597 | 0.1988 | - | - |
1076
+ | 0.7355 | 598 | 0.157 | - | - |
1077
+ | 0.7368 | 599 | 0.146 | - | - |
1078
+ | 0.7380 | 600 | 0.2043 | - | - |
1079
+ | 0.7392 | 601 | 0.1508 | - | - |
1080
+ | 0.7405 | 602 | 0.1946 | - | - |
1081
+ | 0.7417 | 603 | 0.1481 | - | - |
1082
+ | 0.7429 | 604 | 0.0995 | - | - |
1083
+ | 0.7442 | 605 | 0.149 | - | - |
1084
+ | 0.7454 | 606 | 0.1686 | - | - |
1085
+ | 0.7466 | 607 | 0.1555 | - | - |
1086
+ | 0.7478 | 608 | 0.1662 | - | - |
1087
+ | 0.7491 | 609 | 0.1217 | 0.0917 | 0.9064 |
1088
+ | 0.7503 | 610 | 0.0748 | - | - |
1089
+ | 0.7515 | 611 | 0.1723 | - | - |
1090
+ | 0.7528 | 612 | 0.2354 | - | - |
1091
+ | 0.7540 | 613 | 0.1315 | - | - |
1092
+ | 0.7552 | 614 | 0.2913 | - | - |
1093
+ | 0.7565 | 615 | 0.0991 | - | - |
1094
+ | 0.7577 | 616 | 0.1052 | - | - |
1095
+ | 0.7589 | 617 | 0.1496 | - | - |
1096
+ | 0.7601 | 618 | 0.1399 | - | - |
1097
+ | 0.7614 | 619 | 0.1329 | - | - |
1098
+ | 0.7626 | 620 | 0.2287 | - | - |
1099
+ | 0.7638 | 621 | 0.1085 | - | - |
1100
+ | 0.7651 | 622 | 0.1864 | - | - |
1101
+ | 0.7663 | 623 | 0.1577 | - | - |
1102
+ | 0.7675 | 624 | 0.143 | - | - |
1103
+ | 0.7688 | 625 | 0.1886 | - | - |
1104
+ | 0.7700 | 626 | 0.1683 | - | - |
1105
+ | 0.7712 | 627 | 0.212 | - | - |
1106
+ | 0.7724 | 628 | 0.1643 | - | - |
1107
+ | 0.7737 | 629 | 0.1632 | - | - |
1108
+ | 0.7749 | 630 | 0.1384 | 0.0925 | 0.9054 |
1109
+ | 0.7761 | 631 | 0.2133 | - | - |
1110
+ | 0.7774 | 632 | 0.1732 | - | - |
1111
+ | 0.7786 | 633 | 0.1218 | - | - |
1112
+ | 0.7798 | 634 | 0.1581 | - | - |
1113
+ | 0.7811 | 635 | 0.1337 | - | - |
1114
+ | 0.7823 | 636 | 0.1859 | - | - |
1115
+ | 0.7835 | 637 | 0.1616 | - | - |
1116
+ | 0.7847 | 638 | 0.1799 | - | - |
1117
+ | 0.7860 | 639 | 0.1193 | - | - |
1118
+ | 0.7872 | 640 | 0.1471 | - | - |
1119
+ | 0.7884 | 641 | 0.1235 | - | - |
1120
+ | 0.7897 | 642 | 0.1221 | - | - |
1121
+ | 0.7909 | 643 | 0.1379 | - | - |
1122
+ | 0.7921 | 644 | 0.238 | - | - |
1123
+ | 0.7934 | 645 | 0.1671 | - | - |
1124
+ | 0.7946 | 646 | 0.1652 | - | - |
1125
+ | 0.7958 | 647 | 0.1828 | - | - |
1126
+ | 0.7970 | 648 | 0.2207 | - | - |
1127
+ | 0.7983 | 649 | 0.2109 | - | - |
1128
+ | 0.7995 | 650 | 0.1105 | - | - |
1129
+ | 0.8007 | 651 | 0.129 | 0.0933 | 0.9069 |
1130
+ | 0.8020 | 652 | 0.1633 | - | - |
1131
+ | 0.8032 | 653 | 0.201 | - | - |
1132
+ | 0.8044 | 654 | 0.1041 | - | - |
1133
+ | 0.8057 | 655 | 0.1838 | - | - |
1134
+ | 0.8069 | 656 | 0.3044 | - | - |
1135
+ | 0.8081 | 657 | 0.1736 | - | - |
1136
+ | 0.8093 | 658 | 0.1909 | - | - |
1137
+ | 0.8106 | 659 | 0.1413 | - | - |
1138
+ | 0.8118 | 660 | 0.1138 | - | - |
1139
+ | 0.8130 | 661 | 0.1163 | - | - |
1140
+ | 0.8143 | 662 | 0.1725 | - | - |
1141
+ | 0.8155 | 663 | 0.2248 | - | - |
1142
+ | 0.8167 | 664 | 0.1019 | - | - |
1143
+ | 0.8180 | 665 | 0.1138 | - | - |
1144
+ | 0.8192 | 666 | 0.1652 | - | - |
1145
+ | 0.8204 | 667 | 0.1361 | - | - |
1146
+ | 0.8216 | 668 | 0.1769 | - | - |
1147
+ | 0.8229 | 669 | 0.1241 | - | - |
1148
+ | 0.8241 | 670 | 0.1683 | - | - |
1149
+ | 0.8253 | 671 | 0.1315 | - | - |
1150
+ | 0.8266 | 672 | 0.1046 | 0.0940 | 0.9055 |
1151
+ | 0.8278 | 673 | 0.1984 | - | - |
1152
+ | 0.8290 | 674 | 0.1766 | - | - |
1153
+ | 0.8303 | 675 | 0.1245 | - | - |
1154
+ | 0.8315 | 676 | 0.1953 | - | - |
1155
+ | 0.8327 | 677 | 0.1506 | - | - |
1156
+ | 0.8339 | 678 | 0.1145 | - | - |
1157
+ | 0.8352 | 679 | 0.1366 | - | - |
1158
+ | 0.8364 | 680 | 0.1071 | - | - |
1159
+ | 0.8376 | 681 | 0.2142 | - | - |
1160
+ | 0.8389 | 682 | 0.2029 | - | - |
1161
+ | 0.8401 | 683 | 0.1171 | - | - |
1162
+ | 0.8413 | 684 | 0.176 | - | - |
1163
+ | 0.8426 | 685 | 0.1052 | - | - |
1164
+ | 0.8438 | 686 | 0.1892 | - | - |
1165
+ | 0.8450 | 687 | 0.1499 | - | - |
1166
+ | 0.8462 | 688 | 0.1414 | - | - |
1167
+ | 0.8475 | 689 | 0.1193 | - | - |
1168
+ | 0.8487 | 690 | 0.1516 | - | - |
1169
+ | 0.8499 | 691 | 0.1552 | - | - |
1170
+ | 0.8512 | 692 | 0.1168 | - | - |
1171
+ | 0.8524 | 693 | 0.2326 | 0.0932 | 0.9071 |
1172
+ | 0.8536 | 694 | 0.2112 | - | - |
1173
+ | 0.8549 | 695 | 0.0835 | - | - |
1174
+ | 0.8561 | 696 | 0.1512 | - | - |
1175
+ | 0.8573 | 697 | 0.1379 | - | - |
1176
+ | 0.8585 | 698 | 0.1045 | - | - |
1177
+ | 0.8598 | 699 | 0.2045 | - | - |
1178
+ | 0.8610 | 700 | 0.1909 | - | - |
1179
+ | 0.8622 | 701 | 0.1895 | - | - |
1180
+ | 0.8635 | 702 | 0.2077 | - | - |
1181
+ | 0.8647 | 703 | 0.1199 | - | - |
1182
+ | 0.8659 | 704 | 0.1606 | - | - |
1183
+ | 0.8672 | 705 | 0.1501 | - | - |
1184
+ | 0.8684 | 706 | 0.1711 | - | - |
1185
+ | 0.8696 | 707 | 0.222 | - | - |
1186
+ | 0.8708 | 708 | 0.1414 | - | - |
1187
+ | 0.8721 | 709 | 0.1972 | - | - |
1188
+ | 0.8733 | 710 | 0.1074 | - | - |
1189
+ | 0.8745 | 711 | 0.2044 | - | - |
1190
+ | 0.8758 | 712 | 0.0997 | - | - |
1191
+ | 0.8770 | 713 | 0.1178 | - | - |
1192
+ | 0.8782 | 714 | 0.1376 | 0.0929 | 0.9058 |
1193
+ | 0.8795 | 715 | 0.1302 | - | - |
1194
+ | 0.8807 | 716 | 0.1252 | - | - |
1195
+ | 0.8819 | 717 | 0.2365 | - | - |
1196
+ | 0.8831 | 718 | 0.1405 | - | - |
1197
+ | 0.8844 | 719 | 0.1806 | - | - |
1198
+ | 0.8856 | 720 | 0.1495 | - | - |
1199
+ | 0.8868 | 721 | 0.1987 | - | - |
1200
+ | 0.8881 | 722 | 0.096 | - | - |
1201
+ | 0.8893 | 723 | 0.1728 | - | - |
1202
+ | 0.8905 | 724 | 0.2104 | - | - |
1203
+ | 0.8918 | 725 | 0.1562 | - | - |
1204
+ | 0.8930 | 726 | 0.1358 | - | - |
1205
+ | 0.8942 | 727 | 0.1723 | - | - |
1206
+ | 0.8954 | 728 | 0.1947 | - | - |
1207
+ | 0.8967 | 729 | 0.1572 | - | - |
1208
+ | 0.8979 | 730 | 0.1124 | - | - |
1209
+ | 0.8991 | 731 | 0.2272 | - | - |
1210
+ | 0.9004 | 732 | 0.1356 | - | - |
1211
+ | 0.9016 | 733 | 0.1816 | - | - |
1212
+ | 0.9028 | 734 | 0.1011 | - | - |
1213
+ | 0.9041 | 735 | 0.124 | 0.0911 | 0.9051 |
1214
+ | 0.9053 | 736 | 0.1873 | - | - |
1215
+ | 0.9065 | 737 | 0.0702 | - | - |
1216
+ | 0.9077 | 738 | 0.15 | - | - |
1217
+ | 0.9090 | 739 | 0.221 | - | - |
1218
+ | 0.9102 | 740 | 0.1511 | - | - |
1219
+ | 0.9114 | 741 | 0.195 | - | - |
1220
+ | 0.9127 | 742 | 0.1473 | - | - |
1221
+ | 0.9139 | 743 | 0.1311 | - | - |
1222
+ | 0.9151 | 744 | 0.1869 | - | - |
1223
+ | 0.9164 | 745 | 0.1433 | - | - |
1224
+ | 0.9176 | 746 | 0.1286 | - | - |
1225
+ | 0.9188 | 747 | 0.1316 | - | - |
1226
+ | 0.9200 | 748 | 0.1669 | - | - |
1227
+ | 0.9213 | 749 | 0.1691 | - | - |
1228
+ | 0.9225 | 750 | 0.1853 | - | - |
1229
+ | 0.9237 | 751 | 0.1813 | - | - |
1230
+ | 0.9250 | 752 | 0.1754 | - | - |
1231
+ | 0.9262 | 753 | 0.2282 | - | - |
1232
+ | 0.9274 | 754 | 0.1248 | - | - |
1233
+ | 0.9287 | 755 | 0.1182 | - | - |
1234
+ | 0.9299 | 756 | 0.1601 | 0.0903 | 0.9059 |
1235
+ | 0.9311 | 757 | 0.2377 | - | - |
1236
+ | 0.9323 | 758 | 0.1799 | - | - |
1237
+ | 0.9336 | 759 | 0.2016 | - | - |
1238
+ | 0.9348 | 760 | 0.1293 | - | - |
1239
+ | 0.9360 | 761 | 0.2038 | - | - |
1240
+ | 0.9373 | 762 | 0.1384 | - | - |
1241
+ | 0.9385 | 763 | 0.1856 | - | - |
1242
+ | 0.9397 | 764 | 0.2775 | - | - |
1243
+ | 0.9410 | 765 | 0.1651 | - | - |
1244
+ | 0.9422 | 766 | 0.2072 | - | - |
1245
+ | 0.9434 | 767 | 0.1459 | - | - |
1246
+ | 0.9446 | 768 | 0.1277 | - | - |
1247
+ | 0.9459 | 769 | 0.1742 | - | - |
1248
+ | 0.9471 | 770 | 0.1978 | - | - |
1249
+ | 0.9483 | 771 | 0.1992 | - | - |
1250
+ | 0.9496 | 772 | 0.1649 | - | - |
1251
+ | 0.9508 | 773 | 0.2195 | - | - |
1252
+ | 0.9520 | 774 | 0.1348 | - | - |
1253
+ | 0.9533 | 775 | 0.1556 | - | - |
1254
+ | 0.9545 | 776 | 0.2293 | - | - |
1255
+ | 0.9557 | 777 | 0.1585 | 0.0904 | 0.9062 |
1256
+ | 0.9569 | 778 | 0.1029 | - | - |
1257
+ | 0.9582 | 779 | 0.1027 | - | - |
1258
+ | 0.9594 | 780 | 0.1165 | - | - |
1259
+ | 0.9606 | 781 | 0.1654 | - | - |
1260
+ | 0.9619 | 782 | 0.1706 | - | - |
1261
+ | 0.9631 | 783 | 0.102 | - | - |
1262
+ | 0.9643 | 784 | 0.1697 | - | - |
1263
+ | 0.9656 | 785 | 0.177 | - | - |
1264
+ | 0.9668 | 786 | 0.1718 | - | - |
1265
+ | 0.9680 | 787 | 0.1542 | - | - |
1266
+ | 0.9692 | 788 | 0.1654 | - | - |
1267
+ | 0.9705 | 789 | 0.1672 | - | - |
1268
+ | 0.9717 | 790 | 0.1867 | - | - |
1269
+ | 0.9729 | 791 | 0.1717 | - | - |
1270
+ | 0.9742 | 792 | 0.1701 | - | - |
1271
+ | 0.9754 | 793 | 0.1542 | - | - |
1272
+ | 0.9766 | 794 | 0.2153 | - | - |
1273
+ | 0.9779 | 795 | 0.131 | - | - |
1274
+ | 0.9791 | 796 | 0.1448 | - | - |
1275
+ | 0.9803 | 797 | 0.1171 | - | - |
1276
+ | 0.9815 | 798 | 0.1585 | 0.0904 | 0.9063 |
1277
+ | 0.9828 | 799 | 0.1352 | - | - |
1278
+ | 0.9840 | 800 | 0.1146 | - | - |
1279
+ | 0.9852 | 801 | 0.1366 | - | - |
1280
+ | 0.9865 | 802 | 0.1375 | - | - |
1281
+ | 0.9877 | 803 | 0.1588 | - | - |
1282
+ | 0.9889 | 804 | 0.1429 | - | - |
1283
+ | 0.9902 | 805 | 0.1541 | - | - |
1284
+ | 0.9914 | 806 | 0.1171 | - | - |
1285
+ | 0.9926 | 807 | 0.1352 | - | - |
1286
+ | 0.9938 | 808 | 0.1948 | - | - |
1287
+ | 0.9951 | 809 | 0.1628 | - | - |
1288
+ | 0.9963 | 810 | 0.1115 | - | - |
1289
+ | 0.9975 | 811 | 0.0929 | - | - |
1290
+ | 0.9988 | 812 | 0.0955 | - | - |
1291
+ | 1.0 | 813 | 0.0 | 0.0904 | 0.9063 |
1292
+
1293
+ </details>
1294
+
1295
+ ### Framework Versions
1296
+ - Python: 3.10.14
1297
+ - Sentence Transformers: 3.0.1
1298
+ - Transformers: 4.44.0
1299
+ - PyTorch: 2.4.0
1300
+ - Accelerate: 0.33.0
1301
+ - Datasets: 2.21.0
1302
+ - Tokenizers: 0.19.1
1303
+
1304
+ ## Citation
1305
+
1306
+ ### BibTeX
1307
+
1308
+ #### Sentence Transformers
1309
+ ```bibtex
1310
+ @inproceedings{reimers-2019-sentence-bert,
1311
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
1312
+ author = "Reimers, Nils and Gurevych, Iryna",
1313
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
1314
+ month = "11",
1315
+ year = "2019",
1316
+ publisher = "Association for Computational Linguistics",
1317
+ url = "https://arxiv.org/abs/1908.10084",
1318
+ }
1319
+ ```
1320
+
1321
+ <!--
1322
+ ## Glossary
1323
+
1324
+ *Clearly define terms in order to be accessible across audiences.*
1325
+ -->
1326
+
1327
+ <!--
1328
+ ## Model Card Authors
1329
+
1330
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
1331
+ -->
1332
+
1333
+ <!--
1334
+ ## Model Card Contact
1335
+
1336
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
1337
+ -->
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bobox/DeBERTa-small-ST-v1-test-step3",
3
+ "architectures": [
4
+ "DebertaV2Model"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 3072,
12
+ "layer_norm_eps": 1e-07,
13
+ "max_position_embeddings": 512,
14
+ "max_relative_positions": -1,
15
+ "model_type": "deberta-v2",
16
+ "norm_rel_ebd": "layer_norm",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 0,
20
+ "pooler_dropout": 0,
21
+ "pooler_hidden_act": "gelu",
22
+ "pooler_hidden_size": 768,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "position_buckets": 256,
29
+ "relative_attention": true,
30
+ "share_att_key": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.44.0",
33
+ "type_vocab_size": 0,
34
+ "vocab_size": 128100
35
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.0.1",
4
+ "transformers": "4.44.0",
5
+ "pytorch": "2.4.0"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": null
10
+ }
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b56c85d7da789a73c171d3d472df3da4e2aba689378408458a5a1ea53f77121e
3
+ size 565251810
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "[CLS]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "[SEP]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "[MASK]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "[SEP]",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "max_length": 512,
51
+ "model_max_length": 512,
52
+ "pad_to_multiple_of": null,
53
+ "pad_token": "[PAD]",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
+ "sep_token": "[SEP]",
57
+ "sp_model_kwargs": {},
58
+ "split_by_punct": false,
59
+ "stride": 0,
60
+ "tokenizer_class": "DebertaV2Tokenizer",
61
+ "truncation_side": "right",
62
+ "truncation_strategy": "longest_first",
63
+ "unk_token": "[UNK]",
64
+ "vocab_type": "spm"
65
+ }