File size: 227,274 Bytes
e2767b4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 |
---
base_model: microsoft/deberta-v3-small
datasets:
- sentence-transformers/all-nli
- sentence-transformers/stsb
- tals/vitaminc
- nyu-mll/glue
- allenai/scitail
- sentence-transformers/xsum
- sentence-transformers/sentence-compression
- allenai/sciq
- allenai/qasc
- sentence-transformers/msmarco-msmarco-distilbert-base-v3
- sentence-transformers/natural-questions
- sentence-transformers/trivia-qa
- sentence-transformers/quora-duplicates
- sentence-transformers/gooaq
language:
- en
library_name: sentence-transformers
metrics:
- pearson_cosine
- spearman_cosine
- pearson_manhattan
- spearman_manhattan
- pearson_euclidean
- spearman_euclidean
- pearson_dot
- spearman_dot
- pearson_max
- spearman_max
pipeline_tag: sentence-similarity
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- generated_from_trainer
- dataset_size:267363
- loss:AdaptiveLayerLoss
- loss:CoSENTLoss
- loss:GISTEmbedLoss
- loss:OnlineContrastiveLoss
- loss:MultipleNegativesSymmetricRankingLoss
- loss:MultipleNegativesRankingLoss
widget:
- source_sentence: The term technology describes the application of knowledge to real-world
problems and is practiced by engineers.
sentences:
- What term describes the application of knowledge to real-world problems and is
practiced by engineers?
- What bodily function do the triceps help perform?
- Exposure to what can increase the amount of pigment in the skin and make it appear
darker?
- source_sentence: Most of food's chemical energy ultimately comes from sunlight.
sentences:
- Catecholamines are a class of amine hormones synthesised form which amino acid?
- Most of food's chemical energy ultimately comes from what source?
- How many types of bosons are there?
- source_sentence: Someone is shredding cabbage leaves with a knife.
sentences:
- The man is erasing the chalk board.
- Kittens are eating from dishes.
- Someone is chopping some cabbage leaves.
- source_sentence: Three men are dancing.
sentences:
- Women are dancing.
- The woman is pouring oil into the pan.
- A man is dancing.
- source_sentence: What percentage of Warsaw's population was Protestant in 1901?
sentences:
- (The name of the theater refers to a well-known landmark water tower, which is
actually in another nearby area).
- This is primarily accomplished through normal faulting and through the ductile
stretching and thinning.
- After the war, the new communist authorities of Poland discouraged church construction
and only a small number were rebuilt.
model-index:
- name: SentenceTransformer based on microsoft/deberta-v3-small
results:
- task:
type: semantic-similarity
name: Semantic Similarity
dataset:
name: sts test
type: sts-test
metrics:
- type: pearson_cosine
value: 0.7002978854888552
name: Pearson Cosine
- type: spearman_cosine
value: 0.6718756239728468
name: Spearman Cosine
- type: pearson_manhattan
value: 0.7050517306003169
name: Pearson Manhattan
- type: spearman_manhattan
value: 0.6824201536078427
name: Spearman Manhattan
- type: pearson_euclidean
value: 0.6963744527231541
name: Pearson Euclidean
- type: spearman_euclidean
value: 0.6742379556154348
name: Spearman Euclidean
- type: pearson_dot
value: 0.5685392445320393
name: Pearson Dot
- type: spearman_dot
value: 0.5416448961602434
name: Spearman Dot
- type: pearson_max
value: 0.7050517306003169
name: Pearson Max
- type: spearman_max
value: 0.6824201536078427
name: Spearman Max
---
# SentenceTransformer based on microsoft/deberta-v3-small
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) on the [nli-pairs](https://huggingface.co/datasets/sentence-transformers/all-nli), [sts-label](https://huggingface.co/datasets/sentence-transformers/stsb), [vitaminc-pairs](https://huggingface.co/datasets/tals/vitaminc), [qnli-contrastive](https://huggingface.co/datasets/nyu-mll/glue), [scitail-pairs-qa](https://huggingface.co/datasets/allenai/scitail), [scitail-pairs-pos](https://huggingface.co/datasets/allenai/scitail), [xsum-pairs](https://huggingface.co/datasets/sentence-transformers/xsum), [compression-pairs](https://huggingface.co/datasets/sentence-transformers/sentence-compression), [compression-pairs2](https://huggingface.co/datasets/sentence-transformers/sentence-compression), [compression-pairs3](https://huggingface.co/datasets/sentence-transformers/sentence-compression), [sciq_pairs](https://huggingface.co/datasets/allenai/sciq), [qasc_pairs](https://huggingface.co/datasets/allenai/qasc), openbookqa_pairs, [msmarco_pairs](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3), [msmarco_pairs2](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3), [nq_pairs](https://huggingface.co/datasets/sentence-transformers/natural-questions), [nq_pairs2](https://huggingface.co/datasets/sentence-transformers/natural-questions), [trivia_pairs](https://huggingface.co/datasets/sentence-transformers/trivia-qa), [quora_pairs](https://huggingface.co/datasets/sentence-transformers/quora-duplicates), [gooaq_pairs](https://huggingface.co/datasets/sentence-transformers/gooaq), [gooaq_pairs2](https://huggingface.co/datasets/sentence-transformers/gooaq) and [mrpc_pairs](https://huggingface.co/datasets/nyu-mll/glue) datasets. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
## Model Details
### Model Description
- **Model Type:** Sentence Transformer
- **Base model:** [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) <!-- at revision a36c739020e01763fe789b4b85e2df55d6180012 -->
- **Maximum Sequence Length:** 512 tokens
- **Output Dimensionality:** 768 tokens
- **Similarity Function:** Cosine Similarity
- **Training Datasets:**
- [nli-pairs](https://huggingface.co/datasets/sentence-transformers/all-nli)
- [sts-label](https://huggingface.co/datasets/sentence-transformers/stsb)
- [vitaminc-pairs](https://huggingface.co/datasets/tals/vitaminc)
- [qnli-contrastive](https://huggingface.co/datasets/nyu-mll/glue)
- [scitail-pairs-qa](https://huggingface.co/datasets/allenai/scitail)
- [scitail-pairs-pos](https://huggingface.co/datasets/allenai/scitail)
- [xsum-pairs](https://huggingface.co/datasets/sentence-transformers/xsum)
- [compression-pairs](https://huggingface.co/datasets/sentence-transformers/sentence-compression)
- [compression-pairs2](https://huggingface.co/datasets/sentence-transformers/sentence-compression)
- [compression-pairs3](https://huggingface.co/datasets/sentence-transformers/sentence-compression)
- [sciq_pairs](https://huggingface.co/datasets/allenai/sciq)
- [qasc_pairs](https://huggingface.co/datasets/allenai/qasc)
- openbookqa_pairs
- [msmarco_pairs](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3)
- [msmarco_pairs2](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3)
- [nq_pairs](https://huggingface.co/datasets/sentence-transformers/natural-questions)
- [nq_pairs2](https://huggingface.co/datasets/sentence-transformers/natural-questions)
- [trivia_pairs](https://huggingface.co/datasets/sentence-transformers/trivia-qa)
- [quora_pairs](https://huggingface.co/datasets/sentence-transformers/quora-duplicates)
- [gooaq_pairs](https://huggingface.co/datasets/sentence-transformers/gooaq)
- [gooaq_pairs2](https://huggingface.co/datasets/sentence-transformers/gooaq)
- [mrpc_pairs](https://huggingface.co/datasets/nyu-mll/glue)
- **Language:** en
<!-- - **License:** Unknown -->
### Model Sources
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
### Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: DebertaV2Model
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)
```
## Usage
### Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
```bash
pip install -U sentence-transformers
```
Then you can load this model and run inference.
```python
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("bobox/DeBERTa-ST-AllLayers-v3-checkpoints-tmp")
# Run inference
sentences = [
"What percentage of Warsaw's population was Protestant in 1901?",
'After the war, the new communist authorities of Poland discouraged church construction and only a small number were rebuilt.',
'(The name of the theater refers to a well-known landmark water tower, which is actually in another nearby area).',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 768]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
```
<!--
### Direct Usage (Transformers)
<details><summary>Click to see the direct usage in Transformers</summary>
</details>
-->
<!--
### Downstream Usage (Sentence Transformers)
You can finetune this model on your own dataset.
<details><summary>Click to expand</summary>
</details>
-->
<!--
### Out-of-Scope Use
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->
## Evaluation
### Metrics
#### Semantic Similarity
* Dataset: `sts-test`
* Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
| Metric | Value |
|:--------------------|:-----------|
| pearson_cosine | 0.7003 |
| **spearman_cosine** | **0.6719** |
| pearson_manhattan | 0.7051 |
| spearman_manhattan | 0.6824 |
| pearson_euclidean | 0.6964 |
| spearman_euclidean | 0.6742 |
| pearson_dot | 0.5685 |
| spearman_dot | 0.5416 |
| pearson_max | 0.7051 |
| spearman_max | 0.6824 |
<!--
## Bias, Risks and Limitations
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->
<!--
### Recommendations
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->
## Training Details
### Training Datasets
#### nli-pairs
* Dataset: [nli-pairs](https://huggingface.co/datasets/sentence-transformers/all-nli) at [d482672](https://huggingface.co/datasets/sentence-transformers/all-nli/tree/d482672c8e74ce18da116f430137434ba2e52fab)
* Size: 25,000 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 5 tokens</li><li>mean: 16.62 tokens</li><li>max: 62 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 9.46 tokens</li><li>max: 29 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:---------------------------------------------------------------------------|:-------------------------------------------------|
| <code>A person on a horse jumps over a broken down airplane.</code> | <code>A person is outdoors, on a horse.</code> |
| <code>Children smiling and waving at camera</code> | <code>There are children present</code> |
| <code>A boy is jumping on skateboard in the middle of a red bridge.</code> | <code>The boy does a skateboarding trick.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### sts-label
* Dataset: [sts-label](https://huggingface.co/datasets/sentence-transformers/stsb) at [ab7a5ac](https://huggingface.co/datasets/sentence-transformers/stsb/tree/ab7a5ac0e35aa22088bdcf23e7fd99b220e53308)
* Size: 5,749 training samples
* Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>score</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 | score |
|:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
| type | string | string | float |
| details | <ul><li>min: 6 tokens</li><li>mean: 9.81 tokens</li><li>max: 27 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 9.74 tokens</li><li>max: 25 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.54</li><li>max: 1.0</li></ul> |
* Samples:
| sentence1 | sentence2 | score |
|:-----------------------------------------------------------|:----------------------------------------------------------------------|:------------------|
| <code>A plane is taking off.</code> | <code>An air plane is taking off.</code> | <code>1.0</code> |
| <code>A man is playing a large flute.</code> | <code>A man is playing a flute.</code> | <code>0.76</code> |
| <code>A man is spreading shreded cheese on a pizza.</code> | <code>A man is spreading shredded cheese on an uncooked pizza.</code> | <code>0.76</code> |
* Loss: [<code>CoSENTLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosentloss) with these parameters:
```json
{
"scale": 20.0,
"similarity_fct": "pairwise_cos_sim"
}
```
#### vitaminc-pairs
* Dataset: [vitaminc-pairs](https://huggingface.co/datasets/tals/vitaminc) at [be6febb](https://huggingface.co/datasets/tals/vitaminc/tree/be6febb761b0b2807687e61e0b5282e459df2fa0)
* Size: 25,000 training samples
* Columns: <code>claim</code> and <code>evidence</code>
* Approximate statistics based on the first 1000 samples:
| | claim | evidence |
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 6 tokens</li><li>mean: 16.71 tokens</li><li>max: 54 tokens</li></ul> | <ul><li>min: 8 tokens</li><li>mean: 37.14 tokens</li><li>max: 260 tokens</li></ul> |
* Samples:
| claim | evidence |
|:-------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>Manchester had a population of more than 540,000 in 2017 and was the 5th most populous English district .</code> | <code>Manchester ( ) is a major city and metropolitan borough in Greater Manchester , England , with a population of 545,500 as of 2017 ( 5th most populous English district ) .</code> |
| <code>Manchester had a population of less than 540,000 in 2018 and was the 4th most populous English district .</code> | <code>Manchester ( ) is a major city and metropolitan borough in Greater Manchester , England , with a population of 534,982 as of 2018 ( 4th most populous English district ) .</code> |
| <code>Traditional Chinese medicine is founded on more than 4000 years of ancient Chinese medical science and practice .</code> | <code>Traditional Chinese medicine ( TCM ; ) is an ancient system of medical diagnosis and treatment of illnesses with a holistic focus on disease prevention through diet , healthy lifestyle changes , exercise and is built on a patient centered clinically oriented foundation of more than 6,500 years of ancient Chinese medical science and practice that includes various forms of herbal medicine , acupuncture , massage ( tui na ) , exercise ( qigong ) , and dietary therapy , but recently also influenced by modern Western medicine .</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### qnli-contrastive
* Dataset: [qnli-contrastive](https://huggingface.co/datasets/nyu-mll/glue) at [bcdcba7](https://huggingface.co/datasets/nyu-mll/glue/tree/bcdcba79d07bc864c1c254ccfcedcce55bcc9a8c)
* Size: 22,500 training samples
* Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>label</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 | label |
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------|
| type | string | string | int |
| details | <ul><li>min: 6 tokens</li><li>mean: 13.79 tokens</li><li>max: 34 tokens</li></ul> | <ul><li>min: 8 tokens</li><li>mean: 34.86 tokens</li><li>max: 123 tokens</li></ul> | <ul><li>0: 100.00%</li></ul> |
* Samples:
| sentence1 | sentence2 | label |
|:---------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------|
| <code>What is HMG-CoA responsible for producing?</code> | <code>In 1971, Akira Endo, a Japanese biochemist working for the pharmaceutical company Sankyo, identified mevastatin (ML-236B), a molecule produced by the fungus Penicillium citrinum, as an inhibitor of HMG-CoA reductase, a critical enzyme used by the body to produce cholesterol.</code> | <code>0</code> |
| <code>What seperates Tajikistan and Pakistan?</code> | <code>Pakistan lies to the south, separated by the narrow Wakhan Corridor.</code> | <code>0</code> |
| <code>Where was the Agucadoura Wave Farm located?</code> | <code>Since the turn of the 21st century, there has been a trend towards the development of a renewable resource industry and reduction of both consumption and use of fossil fuel resources.</code> | <code>0</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "OnlineContrastiveLoss",
"n_layers_per_step": -1,
"last_layer_weight": 0.75,
"prior_layers_weight": 1,
"kl_div_weight": 0.9,
"kl_temperature": 0.75
}
```
#### scitail-pairs-qa
* Dataset: [scitail-pairs-qa](https://huggingface.co/datasets/allenai/scitail) at [0cc4353](https://huggingface.co/datasets/allenai/scitail/tree/0cc4353235b289165dfde1c7c5d1be983f99ce44)
* Size: 14,537 training samples
* Columns: <code>sentence2</code> and <code>sentence1</code>
* Approximate statistics based on the first 1000 samples:
| | sentence2 | sentence1 |
|:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 7 tokens</li><li>mean: 15.88 tokens</li><li>max: 41 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.0 tokens</li><li>max: 41 tokens</li></ul> |
* Samples:
| sentence2 | sentence1 |
|:------------------------------------------------------------------------|:-----------------------------------------------------------------------|
| <code>Lithium has three electrons.</code> | <code>How many electrons does lithium have?</code> |
| <code>Mammals may either be herbivores, omnivores or herbivores.</code> | <code>Mammals may either be herbivores, omnivores or what else?</code> |
| <code>The fetal period lasts approximately 30 weeks weeks.</code> | <code>Approximately how many weeks does the fetal period last?</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### scitail-pairs-pos
* Dataset: [scitail-pairs-pos](https://huggingface.co/datasets/allenai/scitail) at [0cc4353](https://huggingface.co/datasets/allenai/scitail/tree/0cc4353235b289165dfde1c7c5d1be983f99ce44)
* Size: 8,600 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 7 tokens</li><li>mean: 23.95 tokens</li><li>max: 60 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 15.47 tokens</li><li>max: 36 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:-------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------|
| <code>Ionic bonds are formed when atoms become ions by gaining or losing electrons.</code> | <code>When an atom gains or loses an electron it becames a(n) ion.</code> |
| <code>For example, all atoms of carbon contain six protons, all atoms of oxygen contain eight protons.</code> | <code>All carbon atoms have six protons.</code> |
| <code>form of energy : Forms of energy include heat, light, electrical, mechanical, nuclear, sound and chemical.</code> | <code>Heat, light, and sound are all different forms of energy.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### xsum-pairs
* Dataset: [xsum-pairs](https://huggingface.co/datasets/sentence-transformers/xsum) at [788ddaf](https://huggingface.co/datasets/sentence-transformers/xsum/tree/788ddafe04e539956d56b567bc32a036ee7b9206)
* Size: 10,000 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:-------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 33 tokens</li><li>mean: 191.96 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 25.36 tokens</li><li>max: 71 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>The boat lost contact after departing the eastern state of Sabah on Saturday.<br>The Malaysian Maritime Enforcement Agency said search and rescue efforts were being hampered by bad weather.<br>The incident coincides with the first day of China's week-long Lunar New Year celebration, which is also marked by ethnic Chinese in Malaysia.<br>The catamaran boat left Kota Kinabalu on Saturday at 09:00 local time (01:00 GMT) and was heading towards Pulau Mengalum, a popular tourist island 60km (37 miles) west of the city.<br>The Malaysian Maritime Enforcement Agency said it received a distress call from the boat but contact was lost soon after.<br>"I, like all the relatives of those on board, am hoping for progress in the search and rescue operation," the tourism minister for Sabah state, Masidi Manun, told the AFP news agency.<br>The search area covers 400 nautical square miles between Kota Kinabalu and Pulau Mengalum, according to the New Strait Times.<br>Storms are common in the area at this time of year.<br>Three crew members were on board the vessel, alongside the 31 passengers.</code> | <code>A boat carrying 31 people, including at least 28 Chinese tourists, has gone missing off the Malaysian coast, maritime authorities say.</code> |
| <code>The midfielder, 24, tested positive for Benzoylecgonine, a metabolite of cocaine, after the club's match against Hartlepool United on 22 November.<br>He admitted breaking anti-doping rules.<br>The suspension, which followed an Independent Regulatory Commission Hearing, is effective from 15 December 2016 to 14 February 2018.<br>Lacey, who joined Accrington in July 2016, made 17 appearances in 2016-17 but has not played since 10 December having been immediately suspended by the Lancashire club pending the outcome of the disciplinary action.<br>"Accrington Stanley has strong values on anyone taking any prohibited substances and will always act in the strongest possible way to protect the integrity of the football club," read a club statement.</code> | <code>League Two club Accrington Stanley have terminated the contract of Paddy Lacey after he was given a 14-month drugs ban by the Football Association.</code> |
| <code>They declined to give the reason, saying it was confidential.<br>Mr Assange, an Australian national, had hoped to create a base for Wikileaks in the Nordic country due to its laws protecting whistle-blowers.<br>The rejection comes ahead of the expected publication of some 400,000 Iraq war documents on Wikileaks.<br>The US military has already assembled a 120-member team to prepare for the publication of the documents which are thought to concern battle activity, Iraqi security forces and civilian casualties.<br>Wikileaks' release in July of thousands of documents on the war in Afghanistan prompted US military officials to warn that the whistleblower website might cause the deaths of US soldiers and Afghan civilians because some of the documents contained the names of locals who had helped coalition forces.<br>"We have decided not to grant him (Mr Assange) a residence permit," Sweden's Migration Board official Gunilla Wikstroem told the AFP news agency.<br>"He did not fulfil the requirements," she added without giving any further details.<br>Mr Assange applied for a residence permit on 18 August.<br>He is currently being investigated in Sweden over an alleged sex crime.<br>Mr Assange denies any wrongdoing and says the allegations are part of a smear campaign by opponents of his whistle-blowing website.</code> | <code>The founder of the Wikileaks website, Julian Assange, has been denied residency in Sweden, the country's migration board officials say.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "MultipleNegativesSymmetricRankingLoss",
"n_layers_per_step": -1,
"last_layer_weight": 0.75,
"prior_layers_weight": 1,
"kl_div_weight": 0.9,
"kl_temperature": 0.75
}
```
#### compression-pairs
* Dataset: [compression-pairs](https://huggingface.co/datasets/sentence-transformers/sentence-compression) at [605bc91](https://huggingface.co/datasets/sentence-transformers/sentence-compression/tree/605bc91d95631895ba25b6eda51a3cb596976c90)
* Size: 14,062 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 13 tokens</li><li>mean: 31.52 tokens</li><li>max: 166 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 10.08 tokens</li><li>max: 22 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------|
| <code>A YOUNG Irish construction worker murdered in a suspected Mozambique car-jacking was beaten to death.</code> | <code>Irish worker was beaten to death</code> |
| <code>Virgin Media has turned its first annual profit since its 2006 launch, the ISP said on Wednesday.</code> | <code>Virgin Media turns its first annual profit</code> |
| <code>Videos and reports are surfacing on the Internet that show that two men were thrown out of a Santorum rally for standing up and kissing each other.</code> | <code>Two men thrown out of Santorum rally for kissing</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "MultipleNegativesSymmetricRankingLoss",
"n_layers_per_step": -1,
"last_layer_weight": 0.75,
"prior_layers_weight": 1,
"kl_div_weight": 0.9,
"kl_temperature": 0.75
}
```
#### compression-pairs2
* Dataset: [compression-pairs2](https://huggingface.co/datasets/sentence-transformers/sentence-compression) at [605bc91](https://huggingface.co/datasets/sentence-transformers/sentence-compression/tree/605bc91d95631895ba25b6eda51a3cb596976c90)
* Size: 6,631 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 11 tokens</li><li>mean: 32.3 tokens</li><li>max: 153 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 10.17 tokens</li><li>max: 33 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------|
| <code>Iran rules out pulling out of the Nuclear Non-Proliferation Treaty but insists it will not abandon the right to peaceful nuclear technology.</code> | <code>Iran rules out pulling out of nuclear Non-Proliferation Treaty</code> |
| <code>On the heels of the Walsh Brothers' epic planking around the city, Jimmy Kimmel jumped on the planking bandwagon last night when he ``sent'' old Uncle Frank to plank around Los Angeles.</code> | <code>Jimmy Kimmel's Uncle Frank planks around Los Angeles</code> |
| <code>Israel carried out seven airstrikes in Gaza overnight, killing at least one Palestinian and injuring 33, the official Palestinian Wafa news agency reported Monday.</code> | <code>Israeli airstrikes kill at least one Palestinian,</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": 2,
"last_layer_weight": 0.25,
"prior_layers_weight": 2.5,
"kl_div_weight": 0.75,
"kl_temperature": 0.75
}
```
#### compression-pairs3
* Dataset: [compression-pairs3](https://huggingface.co/datasets/sentence-transformers/sentence-compression) at [605bc91](https://huggingface.co/datasets/sentence-transformers/sentence-compression/tree/605bc91d95631895ba25b6eda51a3cb596976c90)
* Size: 6,631 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 9 tokens</li><li>mean: 32.73 tokens</li><li>max: 302 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 10.09 tokens</li><li>max: 26 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------|
| <code>US President Barack Obama, joined by visiting German Chancellor Angela Merkel, touted a climate change bill being debated by Congress Friday as representing ``enormous progress'' but said more needed to be done.</code> | <code>Obama touts climate change bill</code> |
| <code>Pakistan fast bowler Mohammad Asif on Thursday released from a British jail after completing half of his one-year sentence for his role in a spot-fixing scandal.</code> | <code>Pakistan cricketer Asif released from jail</code> |
| <code>Mostly sunny weather is expected in the morning, partly cloudy after midday, rains and thunderstorm are expected in the western regions, shower is possible in Syunik and Artsakh in the evening.</code> | <code>Mostly sunny weather expected</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "MultipleNegativesRankingLoss",
"n_layers_per_step": 4,
"last_layer_weight": 0.1,
"prior_layers_weight": 10,
"kl_div_weight": 3,
"kl_temperature": 0.25
}
```
#### sciq_pairs
* Dataset: [sciq_pairs](https://huggingface.co/datasets/allenai/sciq) at [2c94ad3](https://huggingface.co/datasets/allenai/sciq/tree/2c94ad3e1aafab77146f384e23536f97a4849815)
* Size: 11,328 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 7 tokens</li><li>mean: 17.3 tokens</li><li>max: 58 tokens</li></ul> | <ul><li>min: 2 tokens</li><li>mean: 84.69 tokens</li><li>max: 512 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:--------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>What is the capacity to cause change?</code> | <code></code> |
| <code>The wavelength of a wave is the distance between corresponding points on what?</code> | <code>Q: The wavelength of a wave is the distance between corresponding points on adjacent waves. For example, it is the distance between two adjacent crests in the transverse waves in the diagram. Infer how wave frequency is related to wavelength.</code> |
| <code>What are modified leaves that bear sporangia?</code> | <code></code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### qasc_pairs
* Dataset: [qasc_pairs](https://huggingface.co/datasets/allenai/qasc) at [a34ba20](https://huggingface.co/datasets/allenai/qasc/tree/a34ba204eb9a33b919c10cc08f4f1c8dae5ec070)
* Size: 7,889 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 4 tokens</li><li>mean: 11.42 tokens</li><li>max: 22 tokens</li></ul> | <ul><li>min: 14 tokens</li><li>mean: 34.82 tokens</li><li>max: 67 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:---------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>Deltas are formed by deposition of what at the mouth of a river by water fanning out?</code> | <code>a delta is formed by deposition of sediment at the mouth of a river by water fanning out. Sand is an example of a clastic sediment.. a delta is formed by deposition of sand at the mouth of a river by water fanning out</code> |
| <code>What is important in preventing heat loss from the body?</code> | <code>Head hair is especially important in preventing heat loss from the body.. Hair is actually composed of a protein called keratin.. Keratin is especially important in preventing heat loss from the body.</code> |
| <code>What can happen when a body of water receives more water then it can hold?</code> | <code>when a body of water receives more water than it can hold , a flood occurs. Flooding can wipe out an entire crop.. when a body of water receives more water then it can hold, it can destroy crops</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### openbookqa_pairs
* Dataset: openbookqa_pairs
* Size: 4,505 training samples
* Columns: <code>question</code> and <code>fact</code>
* Approximate statistics based on the first 1000 samples:
| | question | fact |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 3 tokens</li><li>mean: 13.81 tokens</li><li>max: 78 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 11.49 tokens</li><li>max: 30 tokens</li></ul> |
* Samples:
| question | fact |
|:-----------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|
| <code>What is animal competition?</code> | <code>if two animals eat the same prey then those animals compete for that pey</code> |
| <code>If you wanted to make a metal bed frame, where would you start?</code> | <code>alloys are made of two or more metals</code> |
| <code>Places lacking warmth have few what</code> | <code>cold environments contain few organisms</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### msmarco_pairs
* Dataset: [msmarco_pairs](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3) at [28ff31e](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3/tree/28ff31e4c97cddd53d298497f766e653f1e666f9)
* Size: 13,337 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:---------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 4 tokens</li><li>mean: 8.49 tokens</li><li>max: 19 tokens</li></ul> | <ul><li>min: 14 tokens</li><li>mean: 77.89 tokens</li><li>max: 247 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:----------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>what is the connection between the central nervous system and naegleria fowleri?</code> | <code>N. fowleri is the causative agent of primary amoebic meningoencephalitis (PAM), a rare but nearly always fatal disease of the central nervous system. Cases of PAM resulting from infection with Naegleria fowleri have been reported in over fifteen countries in Africa, Asia, Europe, and North and South America.</code> |
| <code>cortana how many pounds in a ton</code> | <code>âºâº Definition: Ton. The short ton is a unit of mass equal to 2000 lb (exactly 907.18474 kg). In the United States it is often called simply ton without distinguishing it from the metric ton (or tonne) and the long ton rather, the other two are specifically noted.</code> |
| <code>does duloxetine cause anxiety?</code> | <code>The most common side effects of Cymbalta are nausea, dry mouth, constipation, diarrhea, fatigue, drowsiness, difficulty sleeping, loss of appetite, and dizziness. Some patients may experience withdrawal reactions such anxiety, nausea, nervousness, and insomnia.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### msmarco_pairs2
* Dataset: [msmarco_pairs2](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3) at [28ff31e](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3/tree/28ff31e4c97cddd53d298497f766e653f1e666f9)
* Size: 10,913 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:---------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 4 tokens</li><li>mean: 8.62 tokens</li><li>max: 24 tokens</li></ul> | <ul><li>min: 16 tokens</li><li>mean: 77.39 tokens</li><li>max: 229 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:---------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>how long to book driving test</code> | <code>How long do I have to wait. The good news is that waiting lists are getting shorter. We aim to have a national average waiting time for a driving test of 10 weeks.Our ability to meet this target depends on the number of applications received. Accordingly, the average waiting time in driving test centres may vary above or below this 10-week target.ow long do I have to wait. The good news is that waiting lists are getting shorter. We aim to have a national average waiting time for a driving test of 10 weeks.</code> |
| <code>how much does insurance pay a dietitian</code> | <code>1 For patients not covered by health insurance, a one-hour initial consultation with a registered dietitian, or RD, typically costs about $100 to $200 -- usually on the higher end if the dietitian comes to your home. If follow up visits are required, they typically cost $50 to $150 each, depending on length of consultation and whether the dietitian comes to your home.</code> |
| <code>is hydrogen the most flammable gas on earth</code> | <code>Properties: Common hydrogen has a molecular weight of 2,01594 g. As a gas it has a density of 0.071 g/l at 0ºC and 1 atm. Its relative density, compared with that of the air, is 0.0695. Hydrogen is the most flammable of all the known substances. Hydrogen is slightly more soluble in organic solvents than in water. Many metals absorb hydrogen.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": 2,
"last_layer_weight": 0.25,
"prior_layers_weight": 2.5,
"kl_div_weight": 0.75,
"kl_temperature": 0.75
}
```
#### nq_pairs
* Dataset: [nq_pairs](https://huggingface.co/datasets/sentence-transformers/natural-questions) at [f9e894e](https://huggingface.co/datasets/sentence-transformers/natural-questions/tree/f9e894e1081e206e577b4eaa9ee6de2b06ae6f17)
* Size: 18,187 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 10 tokens</li><li>mean: 11.89 tokens</li><li>max: 21 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 128.1 tokens</li><li>max: 512 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:-------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>when does lucas find out about peyton and julian</code> | <code>Julian Baker Julian Baker is first introduced as a director eager to make Lucas's first novel into a movie. He introduces himself as the person interested in making a film out of his book and tells him he wants to get a feel for the place where the book takes place. After a while Lucas agrees. Julian then says it would pay $300,000 grand and Lucas suddenly rethinks. Julian then asks about Peyton. After persuading Lucas to option of his movie, he ran into Peyton (leaving the restrooms at TRIC after he and Lucas sign the contract) and she realizes Julian is the one making the movie and it is soon revealed that he is her ex-boyfriend. Lucas finds out about Julian and Peyton, and punches Julian although they still carry on with the movie although with heated tension between the two men.</code> |
| <code>when did the capitals last won the stanley cup</code> | <code>Washington Capitals The Capitals were founded in 1974 as an expansion franchise, alongside the Kansas City Scouts. Since purchasing the team in 1999, Leonsis revitalized the franchise by drafting star players such as Alexander Ovechkin, Nicklas Backstrom, Mike Green and Braden Holtby. The 2009–10 Capitals won the franchise's first-ever Presidents' Trophy for being the team with the most points at the end of the regular season. They won it a second time in 2015–16, and did so for a third time the following season in 2016–17. In addition to eleven division titles and three Presidents' Trophies, the Capitals have reached the Stanley Cup Finals twice (in 1998 and 2018), winning in 2018.</code> |
| <code>where are the powers of congress listed in the constitution</code> | <code>Powers of the United States Congress Article I of the Constitution sets forth most of the powers of Congress, which include numerous explicit powers enumerated in Section 8. Constitutional amendments have granted Congress additional powers. Congress also has implied powers derived from the Necessary and Proper Clause of the Constitution.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### nq_pairs2
* Dataset: [nq_pairs2](https://huggingface.co/datasets/sentence-transformers/natural-questions) at [f9e894e](https://huggingface.co/datasets/sentence-transformers/natural-questions/tree/f9e894e1081e206e577b4eaa9ee6de2b06ae6f17)
* Size: 6,063 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 9 tokens</li><li>mean: 11.76 tokens</li><li>max: 25 tokens</li></ul> | <ul><li>min: 17 tokens</li><li>mean: 135.17 tokens</li><li>max: 512 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:----------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>have messi and ronaldo ever played against each other</code> | <code>Messi–Ronaldo rivalry At club level, Messi and Ronaldo represented rivals FC Barcelona and Real Madrid C.F., the two players facing each other at least twice every season in the world's most popular regular-season club game, El Clásico (among the most viewed annual sporting events), until Ronaldo's transfer to Italian club Juventus F.C. in 2018.[18][19] Off the field, they are the face of two rival sportswear manufacturers, Messi of Adidas and Ronaldo of Nike, which are also the kit suppliers of their national teams and the opposite for their clubs.[20][21][22] The two highest paid players in football, Ronaldo and Messi are among the world's best paid sports' stars in combined income from salaries, bonuses and off-field earnings. In 2016, Ronaldo was first on Forbes list of the best paid athletes, earning $88 million, with Messi being second at $81.4 million.[23] They have the two biggest social media followings in the world among sportspeople with a combined 211 million Facebook fans, with Ronaldo having 122 million and Messi having 89 million.[24] With a combined Facebook, Instagram and Twitter, Ronaldo has 321 million to be the most famous celebrity on social media and Messi has 181 million excluding Twitter.[25][26]</code> |
| <code>what year did otis redding song sitting on the dock of the bay</code> | <code>(Sittin' On) The Dock of the Bay "(Sittin' On) The Dock of the Bay" is a song co-written by soul singer Otis Redding and guitarist Steve Cropper. It was recorded by Redding twice in 1967, including once just days before his death in a plane crash. The song was released on Stax Records' Volt label in 1968,[2] becoming the first posthumous single to top the charts in the US.[3] It reached number 3 on the UK Singles Chart.</code> |
| <code>why does kansas city have a texas logo</code> | <code>Logos and uniforms of the Kansas City Chiefs When the Texans began playing in 1960, the team's logo consisted of the state of Texas in white with a yellow star marking the location of the city of Dallas. Originally, Hunt chose Columbia Blue and Orange for the Texans' uniforms, but Bud Adams chose the colors for his Houston Oilers franchise.[1] Hunt reverted to red and gold for the Texans' uniforms, which even after the team relocated to Kansas City, remain as the franchise's colors to this day.[1]</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": 2,
"last_layer_weight": 0.25,
"prior_layers_weight": 2.5,
"kl_div_weight": 0.75,
"kl_temperature": 0.75
}
```
#### trivia_pairs
* Dataset: [trivia_pairs](https://huggingface.co/datasets/sentence-transformers/trivia-qa) at [a7c36e3](https://huggingface.co/datasets/sentence-transformers/trivia-qa/tree/a7c36e3c8c8c01526bc094d79bf80d4c848b0ad0)
* Size: 24,250 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 8 tokens</li><li>mean: 18.23 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 45 tokens</li><li>mean: 449.24 tokens</li><li>max: 512 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:-----------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>Who played the title role in the UK television series ‘Dr Quinn, Medicine Woman’?</code> | <code>Dr. Quinn, Medicine Woman (TV Series 1993–1998) - IMDb IMDb Doctor Strange Confirmed to Appear in ‘Thor: Ragnarok’ 7 hours ago There was an error trying to load your rating for this title. Some parts of this page won't work property. Please reload or try later. X Beta I'm Watching This! Keep track of everything you watch; tell your friends. Error The trials and adventures of a female doctor in a small wild west town. Creator: Bolstered by the long-awaited arrival of the railroad, the citizens of Colorado Springs busily prepare for the biggest social event in the town's history, the wedding of Dr. Mike and Sully. The first... 8.7 Robert E and Grace by a house in town. A leader of the KKK comes to town and causes trouble amongst the citizens. 8.7 Brian's dog Pup gets bitten by a rabid raccoon. Then Matthew's fiancée gives the dog water. She also gets bitten. 8.6 "No Small Parts" IMDb Exclusive: "Westworld" Star Thandie Newton Actress Thandie Newton has been nominated for a Golden Globe Award for her performance as Maeve in the HBO's " Westworld ." What other significant parts has she played over the years? Visit IMDb's Golden Globes section for red-carpet photos, videos, and more. a list of 34 titles created 03 Apr 2012 a list of 31 titles created 03 Apr 2013 a list of 34 titles created 05 Apr 2013 a list of 27 images created 01 Jun 2014 a list of 31 titles created 6 months ago Title: Dr. Quinn, Medicine Woman (1993–1998) 6.7/10 Want to share IMDb's rating on your own site? Use the HTML below. You must be a registered user to use the IMDb rating plugin. Won 1 Golden Globe. Another 15 wins & 43 nominations. See more awards » Photos Edit Storyline The fifth daughter of a wealthy Boston physician, Michaela Quinn defies the conventions of post-Civil War society by following in her father's footsteps. After his death, 'Dr. Mike' leaves Boston and moves to the frontier town of Colorado Springs, where she finds the citizens less than thrilled by the concept of a woman doctor. While she struggles to earn their trust, Mike's life is complicated by a growing relationship with mountain man Byron Sully, and the unexpected responsibility of raising three orphaned children. Written by Marg Baskin <[email protected]> 1 January 1993 (USA) See more » Also Known As: Docteur Quinn, femme médecin See more » Filming Locations: Did You Know? Trivia Jane Seymour, Joe Lando, Chad Allen and Shawn Toovey, in their DVD commentaries, all expressed interest in a future "Dr. Quinn" project. Only Allen declined to reprise his role in the two movies that continued the series after its cancellation. See more » Goofs Mr Bray's store doors change from having windows to solid wood and back to having windows. See more » Quotes Dr. Michaela 'Mike' Quinn : You can't close your heart and give up on people. When you shut down the door, no one can get in. (United Kingdom) – See all my reviews Despite being such a simple series, It is probably one of the best for that reason. The 'EastEnders' of nowadays is becoming way too stereotypical and predictable that this masterpiece of a series has a somewhat timelessness to it. I mean, I'm a 16 year old, and I'M ADDICTED TO IT! The acting is of a high standard and there is no part of it I can deem as typical. The best thing is, that if you miss a few episodes and pick it up from a random episode, it still makes sense and you still enjoy watching it. It's not like one of those series that if you miss one episode, you don't really understand what's going on in all the episodes following it. That is why this series is so viewer-friendly. I don't know about you, but I'm considering buying the entire box-set! 27 of 28 people found this review helpful. Was this review helpful to you? Yes</code> |
| <code>Which medieval city was the centre of the Italian earthquake of April 2009?</code> | <code>The Frame: Earthquake in Italy Earthquake in Italy L'AQUILA, Italy (AP) -- A powerful earthquake in mountainous central Italy knocked down whole blocks of buildings early Monday as residents slept, killing more than 70 people in the country's deadliest quake in nearly three decades, officials said. Tens of thousands were homeless and 1,500 were injured. Ambulances screamed through the medieval city L'Aquila as firefighters with dogs worked feverishly to reach people trapped in fallen buildings, including a dormitory where half a dozen university students were believed still inside. Outside the half-collapsed building, tearful young people huddled together, wrapped in blankets, some still in their slippers after being roused from sleep by the quake. Dozens managed to escape as the walls fell around them. L'Aquila, capital of the Abruzzo region, was near the epicenter about 70 miles northeast of Rome. (29 images) updated 6:30 p.m., April 6 Follow The Frame on Twitter at sacbee_theframe Police, volunteers and rescuers work on a collapsed house on April 6, 2009 in the center of the Abruzzo capital L'Aquila, the epicenter of an earthquake measuring 5.8-magnitude on the open-ended Richter scale. At least 27 people were killed in an earthquake that struck central Italy as most people lay sleeping early on April 6, and the death toll was rising steadily after many homes collapsed in the Abruzzo region. AFP / Getty Images / Vincenzo Pinto MORE IMAGES Two men hug each other as people and volunteers stand amidst debris in the city of L'Aquila, after a strong earthquake rocked central Italy early Monday, April 6, 2009. A powerful earthquake struck central Italy early Monday, killing at least 16 people, collapsing buildings and leaving thousands of people homeless, officials and news reports said. Officials said the death toll was likely to rise as rescue crews made their way through the debris. AP / Pier Paolo Cito Rescuers carry a stretcher in the village of Onna, central Italy, Monday, April 6, 2009. A powerful earthquake in mountainous central Italy knocked down whole blocks of buildings early Monday as residents slept, killing at least 50 people and trapping many more, officials said. AP / Sandro Perozzi Rescuers at work following a strong earthquake, in the village of Onna, central Italy, Monday, April 6, 2009. A powerful earthquake in mountainous central Italy knocked down whole blocks of buildings early Monday as residents slept, killing at least 50 people and trapping many more, officials said. The earthquake's epicenter was about 70 miles (110 kilometers) northeast of Rome near the medieval city of L'Aquila. It struck at 3:32 a.m. local time (0132 GMT, EDT Sunday) in a quake-prone region that has had at least nine smaller jolts since the beginning of April. AP / Sandro Perozzi Cars are covered with debris and rubble following a strong earthquake, in the village of Onna, central Italy, Monday, April 6, 2009. A powerful earthquake in mountainous central Italy knocked down whole blocks of buildings early Monday as residents slept, killing at least 50 people and trapping many more, officials said. The earthquake's epicenter was about 70 miles (110 kilometers) northeast of Rome near the medieval city of L'Aquila. It struck at 3:32 a.m. local time (0132 GMT, EDT Sunday) in a quake-prone region that has had at least nine smaller jolts since the beginning of April. AP / Sandro Perozzi A doctors unpacks a rescue pack beside a collapsed building in the centre of L'Aquila on April 6, 2009. A powerful earthquake tore through central Italy devastating historic mountain towns and killing at least 40 people, authorities said. AFP / Getty Images / Filippo Monteforte Antonello Colangeli reacts as rescuers work to remove his son Giulio from the rubbles, in the city of L'Aquila, after a strong earthquake rocked central Italy, early Monday, April 6, 2009. A powerful earthquake struck central Italy early Monday, killing at least 20 people, collapsing buildings and leaving thousands of people homeless, officials and news reports said. AP / Pie</code> |
| <code>Colin Blunstone fronted which 1960s group?</code> | <code>Colin Blunstone - Music on Google Play Colin Blunstone About the artist Colin Edward Michael Blunstone is an English singer-songwriter and musician. In a career spanning more than 50 years, Blunstone came to prominence in the mid 1960s as the lead singer of the English rock band The Zombies, which released four singles that entered the Top 75 charts in the United States during the 1960s, including "She's Not There", "Tell Her No", "She's Coming Home", and "Time of the Season". Blunstone began his solo career in 1969, releasing three singles under a pseudonym of Neil MacArthur. Since then, he has released ten studio albums, and one live album under his real name. His solo hits include "She's Not There", "Say You Don't Mind", "I Don't Believe in Miracles", "How Could We Dare to Be Wrong", "What Becomes of the Brokenhearted", and "The Tracks of My Tears". He is also known for his participation on various albums with the Alan Parsons Project.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### quora_pairs
* Dataset: [quora_pairs](https://huggingface.co/datasets/sentence-transformers/quora-duplicates) at [451a485](https://huggingface.co/datasets/sentence-transformers/quora-duplicates/tree/451a4850bd141edb44ade1b5828c259abd762cdb)
* Size: 5,457 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 6 tokens</li><li>mean: 13.46 tokens</li><li>max: 43 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 13.33 tokens</li><li>max: 46 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------|
| <code>How do you write questions that attract more and better answers?</code> | <code>How can I write more popular questions and answers?</code> |
| <code>How can we take back Pakistan and China occupied kashmir?</code> | <code>Can India get its occupied land back from neighbors like Pakistan and China? If yes, how?</code> |
| <code>How should I prepare for SBI SO Assistant Manager System?</code> | <code>What and how to study for the post of assistant manager (system) in SBI?</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": 2,
"last_layer_weight": 0.25,
"prior_layers_weight": 2.5,
"kl_div_weight": 0.75,
"kl_temperature": 0.75
}
```
#### gooaq_pairs
* Dataset: [gooaq_pairs](https://huggingface.co/datasets/sentence-transformers/gooaq) at [b089f72](https://huggingface.co/datasets/sentence-transformers/gooaq/tree/b089f728748a068b7bc5234e5bcf5b25e3c8279c)
* Size: 18,187 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 8 tokens</li><li>mean: 11.36 tokens</li><li>max: 21 tokens</li></ul> | <ul><li>min: 17 tokens</li><li>mean: 57.44 tokens</li><li>max: 123 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:--------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>what is the health benefits of eating jackfruit?</code> | <code>For example, it's rich in vitamin C and one of the few fruits that's high in B vitamins. Jackfruit also contains folate, niacin, riboflavin, potassium, and magnesium.</code> |
| <code>fscs how is it funded?</code> | <code>The FSCS is funded by the financial services industry. Every firm authorised by the UK regulators is obliged to pay an annual levy, which goes towards our running costs and the compensation payments we make. This levy is split into six broad classes covering each type of firm: deposits.</code> |
| <code>can you get a home equity loan to pay off debt?</code> | <code>Debt consolidation A HELOC or home equity loan can be used to consolidate high-interest debts to a lower interest rate. Homeowners sometimes use home equity to pay off other personal debts such as a car loan or a credit card.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### gooaq_pairs2
* Dataset: [gooaq_pairs2](https://huggingface.co/datasets/sentence-transformers/gooaq) at [b089f72](https://huggingface.co/datasets/sentence-transformers/gooaq/tree/b089f728748a068b7bc5234e5bcf5b25e3c8279c)
* Size: 6,063 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 8 tokens</li><li>mean: 11.36 tokens</li><li>max: 21 tokens</li></ul> | <ul><li>min: 15 tokens</li><li>mean: 57.48 tokens</li><li>max: 129 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:---------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>is going to the gym a hobby?</code> | <code>A hobby is something you are passionate about and pursue seriously. ... Just like reading,playing a specific game or trekking, going to gym can also be listed as hobby. But I would suggest you to use some professional word such as “bodybuilding” “working out/exercising”.</code> |
| <code>are fruit enzymes good for you?</code> | <code>The main functions of drinkable fruit based probiotic enzyme Maintain a high alkaline blood ph. Improve the immune system. Improve the digestive system and relieve constipation. An antioxidant that protects body cells from oxidation by neutralising free radicals.</code> |
| <code>are all haze strains sativa?</code> | <code>HAZE STRAINS — THEIR HISTORY AND ORIGINS. Haze strains are among the most popular sativa-dominant, or nearly pure sativa hybrids. ... The flowering time of Hazes can also be excruciatingly long: some haze strains take up to 16 weeks for the plants to mature.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": 2,
"last_layer_weight": 0.25,
"prior_layers_weight": 2.5,
"kl_div_weight": 0.75,
"kl_temperature": 0.75
}
```
#### mrpc_pairs
* Dataset: [mrpc_pairs](https://huggingface.co/datasets/nyu-mll/glue) at [bcdcba7](https://huggingface.co/datasets/nyu-mll/glue/tree/bcdcba79d07bc864c1c254ccfcedcce55bcc9a8c)
* Size: 2,474 training samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 9 tokens</li><li>mean: 26.56 tokens</li><li>max: 51 tokens</li></ul> | <ul><li>min: 11 tokens</li><li>mean: 26.68 tokens</li><li>max: 52 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:--------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------|
| <code>Not only is this the oldest known planet , it 's also the most distant .</code> | <code>Astronomers have found the oldest and most distant planet known in the universe .</code> |
| <code>Mohcine Douali , who lives in the centre of Algiers , said : " It was a great shock .</code> | <code>" It was a great shock , " said Mohcine Douali , who lives in central Algiers .</code> |
| <code>Malaysia has launched an aggressive media campaign over its water dispute with Singapore .</code> | <code>MALAYSIA will launch a publicity campaign in local newspapers today giving its version of the water dispute with Singapore .</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "MultipleNegativesSymmetricRankingLoss",
"n_layers_per_step": -1,
"last_layer_weight": 0.75,
"prior_layers_weight": 1,
"kl_div_weight": 0.9,
"kl_temperature": 0.75
}
```
### Evaluation Datasets
#### nli-pairs
* Dataset: [nli-pairs](https://huggingface.co/datasets/sentence-transformers/all-nli) at [d482672](https://huggingface.co/datasets/sentence-transformers/all-nli/tree/d482672c8e74ce18da116f430137434ba2e52fab)
* Size: 200 evaluation samples
* Columns: <code>anchor</code> and <code>positive</code>
* Approximate statistics based on the first 1000 samples:
| | anchor | positive |
|:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 5 tokens</li><li>mean: 17.3 tokens</li><li>max: 36 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 9.57 tokens</li><li>max: 21 tokens</li></ul> |
* Samples:
| anchor | positive |
|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------|
| <code>Two women are embracing while holding to go packages.</code> | <code>Two woman are holding packages.</code> |
| <code>Two young children in blue jerseys, one with the number 9 and one with the number 2 are standing on wooden steps in a bathroom and washing their hands in a sink.</code> | <code>Two kids in numbered jerseys wash their hands.</code> |
| <code>A man selling donuts to a customer during a world exhibition event held in the city of Angeles</code> | <code>A man selling donuts to a customer.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### vitaminc-pairs
* Dataset: [vitaminc-pairs](https://huggingface.co/datasets/tals/vitaminc) at [be6febb](https://huggingface.co/datasets/tals/vitaminc/tree/be6febb761b0b2807687e61e0b5282e459df2fa0)
* Size: 166 evaluation samples
* Columns: <code>claim</code> and <code>evidence</code>
* Approximate statistics based on the first 1000 samples:
| | claim | evidence |
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 9 tokens</li><li>mean: 20.58 tokens</li><li>max: 41 tokens</li></ul> | <ul><li>min: 11 tokens</li><li>mean: 34.36 tokens</li><li>max: 79 tokens</li></ul> |
* Samples:
| claim | evidence |
|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>Dragon Con had over 5000 guests .</code> | <code>Among the more than 6000 guests and musical performers at the 2009 convention were such notables as Patrick Stewart , William Shatner , Leonard Nimoy , Terry Gilliam , Bruce Boxleitner , James Marsters , and Mary McDonnell .</code> |
| <code>COVID-19 has reached more than 185 countries .</code> | <code>As of , more than cases of COVID-19 have been reported in more than 190 countries and 200 territories , resulting in more than deaths .</code> |
| <code>In March , Italy had 3.6x times more cases of coronavirus than China .</code> | <code>As of 12 March , among nations with at least one million citizens , Italy has the world 's highest per capita rate of positive coronavirus cases at 206.1 cases per million people ( 3.6x times the rate of China ) and is the country with the second-highest number of positive cases as well as of deaths in the world , after China .</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### sts-label
* Dataset: [sts-label](https://huggingface.co/datasets/sentence-transformers/stsb) at [ab7a5ac](https://huggingface.co/datasets/sentence-transformers/stsb/tree/ab7a5ac0e35aa22088bdcf23e7fd99b220e53308)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>score</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 | score |
|:--------|:---------------------------------------------------------------------------------|:--------------------------------------------------------------------------------|:---------------------------------------------------------------|
| type | string | string | float |
| details | <ul><li>min: 6 tokens</li><li>mean: 9.64 tokens</li><li>max: 18 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 9.6 tokens</li><li>max: 16 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.49</li><li>max: 1.0</li></ul> |
* Samples:
| sentence1 | sentence2 | score |
|:--------------------------------------------------|:------------------------------------------------------|:------------------|
| <code>A man with a hard hat is dancing.</code> | <code>A man wearing a hard hat is dancing.</code> | <code>1.0</code> |
| <code>A young child is riding a horse.</code> | <code>A child is riding a horse.</code> | <code>0.95</code> |
| <code>A man is feeding a mouse to a snake.</code> | <code>The man is feeding a mouse to the snake.</code> | <code>1.0</code> |
* Loss: [<code>CoSENTLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosentloss) with these parameters:
```json
{
"scale": 20.0,
"similarity_fct": "pairwise_cos_sim"
}
```
#### qnli-contrastive
* Dataset: [qnli-contrastive](https://huggingface.co/datasets/nyu-mll/glue) at [bcdcba7](https://huggingface.co/datasets/nyu-mll/glue/tree/bcdcba79d07bc864c1c254ccfcedcce55bcc9a8c)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>label</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 | label |
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------|
| type | string | string | int |
| details | <ul><li>min: 7 tokens</li><li>mean: 14.48 tokens</li><li>max: 29 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 36.17 tokens</li><li>max: 115 tokens</li></ul> | <ul><li>0: 100.00%</li></ul> |
* Samples:
| sentence1 | sentence2 | label |
|:--------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------------|
| <code>What came into force after the new constitution was herald?</code> | <code>As of that day, the new constitution heralding the Second Republic came into force.</code> | <code>0</code> |
| <code>What is the first major city in the stream of the Rhine?</code> | <code>The most important tributaries in this area are the Ill below of Strasbourg, the Neckar in Mannheim and the Main across from Mainz.</code> | <code>0</code> |
| <code>What is the minimum required if you want to teach in Canada?</code> | <code>In most provinces a second Bachelor's Degree such as a Bachelor of Education is required to become a qualified teacher.</code> | <code>0</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "OnlineContrastiveLoss",
"n_layers_per_step": -1,
"last_layer_weight": 0.75,
"prior_layers_weight": 1,
"kl_div_weight": 0.9,
"kl_temperature": 0.75
}
```
#### scitail-pairs-qa
* Dataset: [scitail-pairs-qa](https://huggingface.co/datasets/allenai/scitail) at [0cc4353](https://huggingface.co/datasets/allenai/scitail/tree/0cc4353235b289165dfde1c7c5d1be983f99ce44)
* Size: 200 evaluation samples
* Columns: <code>sentence2</code> and <code>sentence1</code>
* Approximate statistics based on the first 1000 samples:
| | sentence2 | sentence1 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 7 tokens</li><li>mean: 16.21 tokens</li><li>max: 39 tokens</li></ul> | <ul><li>min: 8 tokens</li><li>mean: 15.21 tokens</li><li>max: 34 tokens</li></ul> |
* Samples:
| sentence2 | sentence1 |
|:--------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------|
| <code>Anions are formed by atoms gaining electrons.</code> | <code>What are formed by atoms gaining electrons?</code> |
| <code>Seed dormancy ensures that seeds germinate only when conditions for seedling survival are optimal.</code> | <code>What ensures that seeds germinate only when conditions for seedling survival are optimal?</code> |
| <code>A blizzard usually includes heavy precipitation, strong winds, and surface air temperatures below 0°c.</code> | <code>Which weather event usually includes heavy precipitation, strong winds, and surface air temperatures below 0°C?</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### scitail-pairs-pos
* Dataset: [scitail-pairs-pos](https://huggingface.co/datasets/allenai/scitail) at [0cc4353](https://huggingface.co/datasets/allenai/scitail/tree/0cc4353235b289165dfde1c7c5d1be983f99ce44)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 7 tokens</li><li>mean: 22.95 tokens</li><li>max: 61 tokens</li></ul> | <ul><li>min: 8 tokens</li><li>mean: 15.46 tokens</li><li>max: 36 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:----------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------|
| <code>An introduction to atoms and elements, compounds, atomic structure and bonding, the molecule and chemical reactions.</code> | <code>Replace another in a molecule happens to atoms during a substitution reaction.</code> |
| <code>Wavelength The distance between two consecutive points on a sinusoidal wave that are in phase;</code> | <code>Wavelength is the distance between two corresponding points of adjacent waves called.</code> |
| <code>humans normally have 23 pairs of chromosomes.</code> | <code>Humans typically have 23 pairs pairs of chromosomes.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### xsum-pairs
* Dataset: [xsum-pairs](https://huggingface.co/datasets/sentence-transformers/xsum) at [788ddaf](https://huggingface.co/datasets/sentence-transformers/xsum/tree/788ddafe04e539956d56b567bc32a036ee7b9206)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:-------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 18 tokens</li><li>mean: 194.35 tokens</li><li>max: 377 tokens</li></ul> | <ul><li>min: 13 tokens</li><li>mean: 25.33 tokens</li><li>max: 44 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------|
| <code>Should Christensen remain out, Michael Kightly is likely to keep his place, while former Ipswich striker Luke Varney will hope for a recall.<br>Ipswich midfielder Cole Skuse is set to return to action, having missed two matches after suffering concussion.<br>Forward Tom Lawrence is expected to start after a groin injury.</code> | <code>Burton will check on midfielder Lasse Vigen Christensen, who missed the draw with Aston Villa because of a tight gluteal muscle.</code> |
| <code>The UK's benchmark index closed down 203.2 points at 5673.58, and has now entered a "bear market" having fallen 20% from its record high in April.<br>Indexes across Europe also tumbled, with Germany's Dax down 2.8% and the Cac-40 in Paris dropping 3.5%.<br>Shares in Shell were down about 7% after it said that annual profits would be slightly below City expectations.<br>Oil shares were also hit by the continued fall in crude prices. Brent crude fell 4% $1.16 to $27.60, while US crude dropped more than 5% to $27.01.<br>Crude oil prices have been falling since 2014 but despite that fall, producer countries have maintained output.<br>On Tuesday, the International Energy Agency warned that oil markets could "drown in oversupply" in 2016.<br>Mining shares were also hit hard. Glencore shares fell nearly 10% while BHP Billiton fell more than 7%.<br>BHP Billiton released a production report containing what investors interpreted as gloomy comments about the outlook for commodity prices.<br>Shares in WH Smith led the FTSE 250 higher, with a 5.8% gain. The company said it expects annual profits to be "slightly ahead" of expectations, due to strong sales over the five-week Christmas period.<br>On the currency markets the pound was one fifth of a cent higher against the dollar at $1.4178, and one tenth of a euro cent higher against the euro at €1.2990.</code> | <code>The FTSE 100 slumped 3.5% as investors fretted over global growth prospects and falling oil prices.</code> |
| <code>Alexandra Kinova had four boys and a girl by caesarean section on Sunday, they say.<br>The births took place "without any complications", according to doctors at Prague's Institute for the Care of Mother and Child.<br>The mother and babies were placed in an intensive care unit but are believed to be in a good condition.<br>The Czech Republic's first quintuplets, who were conceived naturally without IVF, have a 95% chance of growing up healthy, the Associated Press quoted Zbynek Stranak, chief doctor at the neonatal section of the institute, as saying.<br>The boys' names are reportedly Deniel, Michael, Alex and Martin, while the girl is called Terezka.<br>Their mother, who is from the town of Milovice, about 20km (12 miles) north-east of the capital, Prague, already had one son.<br>She originally believed she was pregnant with twins, but in March doctors upped it to four - and then five in April.<br>The father of the quintuplets was present at the delivery despite his train being delayed, according to the newspaper Ceske Noviny.<br>"I was crying all the way since I feared I would not manage it," he said.</code> | <code>A 23-year-old woman has given birth to quintuplets in the Czech Republic, officials say, a first for the country.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "MultipleNegativesSymmetricRankingLoss",
"n_layers_per_step": -1,
"last_layer_weight": 0.75,
"prior_layers_weight": 1,
"kl_div_weight": 0.9,
"kl_temperature": 0.75
}
```
#### compression-pairs
* Dataset: [compression-pairs](https://huggingface.co/datasets/sentence-transformers/sentence-compression) at [605bc91](https://huggingface.co/datasets/sentence-transformers/sentence-compression/tree/605bc91d95631895ba25b6eda51a3cb596976c90)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:-----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 12 tokens</li><li>mean: 30.96 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 9.88 tokens</li><li>max: 21 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------|
| <code>Keira Knightley tells us about being spanked by Michael Fassbender for new movie A Dangerous Method.</code> | <code>Keira Knightley spanked by Michael Fassbender</code> |
| <code>The owners of a Segway tour operator in Washington DC sued the city last week in federal court to protest a requirement that their tours guides be licensed, arguing that the city's new regulations infringe upon their right to free speech.</code> | <code>Should tour guides be licensed?</code> |
| <code>A NEW memorial to one of the pioneers of the water cure is being proposed by Malvern Spa Association.</code> | <code>NEW memorial to pioneers of the water cure proposed</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "MultipleNegativesSymmetricRankingLoss",
"n_layers_per_step": -1,
"last_layer_weight": 0.75,
"prior_layers_weight": 1,
"kl_div_weight": 0.9,
"kl_temperature": 0.75
}
```
#### sciq_pairs
* Dataset: [sciq_pairs](https://huggingface.co/datasets/allenai/sciq) at [2c94ad3](https://huggingface.co/datasets/allenai/sciq/tree/2c94ad3e1aafab77146f384e23536f97a4849815)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 7 tokens</li><li>mean: 16.23 tokens</li><li>max: 32 tokens</li></ul> | <ul><li>min: 2 tokens</li><li>mean: 83.69 tokens</li><li>max: 512 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:---------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>An electric transformer connects two circuits with an iron core that becomes what?</code> | <code>An electric transformer connects two circuits with an iron core that becomes an electromagnet.</code> |
| <code>What is the quantity of force multiplied by the time it is applied called?</code> | <code>The quantity of force multiplied by the time it is applied is called impulse.</code> |
| <code>After the amino acid molecule has been bound to its what, protein synthesis can take place?</code> | <code>How can a molecule containing just 4 different nucleotides specify the sequence of the 20 amino acids that occur in proteins? If each nucleotide coded for 1 amino acid, then obviously the nucleic acids could code for only 4 amino acids. What if amino acids were coded for by groups of 2 nucleotides? There are 42, or 16, different combinations of 2 nucleotides (AA, AU, AC, AG, UU, and so on). Such a code is more extensive but still not adequate to code for 20 amino acids. However, if the nucleotides are arranged in groups of 3, the number of different possible combinations is 43, or 64. Here we have a code that is extensive enough to direct the synthesis of the primary structure of a protein molecule. The genetic code can therefore be described as the identification of each group of three nucleotides and its particular amino acid. The sequence of these triplet groups in the mRNA dictates the sequence of the amino acids in the protein. Each individual three-nucleotide coding unit, as we have seen, is called a codon. Protein synthesis is accomplished by orderly interactions between mRNA and the other ribonucleic acids (transfer RNA [tRNA] and ribosomal RNA [rRNA]), the ribosome, and more than 100 enzymes. The mRNA formed in the nucleus during transcription is transported across the nuclear membrane into the cytoplasm to the ribosomes—carrying with it the genetic instructions. The process in which the information encoded in the mRNA is used to direct the sequencing of amino acids and thus ultimately to synthesize a protein is referred to as translation. Before an amino acid can be incorporated into a polypeptide chain, it must be attached to its unique tRNA. This crucial process requires an enzyme known as aminoacyl-tRNA synthetase (Figure 19.12 "Binding of an Amino Acid to Its tRNA"). There is a specific aminoacyl-tRNA synthetase for each amino acid. This high degree of specificity is vital to the incorporation of the correct amino acid into a protein. After the amino acid molecule has been bound to its tRNA carrier, protein synthesis can take place. Figure 19.13 "The Elongation Steps in Protein Synthesis" depicts a schematic stepwise representation of this all-important process. Figure 19.12 Binding of an Amino Acid to Its tRNA Saylor URL: http://www. saylor. org/books.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### qasc_pairs
* Dataset: [qasc_pairs](https://huggingface.co/datasets/allenai/qasc) at [a34ba20](https://huggingface.co/datasets/allenai/qasc/tree/a34ba204eb9a33b919c10cc08f4f1c8dae5ec070)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 5 tokens</li><li>mean: 11.03 tokens</li><li>max: 21 tokens</li></ul> | <ul><li>min: 14 tokens</li><li>mean: 34.2 tokens</li><li>max: 70 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:-------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>what requires water?</code> | <code>Fertilization occurs when sperm swim to an egg inside an archegonium.. Water is needed to transport flagellated sperm to archegonium.. fertilization requires water</code> |
| <code>When does the northern hemisphere experience cooler temperatures?</code> | <code>when a hemisphere is tilted away from the sun , that hemisphere receives less direct sunlight. In winter, the northern hemisphere is tilted away from the sun.. the northern hemisphere gets less sun in the winter</code> |
| <code>what does smoking accelerate?</code> | <code>Aging is associated with the death of cells.. Smoking accelerates the aging process.. smoking accelerates the death of cells</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### openbookqa_pairs
* Dataset: openbookqa_pairs
* Size: 200 evaluation samples
* Columns: <code>question</code> and <code>fact</code>
* Approximate statistics based on the first 1000 samples:
| | question | fact |
|:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 3 tokens</li><li>mean: 13.64 tokens</li><li>max: 45 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 11.2 tokens</li><li>max: 28 tokens</li></ul> |
* Samples:
| question | fact |
|:-----------------------------------------------------------------------|:-----------------------------------------------------------------------------|
| <code>The thermal production of a stove is generically used for</code> | <code>a stove generates heat for cooking usually</code> |
| <code>What creates a valley?</code> | <code>a valley is formed by a river flowing</code> |
| <code>when it turns day and night on a planet, what cause this?</code> | <code>a planet rotating causes cycles of day and night on that planet</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### msmarco_pairs
* Dataset: [msmarco_pairs](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3) at [28ff31e](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3/tree/28ff31e4c97cddd53d298497f766e653f1e666f9)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:---------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 4 tokens</li><li>mean: 8.63 tokens</li><li>max: 20 tokens</li></ul> | <ul><li>min: 28 tokens</li><li>mean: 78.17 tokens</li><li>max: 225 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>how tall is renamon</code> | <code>(Note from anonymous publisher): According to Digimon Masters Online, Renamon is 222.3529411765 centimeters which is 73.54052802' in feet. All digimon have height information on them. If I were to get a Renamon with that is 180 in height, It also provides me the percentage of it's height compared to real height. So I would calculate to get it's real height. â Preceding unsigned comment added by 173.22.75.44 (talk) 20:41, 1 June 2014 (UTC)</code> |
| <code>what is alt level?</code> | <code>The normal level of ALT in the bloodstream is 5 to 45 U/L (units per liter). This range can slightly vary depending where you take the blood test. When a blood test shows elevated ALT levels outside the normal range, even a small amount, fatty liver and other liver disorders might be the cause. If the cause of the ALT enzyme level increase is due to a severe liver disease (like Cirrhosis), the levels would be higher than whatâs found in fatty livers patients. Please note that some labs name ALT as âAlanine Transaminaseâ, âAlanine Aminotransferaseâ or âSGPTâ.</code> |
| <code>where is the decussations of pyramids located and what functions is it involved in</code> | <code>Section of the medulla oblongata at the level of the decussation of the pyramids. The two pyramids contain the motor fibers that pass from the brain to the medulla oblongata and spinal cord. These are the corticobulbar and corticospinal fibers that make up the pyramidal tracts.ibers of the posterior column, which transmit sensory and proprioceptive information, are located behind the pyramids on the medulla oblongata. The medullary pyramids contain motor fibers that are known as the corticobulbar and corticospinal tracts. The corticospinal tracts are on the anterior surface of the pyramids.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### nq_pairs
* Dataset: [nq_pairs](https://huggingface.co/datasets/sentence-transformers/natural-questions) at [f9e894e](https://huggingface.co/datasets/sentence-transformers/natural-questions/tree/f9e894e1081e206e577b4eaa9ee6de2b06ae6f17)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 9 tokens</li><li>mean: 11.79 tokens</li><li>max: 25 tokens</li></ul> | <ul><li>min: 27 tokens</li><li>mean: 126.53 tokens</li><li>max: 396 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:-------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>which team won the world cup 2015 who did they defeat</code> | <code>2015 Cricket World Cup Final It was New Zealand's first World Cup Final.[5] They had previously lost the semi-final on six occasions between 1975 and 2011. Australia played in their record seventh final, having won four (1987, 1999, 2003 and 2007) and lost two (1975, 1996).</code> |
| <code>when did i can only imagine get released</code> | <code>I Can Only Imagine (MercyMe song) "I Can Only Imagine" was released in 2001 as the album's lead single. It gained significant airplay on Christian radio formats before crossing over to mainstream radio formats such as adult contemporary and Top 40 in late 2003 and into 2004; to aid in promotion to these markets, a double A-side physical single (combined with "Word of God Speak") was released in 2003. It charted on several formats, including the Billboard Adult Contemporary (where it peaked at No. 5) and the Hot 100 (where it peaked at No. 71). In 2002, "I Can Only Imagine" earned the Dove Awards for 'Pop/Contemporary Recorded Song of the Year' and 'Song of the Year'; Millard earned the Dove Award 'Songwriter of the Year' at the same ceremony. With 2.5 million copies sold, it is the best-selling Christian single of all time, having been certified 3x platinum by the RIAA. As of 2018, it is the only Christian song to reach that milestone.</code> |
| <code>how did the catherine wheel get its name</code> | <code>Catherine wheel (firework) The firework is named after Saint Catherine of Alexandria who, according to Christian tradition, was condemned to death by “breaking on the wheel”. When she touched the wheel it miraculously flew to pieces.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### trivia_pairs
* Dataset: [trivia_pairs](https://huggingface.co/datasets/sentence-transformers/trivia-qa) at [a7c36e3](https://huggingface.co/datasets/sentence-transformers/trivia-qa/tree/a7c36e3c8c8c01526bc094d79bf80d4c848b0ad0)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 9 tokens</li><li>mean: 17.95 tokens</li><li>max: 42 tokens</li></ul> | <ul><li>min: 25 tokens</li><li>mean: 436.37 tokens</li><li>max: 512 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>Situated in Paris, what was the original name of the ‘Place Charles de Gaulle’?</code> | <code>Place Charles-de-Gaulle was merged with this page 4.5 ★ · 3.7K public ratings</code> |
| <code>Which French phrase commonly used in English means literally 'bottom of the bag' ?</code> | <code>French Words and Expressions in English French Words and Expressions in English Learn the true meanings of French words and expressions commonly used in English Sign Up for Our Free Newsletters Thanks, You're in! What I Learned About Today You might also enjoy: Health Tip of the Day Recipe of the Day There was an error. Please try again. Please select a newsletter. Please enter a valid email address. Did you mean ? Chic sounds more chic than "stylish." cinéma vérité "cinema truth" comme il faut "as it must" The proper way, as it should be cordon bleu "blue ribbon" coup de foudre "bolt of lightning" Love at first sight coup de grâce "mercy blow" Deathblow, final blow, decisive stroke coup de main "stroke of hand" Somehow the English meaning (surprise attack) got completely separated from the French (assistance, helping hand). coup de maître "master stroke" A stroke of genius coup de théâtre "stroke of the theater" Sudden, unexpected turn of events in a play coup d'etat "state blow" Overthrow of the government. Note that the last word is capitalized and accented in French: coup d'État. coup d'œil "stroke of the eye" A glance Baked custard with carmelized crust crème caramel "caramel cream" Synonym of flan - custard lined with caramel crème de cacao "cream of cacao" Chocolate-flavored liqueur crème de la crème "cream of the cream" Synonymous with the English expression "cream of the crop" - refers to the best of the best. continue reading below our video 4 Tips for Improving Test Performance crème de menthe "cream of mint" Mint-flavored liqueur crème fraîche "fresh cream" This is a funny term. Despite its meaning, crème fraîche is in fact slightly fermented, thickened cream. crêpe de Chine "Chinese crepe" Type of silk cri de cœur "cry of heart" The correct way to say "heartfelt cry" in French is cri du cœur (literally, "cry of the heart") crime passionnel "passionate crime" Crime of passion critique "critical, judgment" Critique is an adjective and noun in French, but a noun and verb in English; it refers to a critical review of something or the act of performing such a review. cuisine "kitchen, food style" In English, cuisine refers only to a particular type of food/cooking, such as French cuisine, Southern cuisine, etc. cul-de-sac "bottom (butt) of the bag" Dead-end street debutante "beginner" In French, débutante is the feminine form of débutant - beginner (noun) or beginning (adj). In both languages, it also refers to a young girl making her formal début into society. Interestingly, this usage is not original in French; it was adopted back from English. décolletage, décolleté "low neckline, lowered neckline" The first is a noun, the second an adjective, but both refer to low necklines on women's clothing. dégustation "tasting" The French word simply refers to the act of tasting, while in English "degustation" is used for a tasting event or party, as in wine or cheese tasting. déjà vu "already seen" This is a grammatical structure in French, as in Je l'ai déjà vu=> I've already seen it, whereas in English, déjà vu refers to the phenomenon of feeling like you've already seen or done something when you're sure that you haven't. demimonde "half world" In French, it's hyphenated: demi-monde. In English, there are two meanings: 1. A marginal or disrespectful group 2. Prostitutes and/or kept women demitasse "half cup" In French, it's hyphenated: demi-tasse. Refers to a small cup of espresso or other strong coffee. démodé "out of fashion" Same meaning in both languages: outmoded, out of fashion de rigueur "of rigueur" de trop "of too much" Excessive, superfluous Dieu et mon droit "God and my right" Motto of the British monarch divorcé, divorcée "divorced man, divorced woman" In English, the feminine, divorcée, is far more common, and is often written without the accent: divorcee double entendre "double hearing" A word play or pun. For example, you're looki</code> |
| <code>Developed in 1941, the Nickel-Strunz classification scheme categorizes?</code> | <code>Sabine Grunwald Soil Classification Classification of natural phenomena like soils is typically done for the following purposes: Organize knowledge about the subject to enable investigation and communication to be both logical and comprehensive (structure/organization for scientific/technical development). Provide a framework for establishing relationships among soils and their environment that leads to advancement of theoretical and experimental aspects of soil and related science (academic focus). Establish groupings of soils for which useful and reliable interpretations can be made (utilitarian focus). For example: optimal use(s), hazard/limitation/remediation assessment, potential productivity, framework for technology transfer/information dissemination. There are two different classification approaches, i.e., technical vs. natural classification. Technical classifications are designed for specific applied purposes (e.g., soil engineering classifications, based largely on physical properties), whereas natural classifications attempt to organize the divisions of soils from a more holistic appraisal of soil attributes. The general principles of natural soil classification systems are: A natural system of classification should express general or universal relationships that exist in nature. One should be able to understand, remember, generalize, or predict from information obtained. The scheme should be based on characteristics or attributes of things classified as related to their genesis. It should place similar things together on the basis of their properties. It is technically impossible to use all of the properties of soils to classify them. Judgment based on existing knowledge must be used to determine which properties are most important. Historical Perspective One of the earliest land evaluation systems that incorporated a soil classification was established during the Vao dynasty (2357-2261 B.C.) in China. Soils were graded into nine classes, based on their productivity. It has been suggested that property taxes were based on the size of the individual land holding and soil productivity. In former times (< 1600 A.C.), soil was solely considered as a medium for plant growth. Knowledge of soil behavior and crop growth was passed from generation to generation gained by observation. For example, in the Middle Ages it was well know that manure applied to soils improved crop growth. For instance, the 'Plaggen cultivation' was practiced for a long time in Europe, which left 'Plaggen soils': The top of grassland was peeled off and used as litter in the stables. This material mixed with manure was applied to arable land to improve crop production. In 1840, the German chemist Justus von Liebig initiated a revolution in soil science and agriculture. He proved that plants assimilate mineral nutrients from the soil and proposed the use of mineral fertilizers to fortify deficient soils. Crop production was increased tremendously using mineral fertilizers. Another effect was the shift from extensive to intensive techniques in agriculture, which influenced soils. Thaer (1853) published a classification that combined texture as a primary subdivision with further subdivisions based on agricultural suitability and productivity. Several classifications based largely on geologic origin of soil material were also proposed in the 19th century (Fallou, 1862; Richtofen, 1886). From the 1660s onwards, various members of The Royal Society of London proposed schemes of soil classification that incorporated elements of a natural or scientific approach in their criteria. From this period on, the disciplines of agricultural chemistry (with a strong</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### quora_pairs
* Dataset: [quora_pairs](https://huggingface.co/datasets/sentence-transformers/quora-duplicates) at [451a485](https://huggingface.co/datasets/sentence-transformers/quora-duplicates/tree/451a4850bd141edb44ade1b5828c259abd762cdb)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 6 tokens</li><li>mean: 12.98 tokens</li><li>max: 35 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 13.45 tokens</li><li>max: 36 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:---------------------------------------------------------------|:---------------------------------------------------------------------------------------------------|
| <code>What animals prey on meerkats?</code> | <code>What animals prey of the meerkats?</code> |
| <code>How do you become the top writer on Quora?</code> | <code>If I wanted to make it a goal to be a top writer on Quora, how can I accomplish this?</code> |
| <code>Who do you think would win the 2016 USA Election?</code> | <code>Who will win the Election? Trump or Clinton?</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": 2,
"last_layer_weight": 0.25,
"prior_layers_weight": 2.5,
"kl_div_weight": 0.75,
"kl_temperature": 0.75
}
```
#### gooaq_pairs
* Dataset: [gooaq_pairs](https://huggingface.co/datasets/sentence-transformers/gooaq) at [b089f72](https://huggingface.co/datasets/sentence-transformers/gooaq/tree/b089f728748a068b7bc5234e5bcf5b25e3c8279c)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 8 tokens</li><li>mean: 11.63 tokens</li><li>max: 19 tokens</li></ul> | <ul><li>min: 13 tokens</li><li>mean: 55.31 tokens</li><li>max: 103 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:-----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>what does it mean when you keep having the same person in your dream?</code> | <code>According to a clinical psychologist, having recurring dreams about the same person shouldn't be taken too literally, whether they are your best friend or a sworn enemy. ... These dreams may not mean that you are obsessed with this individual, but may symbolise your feelings and worries.</code> |
| <code>is listerine good for a sore throat?</code> | <code>Can LISTERINE® mouthwash prevent sore throat? No. LISTERINE® mouthwash products are only intended to be used to help prevent common oral health problems like bad breath, plaque, cavities, gingivitis and tooth stains. Please consult with your doctor on how to treat, prevent or relieve the pain of a sore throat.</code> |
| <code>what is the difference between a bb clarinet and a eb clarinet?</code> | <code>The E-flat Clarinet, also called as Piccolo Clarinet, is the small brother of the Clarinet and is the highest instruments of the Clarinet family. The only difference is that it is smaller than the B-flat Clarinet. It is played like the Clarinet and is made of the same materials.</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "GISTEmbedLoss",
"n_layers_per_step": -1,
"last_layer_weight": 1.5,
"prior_layers_weight": 0.5,
"kl_div_weight": 1.25,
"kl_temperature": 1.1
}
```
#### mrpc_pairs
* Dataset: [mrpc_pairs](https://huggingface.co/datasets/nyu-mll/glue) at [bcdcba7](https://huggingface.co/datasets/nyu-mll/glue/tree/bcdcba79d07bc864c1c254ccfcedcce55bcc9a8c)
* Size: 200 evaluation samples
* Columns: <code>sentence1</code> and <code>sentence2</code>
* Approximate statistics based on the first 1000 samples:
| | sentence1 | sentence2 |
|:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
| type | string | string |
| details | <ul><li>min: 14 tokens</li><li>mean: 26.59 tokens</li><li>max: 42 tokens</li></ul> | <ul><li>min: 13 tokens</li><li>mean: 26.07 tokens</li><li>max: 41 tokens</li></ul> |
* Samples:
| sentence1 | sentence2 |
|:---------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------|
| <code>Eric Gagne pitched a perfect ninth for his 23rd save in as many opportunities .</code> | <code>Gagne struck out two in a perfect ninth inning for his 23rd save .</code> |
| <code>University of Michigan President Mary Sue Coleman said in a statement on the university 's Web site , " Our fundamental values haven 't changed .</code> | <code>" Our fundamental values haven 't changed , " Mary Sue Coleman , president of the university , said in a statement in Ann Arbor .</code> |
| <code>Mr Annan also warned the US should not use the war on terror as an excuse to suppress " long-cherished freedoms " .</code> | <code>Annan warned that the dangers of extremism after September 11 should not be used as an excuse to suppress " long-cherished " freedoms .</code> |
* Loss: [<code>AdaptiveLayerLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#adaptivelayerloss) with these parameters:
```json
{
"loss": "MultipleNegativesSymmetricRankingLoss",
"n_layers_per_step": -1,
"last_layer_weight": 0.75,
"prior_layers_weight": 1,
"kl_div_weight": 0.9,
"kl_temperature": 0.75
}
```
### Training Hyperparameters
#### Non-Default Hyperparameters
- `eval_strategy`: steps
- `per_device_train_batch_size`: 42
- `per_device_eval_batch_size`: 42
- `learning_rate`: 3.5e-05
- `weight_decay`: 0.0001
- `lr_scheduler_type`: cosine_with_restarts
- `lr_scheduler_kwargs`: {'num_cycles': 2}
- `warmup_ratio`: 0.25
- `save_safetensors`: False
- `fp16`: True
- `push_to_hub`: True
- `hub_model_id`: bobox/DeBERTa-ST-AllLayers-v3-checkpoints-tmp
- `hub_strategy`: all_checkpoints
- `batch_sampler`: no_duplicates
#### All Hyperparameters
<details><summary>Click to expand</summary>
- `overwrite_output_dir`: False
- `do_predict`: False
- `eval_strategy`: steps
- `prediction_loss_only`: True
- `per_device_train_batch_size`: 42
- `per_device_eval_batch_size`: 42
- `per_gpu_train_batch_size`: None
- `per_gpu_eval_batch_size`: None
- `gradient_accumulation_steps`: 1
- `eval_accumulation_steps`: None
- `learning_rate`: 3.5e-05
- `weight_decay`: 0.0001
- `adam_beta1`: 0.9
- `adam_beta2`: 0.999
- `adam_epsilon`: 1e-08
- `max_grad_norm`: 1.0
- `num_train_epochs`: 3
- `max_steps`: -1
- `lr_scheduler_type`: cosine_with_restarts
- `lr_scheduler_kwargs`: {'num_cycles': 2}
- `warmup_ratio`: 0.25
- `warmup_steps`: 0
- `log_level`: passive
- `log_level_replica`: warning
- `log_on_each_node`: True
- `logging_nan_inf_filter`: True
- `save_safetensors`: False
- `save_on_each_node`: False
- `save_only_model`: False
- `restore_callback_states_from_checkpoint`: False
- `no_cuda`: False
- `use_cpu`: False
- `use_mps_device`: False
- `seed`: 42
- `data_seed`: None
- `jit_mode_eval`: False
- `use_ipex`: False
- `bf16`: False
- `fp16`: True
- `fp16_opt_level`: O1
- `half_precision_backend`: auto
- `bf16_full_eval`: False
- `fp16_full_eval`: False
- `tf32`: None
- `local_rank`: 0
- `ddp_backend`: None
- `tpu_num_cores`: None
- `tpu_metrics_debug`: False
- `debug`: []
- `dataloader_drop_last`: False
- `dataloader_num_workers`: 0
- `dataloader_prefetch_factor`: None
- `past_index`: -1
- `disable_tqdm`: False
- `remove_unused_columns`: True
- `label_names`: None
- `load_best_model_at_end`: False
- `ignore_data_skip`: False
- `fsdp`: []
- `fsdp_min_num_params`: 0
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
- `fsdp_transformer_layer_cls_to_wrap`: None
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
- `deepspeed`: None
- `label_smoothing_factor`: 0.0
- `optim`: adamw_torch
- `optim_args`: None
- `adafactor`: False
- `group_by_length`: False
- `length_column_name`: length
- `ddp_find_unused_parameters`: None
- `ddp_bucket_cap_mb`: None
- `ddp_broadcast_buffers`: False
- `dataloader_pin_memory`: True
- `dataloader_persistent_workers`: False
- `skip_memory_metrics`: True
- `use_legacy_prediction_loop`: False
- `push_to_hub`: True
- `resume_from_checkpoint`: None
- `hub_model_id`: bobox/DeBERTa-ST-AllLayers-v3-checkpoints-tmp
- `hub_strategy`: all_checkpoints
- `hub_private_repo`: False
- `hub_always_push`: False
- `gradient_checkpointing`: False
- `gradient_checkpointing_kwargs`: None
- `include_inputs_for_metrics`: False
- `eval_do_concat_batches`: True
- `fp16_backend`: auto
- `push_to_hub_model_id`: None
- `push_to_hub_organization`: None
- `mp_parameters`:
- `auto_find_batch_size`: False
- `full_determinism`: False
- `torchdynamo`: None
- `ray_scope`: last
- `ddp_timeout`: 1800
- `torch_compile`: False
- `torch_compile_backend`: None
- `torch_compile_mode`: None
- `dispatch_batches`: None
- `split_batches`: None
- `include_tokens_per_second`: False
- `include_num_input_tokens_seen`: False
- `neftune_noise_alpha`: None
- `optim_target_modules`: None
- `batch_eval_metrics`: False
- `batch_sampler`: no_duplicates
- `multi_dataset_batch_sampler`: proportional
</details>
### Training Logs
| Epoch | Step | Training Loss | openbookqa pairs loss | mrpc pairs loss | scitail-pairs-pos loss | compression-pairs loss | nli-pairs loss | sts-label loss | vitaminc-pairs loss | gooaq pairs loss | quora pairs loss | trivia pairs loss | qasc pairs loss | scitail-pairs-qa loss | sciq pairs loss | msmarco pairs loss | nq pairs loss | xsum-pairs loss | qnli-contrastive loss | sts-test_spearman_cosine |
|:------:|:----:|:-------------:|:---------------------:|:---------------:|:----------------------:|:----------------------:|:--------------:|:--------------:|:-------------------:|:----------------:|:----------------:|:-----------------:|:---------------:|:---------------------:|:---------------:|:------------------:|:-------------:|:---------------:|:---------------------:|:------------------------:|
| 0.0075 | 48 | 12.3074 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0151 | 96 | 15.7221 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0226 | 144 | 10.8027 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0301 | 192 | 8.9559 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0376 | 240 | 8.8511 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0452 | 288 | 9.3478 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0527 | 336 | 8.8892 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0602 | 384 | 8.3008 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0678 | 432 | 7.3455 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0751 | 479 | - | 7.6210 | 2.1153 | 4.2950 | 2.8060 | 6.5910 | 3.3717 | 6.2479 | 7.4574 | 1.1983 | 8.1779 | 7.7198 | 5.4017 | 10.6215 | 8.3536 | 7.9954 | 3.5407 | 3.5311 | 0.3998 |
| 0.0753 | 480 | 8.0369 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0828 | 528 | 6.2732 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0903 | 576 | 7.8529 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.0979 | 624 | 5.8643 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1054 | 672 | 6.3179 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1129 | 720 | 6.1175 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1205 | 768 | 5.2392 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1280 | 816 | 5.8324 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1355 | 864 | 5.1523 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1430 | 912 | 6.0303 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1503 | 958 | - | 5.0058 | 1.0854 | 2.4681 | 1.6288 | 4.3176 | 3.7677 | 6.3070 | 3.9322 | 0.9212 | 4.9134 | 3.7815 | 1.4301 | 9.7878 | 4.5749 | 5.2812 | 2.1537 | 2.8634 | 0.6459 |
| 0.1506 | 960 | 5.7748 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1581 | 1008 | 4.8728 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1656 | 1056 | 4.7375 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1731 | 1104 | 4.6766 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1807 | 1152 | 4.3209 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1882 | 1200 | 3.7761 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.1957 | 1248 | 4.2161 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2033 | 1296 | 4.9089 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2108 | 1344 | 4.3406 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2183 | 1392 | 3.5664 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2254 | 1437 | - | 4.3604 | 0.7371 | 1.9232 | 1.1713 | 3.2246 | 4.0164 | 6.7152 | 3.0770 | 0.8747 | 3.7819 | 2.9044 | 1.0403 | 9.4438 | 3.5160 | 4.0163 | 1.6821 | 1.9995 | 0.6719 |
| 0.2258 | 1440 | 4.7194 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2334 | 1488 | 3.6345 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2409 | 1536 | 3.5947 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2484 | 1584 | 4.0526 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2560 | 1632 | 3.7962 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2635 | 1680 | 4.1927 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2710 | 1728 | 3.6351 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2785 | 1776 | 3.4256 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2861 | 1824 | 3.3175 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
| 0.2936 | 1872 | 3.4984 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
### Framework Versions
- Python: 3.10.12
- Sentence Transformers: 3.0.1
- Transformers: 4.41.2
- PyTorch: 2.3.0+cu121
- Accelerate: 0.32.1
- Datasets: 2.20.0
- Tokenizers: 0.19.1
## Citation
### BibTeX
#### Sentence Transformers
```bibtex
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
```
#### AdaptiveLayerLoss
```bibtex
@misc{li20242d,
title={2D Matryoshka Sentence Embeddings},
author={Xianming Li and Zongxi Li and Jing Li and Haoran Xie and Qing Li},
year={2024},
eprint={2402.14776},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
#### CoSENTLoss
```bibtex
@online{kexuefm-8847,
title={CoSENT: A more efficient sentence vector scheme than Sentence-BERT},
author={Su Jianlin},
year={2022},
month={Jan},
url={https://kexue.fm/archives/8847},
}
```
#### GISTEmbedLoss
```bibtex
@misc{solatorio2024gistembed,
title={GISTEmbed: Guided In-sample Selection of Training Negatives for Text Embedding Fine-tuning},
author={Aivin V. Solatorio},
year={2024},
eprint={2402.16829},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
```
#### MultipleNegativesRankingLoss
```bibtex
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
<!--
## Glossary
*Clearly define terms in order to be accessible across audiences.*
-->
<!--
## Model Card Authors
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->
<!--
## Model Card Contact
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
--> |