eval/beir-arguana_ndcg@10 = 0.37881 eval/beir-arguana_recall@100 = 0.9175 eval/beir-avg_ndcg@10 = 0.36036 eval/beir-avg_recall@10 = 0.42374800000000007 eval/beir-avg_recall@100 = 0.651238 eval/beir-avg_recall@20 = 0.49372999999999995 eval/beir-climate-fever_ndcg@10 = 0.17148 eval/beir-climate-fever_recall@10 = 0.21584 eval/beir-climate-fever_recall@100 = 0.47004 eval/beir-climate-fever_recall@20 = 0.28686 eval/beir-cqadupstack_ndcg@10 = 0.2803316666666667 eval/beir-cqadupstack_recall@100 = 0.604765 eval/beir-dbpedia-entity_ndcg@10 = 0.30952 eval/beir-dbpedia-entity_recall@10 = 0.21 eval/beir-dbpedia-entity_recall@100 = 0.47187 eval/beir-dbpedia-entity_recall@20 = 0.2695 eval/beir-fever_ndcg@10 = 0.59083 eval/beir-fever_recall@10 = 0.77263 eval/beir-fever_recall@100 = 0.90212 eval/beir-fever_recall@20 = 0.82815 eval/beir-fiqa_ndcg@10 = 0.25271 eval/beir-fiqa_recall@100 = 0.57736 eval/beir-hotpotqa_ndcg@10 = 0.51242 eval/beir-hotpotqa_recall@10 = 0.54774 eval/beir-hotpotqa_recall@100 = 0.71269 eval/beir-hotpotqa_recall@20 = 0.60459 eval/beir-msmarco_ndcg@10 = 0.21755 eval/beir-msmarco_recall@10 = 0.37253 eval/beir-msmarco_recall@100 = 0.69947 eval/beir-msmarco_recall@20 = 0.47955 eval/beir-nfcorpus_ndcg@10 = 0.29721 eval/beir-nfcorpus_recall@100 = 0.27314 eval/beir-nq_ndcg@10 = 0.28932 eval/beir-nq_recall@100 = 0.80159 eval/beir-quora_ndcg@10 = 0.81279 eval/beir-quora_recall@100 = 0.98333 eval/beir-scidocs_ndcg@10 = 0.1582 eval/beir-scidocs_recall@100 = 0.37162 eval/beir-scifact_ndcg@10 = 0.63462 eval/beir-scifact_recall@100 = 0.90322 eval/beir-trec-covid_ndcg@10 = 0.5527 eval/beir-trec-covid_recall@100 = 0.10295 eval/beir-webis-touche2020_ndcg@10 = 0.18394 eval/beir-webis-touche2020_recall@100 = 0.44207 eval/qa-curatedtrec-test-acc@100 = 0.9351585014409222 eval/qa-curatedtrec-test-acc@20 = 0.8515850144092219 eval/qa-curatedtrec-test-acc@5 = 0.7046109510086456 eval/qa-entityqs-macro-acc@100 = 0.7602876877587357 eval/qa-entityqs-macro-acc@20 = 0.6447794141198262 eval/qa-entityqs-macro-acc@5 = 0.5100471129421901 eval/qa-nq-test-acc@100 = 0.8013850415512466 eval/qa-nq-test-acc@20 = 0.6750692520775623 eval/qa-nq-test-acc@5 = 0.4692520775623269 eval/qa-squad1-test-acc@100 = 0.7788079470198676 eval/qa-squad1-test-acc@20 = 0.6403973509933775 eval/qa-squad1-test-acc@5 = 0.4489120151371807 eval/qa-trivia-test-acc@100 = 0.8371784672500663 eval/qa-trivia-test-acc@20 = 0.7591266684345443 eval/qa-trivia-test-acc@5 = 0.6233536639264563 eval/qa-webq-test-acc@100 = 0.8188976377952756 eval/qa-webq-test-acc@20 = 0.7052165354330708 eval/qa-webq-test-acc@5 = 0.5068897637795275 eval/senteval-CR = 87.01 eval/senteval-MPQA = 88.89 eval/senteval-MR = 81.09 eval/senteval-MRPC = 70.8 eval/senteval-SICKRelatedness = 0.7025121676824595 eval/senteval-SST2 = 84.06 eval/senteval-STS12 = 0.6198695741812278 eval/senteval-STS13 = 0.7458274512781141 eval/senteval-STS14 = 0.6864467990669687 eval/senteval-STS15 = 0.7986117512115241 eval/senteval-STS16 = 0.7823935956761426 eval/senteval-STSBenchmark = 0.7668425138024548 eval/senteval-SUBJ = 95.34 eval/senteval-TREC = 82.7 eval/senteval-avg_sts_7 = 0.7289291218426988 eval/senteval-avg_transfer = 84.27 train/global_step = 100000