Pringled commited on
Commit
5525871
1 Parent(s): b0a2f11

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +433 -434
README.md CHANGED
@@ -1,15 +1,21 @@
1
  ---
2
  library_name: model2vec
3
  license: mit
 
 
 
 
4
  model-index:
5
  - name: potion-base-8M
6
  results:
7
- - dataset:
8
- config: en-ext
 
9
  name: MTEB AmazonCounterfactualClassification (en-ext)
10
- revision: e8379541af4e31359cca9fbcf4b00f2671dba205
11
- split: test
12
  type: mteb/amazon_counterfactual
 
 
 
13
  metrics:
14
  - type: accuracy
15
  value: 72.15142428785607
@@ -23,14 +29,14 @@ model-index:
23
  value: 77.34906471545477
24
  - type: main_score
25
  value: 72.15142428785607
26
- task:
27
  type: Classification
28
- - dataset:
29
- config: en
30
  name: MTEB AmazonCounterfactualClassification (en)
31
- revision: e8379541af4e31359cca9fbcf4b00f2671dba205
32
- split: test
33
  type: mteb/amazon_counterfactual
 
 
 
34
  metrics:
35
  - type: accuracy
36
  value: 71.7910447761194
@@ -44,14 +50,14 @@ model-index:
44
  value: 74.32788084269461
45
  - type: main_score
46
  value: 71.7910447761194
47
- task:
48
  type: Classification
49
- - dataset:
50
- config: default
51
  name: MTEB AmazonPolarityClassification (default)
52
- revision: e2d317d38cd51312af73b3d32a06d1a08b442046
53
- split: test
54
  type: mteb/amazon_polarity
 
 
 
55
  metrics:
56
  - type: accuracy
57
  value: 72.47644999999999
@@ -65,14 +71,14 @@ model-index:
65
  value: 72.2600863044581
66
  - type: main_score
67
  value: 72.47644999999999
68
- task:
69
  type: Classification
70
- - dataset:
71
- config: en
72
  name: MTEB AmazonReviewsClassification (en)
73
- revision: 1399c76144fd37290681b995c656ef9b2e06e26d
74
- split: test
75
  type: mteb/amazon_reviews_multi
 
 
 
76
  metrics:
77
  - type: accuracy
78
  value: 36.012
@@ -82,14 +88,14 @@ model-index:
82
  value: 35.38209336470206
83
  - type: main_score
84
  value: 36.012
85
- task:
86
- type: Classification
87
- - dataset:
88
- config: default
89
  name: MTEB ArguAna (default)
90
- revision: c22ab2a51041ffd869aaddef7af8d8215647e41a
91
- split: test
92
  type: mteb/arguana
 
 
 
93
  metrics:
94
  - type: main_score
95
  value: 41.966
@@ -373,14 +379,14 @@ model-index:
373
  value: 42.888
374
  - type: recall_at_5
375
  value: 52.063
376
- task:
377
- type: Retrieval
378
- - dataset:
379
- config: default
380
  name: MTEB ArxivClusteringP2P (default)
381
- revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
382
- split: test
383
  type: mteb/arxiv-clustering-p2p
 
 
 
384
  metrics:
385
  - type: main_score
386
  value: 35.387660145946825
@@ -388,14 +394,14 @@ model-index:
388
  value: 35.387660145946825
389
  - type: v_measure_std
390
  value: 14.022525689022785
391
- task:
392
  type: Clustering
393
- - dataset:
394
- config: default
395
  name: MTEB ArxivClusteringS2S (default)
396
- revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
397
- split: test
398
  type: mteb/arxiv-clustering-s2s
 
 
 
399
  metrics:
400
  - type: main_score
401
  value: 25.26058942964131
@@ -403,14 +409,14 @@ model-index:
403
  value: 25.26058942964131
404
  - type: v_measure_std
405
  value: 14.850432186356857
406
- task:
407
- type: Clustering
408
- - dataset:
409
- config: default
410
  name: MTEB AskUbuntuDupQuestions (default)
411
- revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
412
- split: test
413
  type: mteb/askubuntudupquestions-reranking
 
 
 
414
  metrics:
415
  - type: main_score
416
  value: 54.13950871400633
@@ -430,14 +436,14 @@ model-index:
430
  value: 21.861692980537956
431
  - type: nAUC_mrr_std
432
  value: 8.256966784037171
433
- task:
434
- type: Reranking
435
- - dataset:
436
- config: default
437
  name: MTEB BIOSSES (default)
438
- revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
439
- split: test
440
  type: mteb/biosses-sts
 
 
 
441
  metrics:
442
  - type: cosine_pearson
443
  value: 79.11612010879227
@@ -457,14 +463,14 @@ model-index:
457
  value: 79.11612010879227
458
  - type: spearman
459
  value: 75.85775256673794
460
- task:
461
- type: STS
462
- - dataset:
463
- config: default
464
  name: MTEB Banking77Classification (default)
465
- revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
466
- split: test
467
  type: mteb/banking77
 
 
 
468
  metrics:
469
  - type: accuracy
470
  value: 72.63636363636363
@@ -474,14 +480,14 @@ model-index:
474
  value: 71.69751597573539
475
  - type: main_score
476
  value: 72.63636363636363
477
- task:
478
- type: Classification
479
- - dataset:
480
- config: default
481
  name: MTEB BiorxivClusteringP2P (default)
482
- revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
483
- split: test
484
  type: mteb/biorxiv-clustering-p2p
 
 
 
485
  metrics:
486
  - type: main_score
487
  value: 30.861840536151014
@@ -489,14 +495,14 @@ model-index:
489
  value: 30.861840536151014
490
  - type: v_measure_std
491
  value: 0.8096483751274005
492
- task:
493
  type: Clustering
494
- - dataset:
495
- config: default
496
  name: MTEB BiorxivClusteringS2S (default)
497
- revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
498
- split: test
499
  type: mteb/biorxiv-clustering-s2s
 
 
 
500
  metrics:
501
  - type: main_score
502
  value: 20.219544420664455
@@ -504,14 +510,14 @@ model-index:
504
  value: 20.219544420664455
505
  - type: v_measure_std
506
  value: 0.7431903039116942
507
- task:
508
- type: Clustering
509
- - dataset:
510
- config: default
511
  name: MTEB CQADupstackAndroidRetrieval (default)
512
- revision: f46a197baaae43b4f621051089b82a364682dfeb
513
- split: test
514
  type: mteb/cqadupstack-android
 
 
 
515
  metrics:
516
  - type: main_score
517
  value: 31.835
@@ -795,14 +801,14 @@ model-index:
795
  value: 29.301
796
  - type: recall_at_5
797
  value: 33.918
798
- task:
799
  type: Retrieval
800
- - dataset:
801
- config: default
802
  name: MTEB CQADupstackEnglishRetrieval (default)
803
- revision: ad9991cb51e31e31e430383c75ffb2885547b5f0
804
- split: test
805
  type: mteb/cqadupstack-english
 
 
 
806
  metrics:
807
  - type: main_score
808
  value: 25.734
@@ -1086,14 +1092,14 @@ model-index:
1086
  value: 23.54
1087
  - type: recall_at_5
1088
  value: 28.054000000000002
1089
- task:
1090
  type: Retrieval
1091
- - dataset:
1092
- config: default
1093
  name: MTEB CQADupstackGamingRetrieval (default)
1094
- revision: 4885aa143210c98657558c04aaf3dc47cfb54340
1095
- split: test
1096
  type: mteb/cqadupstack-gaming
 
 
 
1097
  metrics:
1098
  - type: main_score
1099
  value: 39.235
@@ -1377,14 +1383,14 @@ model-index:
1377
  value: 37.830999999999996
1378
  - type: recall_at_5
1379
  value: 43.811
1380
- task:
1381
  type: Retrieval
1382
- - dataset:
1383
- config: default
1384
  name: MTEB CQADupstackGisRetrieval (default)
1385
- revision: 5003b3064772da1887988e05400cf3806fe491f2
1386
- split: test
1387
  type: mteb/cqadupstack-gis
 
 
 
1388
  metrics:
1389
  - type: main_score
1390
  value: 19.197
@@ -1668,14 +1674,14 @@ model-index:
1668
  value: 19.417
1669
  - type: recall_at_5
1670
  value: 23.577
1671
- task:
1672
  type: Retrieval
1673
- - dataset:
1674
- config: default
1675
  name: MTEB CQADupstackMathematicaRetrieval (default)
1676
- revision: 90fceea13679c63fe563ded68f3b6f06e50061de
1677
- split: test
1678
  type: mteb/cqadupstack-mathematica
 
 
 
1679
  metrics:
1680
  - type: main_score
1681
  value: 12.145
@@ -1959,14 +1965,14 @@ model-index:
1959
  value: 11.178
1960
  - type: recall_at_5
1961
  value: 13.877
1962
- task:
1963
  type: Retrieval
1964
- - dataset:
1965
- config: default
1966
  name: MTEB CQADupstackPhysicsRetrieval (default)
1967
- revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4
1968
- split: test
1969
  type: mteb/cqadupstack-physics
 
 
 
1970
  metrics:
1971
  - type: main_score
1972
  value: 26.667999999999996
@@ -2250,14 +2256,14 @@ model-index:
2250
  value: 24.524
2251
  - type: recall_at_5
2252
  value: 28.787000000000003
2253
- task:
2254
  type: Retrieval
2255
- - dataset:
2256
- config: default
2257
  name: MTEB CQADupstackProgrammersRetrieval (default)
2258
- revision: 6184bc1440d2dbc7612be22b50686b8826d22b32
2259
- split: test
2260
  type: mteb/cqadupstack-programmers
 
 
 
2261
  metrics:
2262
  - type: main_score
2263
  value: 21.66
@@ -2541,27 +2547,27 @@ model-index:
2541
  value: 19.79
2542
  - type: recall_at_5
2543
  value: 23.376
2544
- task:
2545
  type: Retrieval
2546
- - dataset:
2547
- config: default
2548
  name: MTEB CQADupstackRetrieval (default)
2549
- revision: CQADupstackRetrieval_is_a_combined_dataset
2550
- split: test
2551
  type: CQADupstackRetrieval_is_a_combined_dataset
 
 
 
2552
  metrics:
2553
  - type: main_score
2554
  value: 22.302333333333333
2555
  - type: ndcg_at_10
2556
  value: 22.302333333333333
2557
- task:
2558
  type: Retrieval
2559
- - dataset:
2560
- config: default
2561
  name: MTEB CQADupstackStatsRetrieval (default)
2562
- revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a
2563
- split: test
2564
  type: mteb/cqadupstack-stats
 
 
 
2565
  metrics:
2566
  - type: main_score
2567
  value: 17.253
@@ -2845,14 +2851,14 @@ model-index:
2845
  value: 17.134
2846
  - type: recall_at_5
2847
  value: 19.72
2848
- task:
2849
  type: Retrieval
2850
- - dataset:
2851
- config: default
2852
  name: MTEB CQADupstackTexRetrieval (default)
2853
- revision: 46989137a86843e03a6195de44b09deda022eec7
2854
- split: test
2855
  type: mteb/cqadupstack-tex
 
 
 
2856
  metrics:
2857
  - type: main_score
2858
  value: 13.308
@@ -3136,14 +3142,14 @@ model-index:
3136
  value: 12.199
3137
  - type: recall_at_5
3138
  value: 14.87
3139
- task:
3140
  type: Retrieval
3141
- - dataset:
3142
- config: default
3143
  name: MTEB CQADupstackUnixRetrieval (default)
3144
- revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53
3145
- split: test
3146
  type: mteb/cqadupstack-unix
 
 
 
3147
  metrics:
3148
  - type: main_score
3149
  value: 20.061999999999998
@@ -3427,14 +3433,14 @@ model-index:
3427
  value: 18.859
3428
  - type: recall_at_5
3429
  value: 22.435
3430
- task:
3431
  type: Retrieval
3432
- - dataset:
3433
- config: default
3434
  name: MTEB CQADupstackWebmastersRetrieval (default)
3435
- revision: 160c094312a0e1facb97e55eeddb698c0abe3571
3436
- split: test
3437
  type: mteb/cqadupstack-webmasters
 
 
 
3438
  metrics:
3439
  - type: main_score
3440
  value: 23.735999999999997
@@ -3718,14 +3724,14 @@ model-index:
3718
  value: 21.397
3719
  - type: recall_at_5
3720
  value: 24.863
3721
- task:
3722
  type: Retrieval
3723
- - dataset:
3724
- config: default
3725
  name: MTEB CQADupstackWordpressRetrieval (default)
3726
- revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4
3727
- split: test
3728
  type: mteb/cqadupstack-wordpress
 
 
 
3729
  metrics:
3730
  - type: main_score
3731
  value: 16.794999999999998
@@ -4009,14 +4015,14 @@ model-index:
4009
  value: 15.440999999999999
4010
  - type: recall_at_5
4011
  value: 18.829
4012
- task:
4013
  type: Retrieval
4014
- - dataset:
4015
- config: default
4016
  name: MTEB ClimateFEVER (default)
4017
- revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380
4018
- split: test
4019
  type: mteb/climate-fever
 
 
 
4020
  metrics:
4021
  - type: main_score
4022
  value: 17.288999999999998
@@ -4300,14 +4306,14 @@ model-index:
4300
  value: 12.232
4301
  - type: recall_at_5
4302
  value: 15.540999999999999
4303
- task:
4304
  type: Retrieval
4305
- - dataset:
4306
- config: default
4307
  name: MTEB DBPedia (default)
4308
- revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659
4309
- split: test
4310
  type: mteb/dbpedia
 
 
 
4311
  metrics:
4312
  - type: main_score
4313
  value: 24.235
@@ -4591,14 +4597,14 @@ model-index:
4591
  value: 8.698
4592
  - type: recall_at_5
4593
  value: 11.052
4594
- task:
4595
- type: Retrieval
4596
- - dataset:
4597
- config: default
4598
  name: MTEB EmotionClassification (default)
4599
- revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
4600
- split: test
4601
  type: mteb/emotion
 
 
 
4602
  metrics:
4603
  - type: accuracy
4604
  value: 45.09
@@ -4608,14 +4614,14 @@ model-index:
4608
  value: 47.129694558751545
4609
  - type: main_score
4610
  value: 45.09
4611
- task:
4612
- type: Classification
4613
- - dataset:
4614
- config: default
4615
  name: MTEB FEVER (default)
4616
- revision: bea83ef9e8fb933d90a2f1d5515737465d613e12
4617
- split: test
4618
  type: mteb/fever
 
 
 
4619
  metrics:
4620
  - type: main_score
4621
  value: 30.267
@@ -4899,14 +4905,14 @@ model-index:
4899
  value: 29.42
4900
  - type: recall_at_5
4901
  value: 36.112
4902
- task:
4903
  type: Retrieval
4904
- - dataset:
4905
- config: default
4906
  name: MTEB FiQA2018 (default)
4907
- revision: 27a168819829fe9bcd655c2df245fb19452e8e06
4908
- split: test
4909
  type: mteb/fiqa
 
 
 
4910
  metrics:
4911
  - type: main_score
4912
  value: 16.619
@@ -5190,14 +5196,14 @@ model-index:
5190
  value: 12.485
5191
  - type: recall_at_5
5192
  value: 15.862000000000002
5193
- task:
5194
  type: Retrieval
5195
- - dataset:
5196
- config: default
5197
  name: MTEB HotpotQA (default)
5198
- revision: ab518f4d6fcca38d87c25209f94beba119d02014
5199
- split: test
5200
  type: mteb/hotpotqa
 
 
 
5201
  metrics:
5202
  - type: main_score
5203
  value: 36.217
@@ -5481,14 +5487,14 @@ model-index:
5481
  value: 30.311
5482
  - type: recall_at_5
5483
  value: 34.092
5484
- task:
5485
- type: Retrieval
5486
- - dataset:
5487
- config: default
5488
  name: MTEB ImdbClassification (default)
5489
- revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
5490
- split: test
5491
  type: mteb/imdb
 
 
 
5492
  metrics:
5493
  - type: accuracy
5494
  value: 70.78
@@ -5502,14 +5508,14 @@ model-index:
5502
  value: 70.56498155979408
5503
  - type: main_score
5504
  value: 70.78
5505
- task:
5506
- type: Classification
5507
- - dataset:
5508
- config: default
5509
  name: MTEB MSMARCO (default)
5510
- revision: c5a29a104738b98a9e76336939199e264163d4a0
5511
- split: test
5512
  type: mteb/msmarco
 
 
 
5513
  metrics:
5514
  - type: main_score
5515
  value: 34.981
@@ -5793,14 +5799,14 @@ model-index:
5793
  value: 2.8240000000000003
5794
  - type: recall_at_5
5795
  value: 4.476999999999999
5796
- task:
5797
- type: Retrieval
5798
- - dataset:
5799
- config: en
5800
  name: MTEB MTOPDomainClassification (en)
5801
- revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
5802
- split: test
5803
  type: mteb/mtop_domain
 
 
 
5804
  metrics:
5805
  - type: accuracy
5806
  value: 88.95576835385319
@@ -5810,14 +5816,14 @@ model-index:
5810
  value: 89.00721562093213
5811
  - type: main_score
5812
  value: 88.95576835385319
5813
- task:
5814
  type: Classification
5815
- - dataset:
5816
- config: en
5817
  name: MTEB MTOPIntentClassification (en)
5818
- revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
5819
- split: test
5820
  type: mteb/mtop_intent
 
 
 
5821
  metrics:
5822
  - type: accuracy
5823
  value: 56.99726402188783
@@ -5827,14 +5833,14 @@ model-index:
5827
  value: 59.96788951671549
5828
  - type: main_score
5829
  value: 56.99726402188783
5830
- task:
5831
  type: Classification
5832
- - dataset:
5833
- config: en
5834
  name: MTEB MassiveIntentClassification (en)
5835
- revision: 4672e20407010da34463acc759c162ca9734bca6
5836
- split: test
5837
  type: mteb/amazon_massive_intent
 
 
 
5838
  metrics:
5839
  - type: accuracy
5840
  value: 63.79287155346336
@@ -5844,14 +5850,14 @@ model-index:
5844
  value: 62.567311481126055
5845
  - type: main_score
5846
  value: 63.79287155346336
5847
- task:
5848
  type: Classification
5849
- - dataset:
5850
- config: en
5851
  name: MTEB MassiveScenarioClassification (en)
5852
- revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8
5853
- split: test
5854
  type: mteb/amazon_massive_scenario
 
 
 
5855
  metrics:
5856
  - type: accuracy
5857
  value: 70.30934767989241
@@ -5861,14 +5867,14 @@ model-index:
5861
  value: 70.1128179307388
5862
  - type: main_score
5863
  value: 70.30934767989241
5864
- task:
5865
- type: Classification
5866
- - dataset:
5867
- config: default
5868
  name: MTEB MedrxivClusteringP2P (default)
5869
- revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
5870
- split: test
5871
  type: mteb/medrxiv-clustering-p2p
 
 
 
5872
  metrics:
5873
  - type: main_score
5874
  value: 27.61734940907637
@@ -5876,14 +5882,14 @@ model-index:
5876
  value: 27.61734940907637
5877
  - type: v_measure_std
5878
  value: 1.2248100208316097
5879
- task:
5880
  type: Clustering
5881
- - dataset:
5882
- config: default
5883
  name: MTEB MedrxivClusteringS2S (default)
5884
- revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
5885
- split: test
5886
  type: mteb/medrxiv-clustering-s2s
 
 
 
5887
  metrics:
5888
  - type: main_score
5889
  value: 23.802943866708308
@@ -5891,14 +5897,14 @@ model-index:
5891
  value: 23.802943866708308
5892
  - type: v_measure_std
5893
  value: 1.4975518910969763
5894
- task:
5895
- type: Clustering
5896
- - dataset:
5897
- config: default
5898
  name: MTEB MindSmallReranking (default)
5899
- revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
5900
- split: test
5901
  type: mteb/mind_small
 
 
 
5902
  metrics:
5903
  - type: main_score
5904
  value: 29.431722284942175
@@ -5918,14 +5924,14 @@ model-index:
5918
  value: -17.42779158552557
5919
  - type: nAUC_mrr_std
5920
  value: -5.997215692334967
5921
- task:
5922
- type: Reranking
5923
- - dataset:
5924
- config: default
5925
  name: MTEB NFCorpus (default)
5926
- revision: ec0fa4fe99da2ff19ca1214b7966684033a58814
5927
- split: test
5928
  type: mteb/nfcorpus
 
 
 
5929
  metrics:
5930
  - type: main_score
5931
  value: 24.267
@@ -6209,14 +6215,14 @@ model-index:
6209
  value: 7.236
6210
  - type: recall_at_5
6211
  value: 9.003
6212
- task:
6213
  type: Retrieval
6214
- - dataset:
6215
- config: default
6216
  name: MTEB NQ (default)
6217
- revision: b774495ed302d8c44a3a7ea25c90dbce03968f31
6218
- split: test
6219
  type: mteb/nq
 
 
 
6220
  metrics:
6221
  - type: main_score
6222
  value: 19.543
@@ -6500,14 +6506,14 @@ model-index:
6500
  value: 17.669
6501
  - type: recall_at_5
6502
  value: 23.488999999999997
6503
- task:
6504
  type: Retrieval
6505
- - dataset:
6506
- config: default
6507
  name: MTEB QuoraRetrieval (default)
6508
- revision: e4e08e0b7dbe3c8700f0daef558ff32256715259
6509
- split: test
6510
  type: mteb/quora
 
 
 
6511
  metrics:
6512
  - type: main_score
6513
  value: 80.598
@@ -6791,14 +6797,14 @@ model-index:
6791
  value: 79.074
6792
  - type: recall_at_5
6793
  value: 84.143
6794
- task:
6795
- type: Retrieval
6796
- - dataset:
6797
- config: default
6798
  name: MTEB RedditClustering (default)
6799
- revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
6800
- split: test
6801
  type: mteb/reddit-clustering
 
 
 
6802
  metrics:
6803
  - type: main_score
6804
  value: 39.5465127563479
@@ -6806,14 +6812,14 @@ model-index:
6806
  value: 39.5465127563479
6807
  - type: v_measure_std
6808
  value: 5.038703300031419
6809
- task:
6810
  type: Clustering
6811
- - dataset:
6812
- config: default
6813
  name: MTEB RedditClusteringP2P (default)
6814
- revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
6815
- split: test
6816
  type: mteb/reddit-clustering-p2p
 
 
 
6817
  metrics:
6818
  - type: main_score
6819
  value: 47.07911795189491
@@ -6821,14 +6827,14 @@ model-index:
6821
  value: 47.07911795189491
6822
  - type: v_measure_std
6823
  value: 11.546436135362846
6824
- task:
6825
- type: Clustering
6826
- - dataset:
6827
- config: default
6828
  name: MTEB SCIDOCS (default)
6829
- revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88
6830
- split: test
6831
  type: mteb/scidocs
 
 
 
6832
  metrics:
6833
  - type: main_score
6834
  value: 12.386999999999999
@@ -7112,14 +7118,14 @@ model-index:
7112
  value: 6.553000000000001
7113
  - type: recall_at_5
7114
  value: 9.013
7115
- task:
7116
- type: Retrieval
7117
- - dataset:
7118
- config: default
7119
  name: MTEB SICK-R (default)
7120
- revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
7121
- split: test
7122
  type: mteb/sickr-sts
 
 
 
7123
  metrics:
7124
  - type: cosine_pearson
7125
  value: 75.67336823619708
@@ -7139,14 +7145,14 @@ model-index:
7139
  value: 75.67336823619708
7140
  - type: spearman
7141
  value: 64.6753400763881
7142
- task:
7143
  type: STS
7144
- - dataset:
7145
- config: default
7146
  name: MTEB STS12 (default)
7147
- revision: a0d554a64d88156834ff5ae9920b964011b16384
7148
- split: test
7149
  type: mteb/sts12-sts
 
 
 
7150
  metrics:
7151
  - type: cosine_pearson
7152
  value: 72.06681927996405
@@ -7166,14 +7172,14 @@ model-index:
7166
  value: 72.06681927996405
7167
  - type: spearman
7168
  value: 62.248985055530525
7169
- task:
7170
  type: STS
7171
- - dataset:
7172
- config: default
7173
  name: MTEB STS13 (default)
7174
- revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
7175
- split: test
7176
  type: mteb/sts13-sts
 
 
 
7177
  metrics:
7178
  - type: cosine_pearson
7179
  value: 76.53983680018591
@@ -7193,14 +7199,14 @@ model-index:
7193
  value: 76.53983680018591
7194
  - type: spearman
7195
  value: 77.27600787572996
7196
- task:
7197
  type: STS
7198
- - dataset:
7199
- config: default
7200
  name: MTEB STS14 (default)
7201
- revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
7202
- split: test
7203
  type: mteb/sts14-sts
 
 
 
7204
  metrics:
7205
  - type: cosine_pearson
7206
  value: 76.20854411766629
@@ -7220,14 +7226,14 @@ model-index:
7220
  value: 76.20854411766629
7221
  - type: spearman
7222
  value: 71.914099628002
7223
- task:
7224
  type: STS
7225
- - dataset:
7226
- config: default
7227
  name: MTEB STS15 (default)
7228
- revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
7229
- split: test
7230
  type: mteb/sts15-sts
 
 
 
7231
  metrics:
7232
  - type: cosine_pearson
7233
  value: 79.24273419832653
@@ -7247,14 +7253,14 @@ model-index:
7247
  value: 79.24273419832653
7248
  - type: spearman
7249
  value: 79.75345871163103
7250
- task:
7251
  type: STS
7252
- - dataset:
7253
- config: default
7254
  name: MTEB STS16 (default)
7255
- revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
7256
- split: test
7257
  type: mteb/sts16-sts
 
 
 
7258
  metrics:
7259
  - type: cosine_pearson
7260
  value: 75.64452330127995
@@ -7274,14 +7280,14 @@ model-index:
7274
  value: 75.64452330127995
7275
  - type: spearman
7276
  value: 76.26343823222666
7277
- task:
7278
  type: STS
7279
- - dataset:
7280
- config: es-en
7281
  name: MTEB STS17 (es-en)
7282
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7283
- split: test
7284
  type: mteb/sts17-crosslingual-sts
 
 
 
7285
  metrics:
7286
  - type: cosine_pearson
7287
  value: 17.52217310066287
@@ -7301,14 +7307,14 @@ model-index:
7301
  value: 17.52217310066287
7302
  - type: spearman
7303
  value: 14.729958484232528
7304
- task:
7305
  type: STS
7306
- - dataset:
7307
- config: en-en
7308
  name: MTEB STS17 (en-en)
7309
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7310
- split: test
7311
  type: mteb/sts17-crosslingual-sts
 
 
 
7312
  metrics:
7313
  - type: cosine_pearson
7314
  value: 84.67406984717113
@@ -7328,14 +7334,14 @@ model-index:
7328
  value: 84.67406984717113
7329
  - type: spearman
7330
  value: 85.96709815630739
7331
- task:
7332
  type: STS
7333
- - dataset:
7334
- config: fr-en
7335
  name: MTEB STS17 (fr-en)
7336
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7337
- split: test
7338
  type: mteb/sts17-crosslingual-sts
 
 
 
7339
  metrics:
7340
  - type: cosine_pearson
7341
  value: 36.02331692863771
@@ -7355,14 +7361,14 @@ model-index:
7355
  value: 36.02331692863771
7356
  - type: spearman
7357
  value: 34.28540470062557
7358
- task:
7359
  type: STS
7360
- - dataset:
7361
- config: en-tr
7362
  name: MTEB STS17 (en-tr)
7363
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7364
- split: test
7365
  type: mteb/sts17-crosslingual-sts
 
 
 
7366
  metrics:
7367
  - type: cosine_pearson
7368
  value: 13.925983981770388
@@ -7382,14 +7388,14 @@ model-index:
7382
  value: 13.925983981770388
7383
  - type: spearman
7384
  value: 11.193291331109325
7385
- task:
7386
  type: STS
7387
- - dataset:
7388
- config: en-de
7389
  name: MTEB STS17 (en-de)
7390
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7391
- split: test
7392
  type: mteb/sts17-crosslingual-sts
 
 
 
7393
  metrics:
7394
  - type: cosine_pearson
7395
  value: 26.77839285232968
@@ -7409,14 +7415,14 @@ model-index:
7409
  value: 26.77839285232968
7410
  - type: spearman
7411
  value: 23.010015986939717
7412
- task:
7413
  type: STS
7414
- - dataset:
7415
- config: it-en
7416
  name: MTEB STS17 (it-en)
7417
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7418
- split: test
7419
  type: mteb/sts17-crosslingual-sts
 
 
 
7420
  metrics:
7421
  - type: cosine_pearson
7422
  value: 25.330935194314364
@@ -7436,14 +7442,14 @@ model-index:
7436
  value: 25.330935194314364
7437
  - type: spearman
7438
  value: 23.143555348782797
7439
- task:
7440
  type: STS
7441
- - dataset:
7442
- config: en-ar
7443
  name: MTEB STS17 (en-ar)
7444
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7445
- split: test
7446
  type: mteb/sts17-crosslingual-sts
 
 
 
7447
  metrics:
7448
  - type: cosine_pearson
7449
  value: 6.61651078645899
@@ -7463,14 +7469,14 @@ model-index:
7463
  value: 6.61651078645899
7464
  - type: spearman
7465
  value: 5.415104433010482
7466
- task:
7467
  type: STS
7468
- - dataset:
7469
- config: nl-en
7470
  name: MTEB STS17 (nl-en)
7471
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7472
- split: test
7473
  type: mteb/sts17-crosslingual-sts
 
 
 
7474
  metrics:
7475
  - type: cosine_pearson
7476
  value: 32.718045784523184
@@ -7490,14 +7496,14 @@ model-index:
7490
  value: 32.718045784523184
7491
  - type: spearman
7492
  value: 27.52844368619317
7493
- task:
7494
  type: STS
7495
- - dataset:
7496
- config: es-en
7497
  name: MTEB STS22 (es-en)
7498
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7499
- split: test
7500
  type: mteb/sts22-crosslingual-sts
 
 
 
7501
  metrics:
7502
  - type: cosine_pearson
7503
  value: 9.98410299881163
@@ -7517,14 +7523,14 @@ model-index:
7517
  value: 9.98410299881163
7518
  - type: spearman
7519
  value: 10.98684405086525
7520
- task:
7521
  type: STS
7522
- - dataset:
7523
- config: en
7524
  name: MTEB STS22 (en)
7525
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7526
- split: test
7527
  type: mteb/sts22-crosslingual-sts
 
 
 
7528
  metrics:
7529
  - type: cosine_pearson
7530
  value: 59.3180680265132
@@ -7544,14 +7550,14 @@ model-index:
7544
  value: 59.3180680265132
7545
  - type: spearman
7546
  value: 63.07956002739231
7547
- task:
7548
  type: STS
7549
- - dataset:
7550
- config: de-en
7551
  name: MTEB STS22 (de-en)
7552
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7553
- split: test
7554
  type: mteb/sts22-crosslingual-sts
 
 
 
7555
  metrics:
7556
  - type: cosine_pearson
7557
  value: 29.061215770374826
@@ -7571,14 +7577,14 @@ model-index:
7571
  value: 29.061215770374826
7572
  - type: spearman
7573
  value: 36.21441725938738
7574
- task:
7575
  type: STS
7576
- - dataset:
7577
- config: zh-en
7578
  name: MTEB STS22 (zh-en)
7579
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7580
- split: test
7581
  type: mteb/sts22-crosslingual-sts
 
 
 
7582
  metrics:
7583
  - type: cosine_pearson
7584
  value: 11.266385865086239
@@ -7598,14 +7604,14 @@ model-index:
7598
  value: 11.266385865086239
7599
  - type: spearman
7600
  value: 17.291293843893733
7601
- task:
7602
  type: STS
7603
- - dataset:
7604
- config: pl-en
7605
  name: MTEB STS22 (pl-en)
7606
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7607
- split: test
7608
  type: mteb/sts22-crosslingual-sts
 
 
 
7609
  metrics:
7610
  - type: cosine_pearson
7611
  value: 9.647587208410648
@@ -7625,14 +7631,14 @@ model-index:
7625
  value: 9.647587208410648
7626
  - type: spearman
7627
  value: 21.33739699413266
7628
- task:
7629
  type: STS
7630
- - dataset:
7631
- config: default
7632
  name: MTEB STSBenchmark (default)
7633
- revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
7634
- split: test
7635
  type: mteb/stsbenchmark-sts
 
 
 
7636
  metrics:
7637
  - type: cosine_pearson
7638
  value: 77.2598255013409
@@ -7652,14 +7658,14 @@ model-index:
7652
  value: 77.2598255013409
7653
  - type: spearman
7654
  value: 75.40519061413276
7655
- task:
7656
- type: STS
7657
- - dataset:
7658
- config: default
7659
  name: MTEB SciDocsRR (default)
7660
- revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
7661
- split: test
7662
  type: mteb/scidocs-reranking
 
 
 
7663
  metrics:
7664
  - type: main_score
7665
  value: 72.10127087089839
@@ -7679,14 +7685,14 @@ model-index:
7679
  value: 74.90946175462605
7680
  - type: nAUC_mrr_std
7681
  value: 71.81332269641806
7682
- task:
7683
- type: Reranking
7684
- - dataset:
7685
- config: default
7686
  name: MTEB SciFact (default)
7687
- revision: 0228b52cf27578f30900b9e5271d331663a030d7
7688
- split: test
7689
  type: mteb/scifact
 
 
 
7690
  metrics:
7691
  - type: main_score
7692
  value: 50.63999999999999
@@ -7970,14 +7976,14 @@ model-index:
7970
  value: 50.306
7971
  - type: recall_at_5
7972
  value: 57.443999999999996
7973
- task:
7974
- type: Retrieval
7975
- - dataset:
7976
- config: default
7977
  name: MTEB SprintDuplicateQuestions (default)
7978
- revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
7979
- split: test
7980
  type: mteb/sprintduplicatequestions-pairclassification
 
 
 
7981
  metrics:
7982
  - type: cosine_accuracy
7983
  value: 99.71386138613862
@@ -8061,14 +8067,14 @@ model-index:
8061
  value: 85.7727737973388
8062
  - type: similarity_recall
8063
  value: 83.8
8064
- task:
8065
- type: PairClassification
8066
- - dataset:
8067
- config: default
8068
  name: MTEB StackExchangeClustering (default)
8069
- revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
8070
- split: test
8071
  type: mteb/stackexchange-clustering
 
 
 
8072
  metrics:
8073
  - type: main_score
8074
  value: 48.18939518021159
@@ -8076,14 +8082,14 @@ model-index:
8076
  value: 48.18939518021159
8077
  - type: v_measure_std
8078
  value: 4.6189444340187995
8079
- task:
8080
  type: Clustering
8081
- - dataset:
8082
- config: default
8083
  name: MTEB StackExchangeClusteringP2P (default)
8084
- revision: 815ca46b2622cec33ccafc3735d572c266efdb44
8085
- split: test
8086
  type: mteb/stackexchange-clustering-p2p
 
 
 
8087
  metrics:
8088
  - type: main_score
8089
  value: 30.743938802421265
@@ -8091,14 +8097,14 @@ model-index:
8091
  value: 30.743938802421265
8092
  - type: v_measure_std
8093
  value: 1.4645401677053824
8094
- task:
8095
- type: Clustering
8096
- - dataset:
8097
- config: default
8098
  name: MTEB StackOverflowDupQuestions (default)
8099
- revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
8100
- split: test
8101
  type: mteb/stackoverflowdupquestions-reranking
 
 
 
8102
  metrics:
8103
  - type: main_score
8104
  value: 43.254152892780986
@@ -8118,14 +8124,14 @@ model-index:
8118
  value: 14.093758435205075
8119
  - type: nAUC_mrr_std
8120
  value: 4.198791420159734
8121
- task:
8122
- type: Reranking
8123
- - dataset:
8124
- config: default
8125
  name: MTEB SummEval (default)
8126
- revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
8127
- split: test
8128
  type: mteb/summeval
 
 
 
8129
  metrics:
8130
  - type: cosine_pearson
8131
  value: 29.88360050203766
@@ -8141,14 +8147,14 @@ model-index:
8141
  value: 29.88360050203766
8142
  - type: spearman
8143
  value: 29.275185932109494
8144
- task:
8145
- type: Summarization
8146
- - dataset:
8147
- config: default
8148
  name: MTEB TRECCOVID (default)
8149
- revision: bb9466bac8153a0349341eb1b22e06409e78ef4e
8150
- split: test
8151
  type: mteb/trec-covid
 
 
 
8152
  metrics:
8153
  - type: main_score
8154
  value: 45.747
@@ -8432,14 +8438,14 @@ model-index:
8432
  value: 0.395
8433
  - type: recall_at_5
8434
  value: 0.628
8435
- task:
8436
  type: Retrieval
8437
- - dataset:
8438
- config: default
8439
  name: MTEB Touche2020 (default)
8440
- revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f
8441
- split: test
8442
  type: mteb/touche2020
 
 
 
8443
  metrics:
8444
  - type: main_score
8445
  value: 18.618000000000002
@@ -8723,14 +8729,14 @@ model-index:
8723
  value: 4.2540000000000004
8724
  - type: recall_at_5
8725
  value: 7.142999999999999
8726
- task:
8727
- type: Retrieval
8728
- - dataset:
8729
- config: default
8730
  name: MTEB ToxicConversationsClassification (default)
8731
- revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
8732
- split: test
8733
  type: mteb/toxic_conversations_50k
 
 
 
8734
  metrics:
8735
  - type: accuracy
8736
  value: 69.0625
@@ -8744,14 +8750,14 @@ model-index:
8744
  value: 75.91632061778698
8745
  - type: main_score
8746
  value: 69.0625
8747
- task:
8748
  type: Classification
8749
- - dataset:
8750
- config: default
8751
  name: MTEB TweetSentimentExtractionClassification (default)
8752
- revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
8753
- split: test
8754
  type: mteb/tweet_sentiment_extraction
 
 
 
8755
  metrics:
8756
  - type: accuracy
8757
  value: 55.387662705149964
@@ -8761,14 +8767,14 @@ model-index:
8761
  value: 55.01561915660653
8762
  - type: main_score
8763
  value: 55.387662705149964
8764
- task:
8765
- type: Classification
8766
- - dataset:
8767
- config: default
8768
  name: MTEB TwentyNewsgroupsClustering (default)
8769
- revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
8770
- split: test
8771
  type: mteb/twentynewsgroups-clustering
 
 
 
8772
  metrics:
8773
  - type: main_score
8774
  value: 33.535908963951435
@@ -8776,14 +8782,14 @@ model-index:
8776
  value: 33.535908963951435
8777
  - type: v_measure_std
8778
  value: 1.8862804680454297
8779
- task:
8780
- type: Clustering
8781
- - dataset:
8782
- config: default
8783
  name: MTEB TwitterSemEval2015 (default)
8784
- revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
8785
- split: test
8786
  type: mteb/twittersemeval2015-pairclassification
 
 
 
8787
  metrics:
8788
  - type: cosine_accuracy
8789
  value: 81.57000655659535
@@ -8867,14 +8873,14 @@ model-index:
8867
  value: 47.93632075471698
8868
  - type: similarity_recall
8869
  value: 64.35356200527704
8870
- task:
8871
  type: PairClassification
8872
- - dataset:
8873
- config: default
8874
  name: MTEB TwitterURLCorpus (default)
8875
- revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
8876
- split: test
8877
  type: mteb/twitterurlcorpus-pairclassification
 
 
 
8878
  metrics:
8879
  - type: cosine_accuracy
8880
  value: 87.09977878682035
@@ -8958,13 +8964,6 @@ model-index:
8958
  value: 70.7599466704091
8959
  - type: similarity_recall
8960
  value: 77.64089929165382
8961
- task:
8962
- type: PairClassification
8963
- model_name: potion-base-8M
8964
- tags:
8965
- - embeddings
8966
- - static-embeddings
8967
- - mteb
8968
  ---
8969
 
8970
  # potion-base-8M Model Card
 
1
  ---
2
  library_name: model2vec
3
  license: mit
4
+ tags:
5
+ - embeddings
6
+ - static-embeddings
7
+ - mteb
8
  model-index:
9
  - name: potion-base-8M
10
  results:
11
+ - task:
12
+ type: Classification
13
+ dataset:
14
  name: MTEB AmazonCounterfactualClassification (en-ext)
 
 
15
  type: mteb/amazon_counterfactual
16
+ config: en-ext
17
+ split: test
18
+ revision: e8379541af4e31359cca9fbcf4b00f2671dba205
19
  metrics:
20
  - type: accuracy
21
  value: 72.15142428785607
 
29
  value: 77.34906471545477
30
  - type: main_score
31
  value: 72.15142428785607
32
+ - task:
33
  type: Classification
34
+ dataset:
 
35
  name: MTEB AmazonCounterfactualClassification (en)
 
 
36
  type: mteb/amazon_counterfactual
37
+ config: en
38
+ split: test
39
+ revision: e8379541af4e31359cca9fbcf4b00f2671dba205
40
  metrics:
41
  - type: accuracy
42
  value: 71.7910447761194
 
50
  value: 74.32788084269461
51
  - type: main_score
52
  value: 71.7910447761194
53
+ - task:
54
  type: Classification
55
+ dataset:
 
56
  name: MTEB AmazonPolarityClassification (default)
 
 
57
  type: mteb/amazon_polarity
58
+ config: default
59
+ split: test
60
+ revision: e2d317d38cd51312af73b3d32a06d1a08b442046
61
  metrics:
62
  - type: accuracy
63
  value: 72.47644999999999
 
71
  value: 72.2600863044581
72
  - type: main_score
73
  value: 72.47644999999999
74
+ - task:
75
  type: Classification
76
+ dataset:
 
77
  name: MTEB AmazonReviewsClassification (en)
 
 
78
  type: mteb/amazon_reviews_multi
79
+ config: en
80
+ split: test
81
+ revision: 1399c76144fd37290681b995c656ef9b2e06e26d
82
  metrics:
83
  - type: accuracy
84
  value: 36.012
 
88
  value: 35.38209336470206
89
  - type: main_score
90
  value: 36.012
91
+ - task:
92
+ type: Retrieval
93
+ dataset:
 
94
  name: MTEB ArguAna (default)
 
 
95
  type: mteb/arguana
96
+ config: default
97
+ split: test
98
+ revision: c22ab2a51041ffd869aaddef7af8d8215647e41a
99
  metrics:
100
  - type: main_score
101
  value: 41.966
 
379
  value: 42.888
380
  - type: recall_at_5
381
  value: 52.063
382
+ - task:
383
+ type: Clustering
384
+ dataset:
 
385
  name: MTEB ArxivClusteringP2P (default)
 
 
386
  type: mteb/arxiv-clustering-p2p
387
+ config: default
388
+ split: test
389
+ revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
390
  metrics:
391
  - type: main_score
392
  value: 35.387660145946825
 
394
  value: 35.387660145946825
395
  - type: v_measure_std
396
  value: 14.022525689022785
397
+ - task:
398
  type: Clustering
399
+ dataset:
 
400
  name: MTEB ArxivClusteringS2S (default)
 
 
401
  type: mteb/arxiv-clustering-s2s
402
+ config: default
403
+ split: test
404
+ revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
405
  metrics:
406
  - type: main_score
407
  value: 25.26058942964131
 
409
  value: 25.26058942964131
410
  - type: v_measure_std
411
  value: 14.850432186356857
412
+ - task:
413
+ type: Reranking
414
+ dataset:
 
415
  name: MTEB AskUbuntuDupQuestions (default)
 
 
416
  type: mteb/askubuntudupquestions-reranking
417
+ config: default
418
+ split: test
419
+ revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
420
  metrics:
421
  - type: main_score
422
  value: 54.13950871400633
 
436
  value: 21.861692980537956
437
  - type: nAUC_mrr_std
438
  value: 8.256966784037171
439
+ - task:
440
+ type: STS
441
+ dataset:
 
442
  name: MTEB BIOSSES (default)
 
 
443
  type: mteb/biosses-sts
444
+ config: default
445
+ split: test
446
+ revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
447
  metrics:
448
  - type: cosine_pearson
449
  value: 79.11612010879227
 
463
  value: 79.11612010879227
464
  - type: spearman
465
  value: 75.85775256673794
466
+ - task:
467
+ type: Classification
468
+ dataset:
 
469
  name: MTEB Banking77Classification (default)
 
 
470
  type: mteb/banking77
471
+ config: default
472
+ split: test
473
+ revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
474
  metrics:
475
  - type: accuracy
476
  value: 72.63636363636363
 
480
  value: 71.69751597573539
481
  - type: main_score
482
  value: 72.63636363636363
483
+ - task:
484
+ type: Clustering
485
+ dataset:
 
486
  name: MTEB BiorxivClusteringP2P (default)
 
 
487
  type: mteb/biorxiv-clustering-p2p
488
+ config: default
489
+ split: test
490
+ revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
491
  metrics:
492
  - type: main_score
493
  value: 30.861840536151014
 
495
  value: 30.861840536151014
496
  - type: v_measure_std
497
  value: 0.8096483751274005
498
+ - task:
499
  type: Clustering
500
+ dataset:
 
501
  name: MTEB BiorxivClusteringS2S (default)
 
 
502
  type: mteb/biorxiv-clustering-s2s
503
+ config: default
504
+ split: test
505
+ revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
506
  metrics:
507
  - type: main_score
508
  value: 20.219544420664455
 
510
  value: 20.219544420664455
511
  - type: v_measure_std
512
  value: 0.7431903039116942
513
+ - task:
514
+ type: Retrieval
515
+ dataset:
 
516
  name: MTEB CQADupstackAndroidRetrieval (default)
 
 
517
  type: mteb/cqadupstack-android
518
+ config: default
519
+ split: test
520
+ revision: f46a197baaae43b4f621051089b82a364682dfeb
521
  metrics:
522
  - type: main_score
523
  value: 31.835
 
801
  value: 29.301
802
  - type: recall_at_5
803
  value: 33.918
804
+ - task:
805
  type: Retrieval
806
+ dataset:
 
807
  name: MTEB CQADupstackEnglishRetrieval (default)
 
 
808
  type: mteb/cqadupstack-english
809
+ config: default
810
+ split: test
811
+ revision: ad9991cb51e31e31e430383c75ffb2885547b5f0
812
  metrics:
813
  - type: main_score
814
  value: 25.734
 
1092
  value: 23.54
1093
  - type: recall_at_5
1094
  value: 28.054000000000002
1095
+ - task:
1096
  type: Retrieval
1097
+ dataset:
 
1098
  name: MTEB CQADupstackGamingRetrieval (default)
 
 
1099
  type: mteb/cqadupstack-gaming
1100
+ config: default
1101
+ split: test
1102
+ revision: 4885aa143210c98657558c04aaf3dc47cfb54340
1103
  metrics:
1104
  - type: main_score
1105
  value: 39.235
 
1383
  value: 37.830999999999996
1384
  - type: recall_at_5
1385
  value: 43.811
1386
+ - task:
1387
  type: Retrieval
1388
+ dataset:
 
1389
  name: MTEB CQADupstackGisRetrieval (default)
 
 
1390
  type: mteb/cqadupstack-gis
1391
+ config: default
1392
+ split: test
1393
+ revision: 5003b3064772da1887988e05400cf3806fe491f2
1394
  metrics:
1395
  - type: main_score
1396
  value: 19.197
 
1674
  value: 19.417
1675
  - type: recall_at_5
1676
  value: 23.577
1677
+ - task:
1678
  type: Retrieval
1679
+ dataset:
 
1680
  name: MTEB CQADupstackMathematicaRetrieval (default)
 
 
1681
  type: mteb/cqadupstack-mathematica
1682
+ config: default
1683
+ split: test
1684
+ revision: 90fceea13679c63fe563ded68f3b6f06e50061de
1685
  metrics:
1686
  - type: main_score
1687
  value: 12.145
 
1965
  value: 11.178
1966
  - type: recall_at_5
1967
  value: 13.877
1968
+ - task:
1969
  type: Retrieval
1970
+ dataset:
 
1971
  name: MTEB CQADupstackPhysicsRetrieval (default)
 
 
1972
  type: mteb/cqadupstack-physics
1973
+ config: default
1974
+ split: test
1975
+ revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4
1976
  metrics:
1977
  - type: main_score
1978
  value: 26.667999999999996
 
2256
  value: 24.524
2257
  - type: recall_at_5
2258
  value: 28.787000000000003
2259
+ - task:
2260
  type: Retrieval
2261
+ dataset:
 
2262
  name: MTEB CQADupstackProgrammersRetrieval (default)
 
 
2263
  type: mteb/cqadupstack-programmers
2264
+ config: default
2265
+ split: test
2266
+ revision: 6184bc1440d2dbc7612be22b50686b8826d22b32
2267
  metrics:
2268
  - type: main_score
2269
  value: 21.66
 
2547
  value: 19.79
2548
  - type: recall_at_5
2549
  value: 23.376
2550
+ - task:
2551
  type: Retrieval
2552
+ dataset:
 
2553
  name: MTEB CQADupstackRetrieval (default)
 
 
2554
  type: CQADupstackRetrieval_is_a_combined_dataset
2555
+ config: default
2556
+ split: test
2557
+ revision: CQADupstackRetrieval_is_a_combined_dataset
2558
  metrics:
2559
  - type: main_score
2560
  value: 22.302333333333333
2561
  - type: ndcg_at_10
2562
  value: 22.302333333333333
2563
+ - task:
2564
  type: Retrieval
2565
+ dataset:
 
2566
  name: MTEB CQADupstackStatsRetrieval (default)
 
 
2567
  type: mteb/cqadupstack-stats
2568
+ config: default
2569
+ split: test
2570
+ revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a
2571
  metrics:
2572
  - type: main_score
2573
  value: 17.253
 
2851
  value: 17.134
2852
  - type: recall_at_5
2853
  value: 19.72
2854
+ - task:
2855
  type: Retrieval
2856
+ dataset:
 
2857
  name: MTEB CQADupstackTexRetrieval (default)
 
 
2858
  type: mteb/cqadupstack-tex
2859
+ config: default
2860
+ split: test
2861
+ revision: 46989137a86843e03a6195de44b09deda022eec7
2862
  metrics:
2863
  - type: main_score
2864
  value: 13.308
 
3142
  value: 12.199
3143
  - type: recall_at_5
3144
  value: 14.87
3145
+ - task:
3146
  type: Retrieval
3147
+ dataset:
 
3148
  name: MTEB CQADupstackUnixRetrieval (default)
 
 
3149
  type: mteb/cqadupstack-unix
3150
+ config: default
3151
+ split: test
3152
+ revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53
3153
  metrics:
3154
  - type: main_score
3155
  value: 20.061999999999998
 
3433
  value: 18.859
3434
  - type: recall_at_5
3435
  value: 22.435
3436
+ - task:
3437
  type: Retrieval
3438
+ dataset:
 
3439
  name: MTEB CQADupstackWebmastersRetrieval (default)
 
 
3440
  type: mteb/cqadupstack-webmasters
3441
+ config: default
3442
+ split: test
3443
+ revision: 160c094312a0e1facb97e55eeddb698c0abe3571
3444
  metrics:
3445
  - type: main_score
3446
  value: 23.735999999999997
 
3724
  value: 21.397
3725
  - type: recall_at_5
3726
  value: 24.863
3727
+ - task:
3728
  type: Retrieval
3729
+ dataset:
 
3730
  name: MTEB CQADupstackWordpressRetrieval (default)
 
 
3731
  type: mteb/cqadupstack-wordpress
3732
+ config: default
3733
+ split: test
3734
+ revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4
3735
  metrics:
3736
  - type: main_score
3737
  value: 16.794999999999998
 
4015
  value: 15.440999999999999
4016
  - type: recall_at_5
4017
  value: 18.829
4018
+ - task:
4019
  type: Retrieval
4020
+ dataset:
 
4021
  name: MTEB ClimateFEVER (default)
 
 
4022
  type: mteb/climate-fever
4023
+ config: default
4024
+ split: test
4025
+ revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380
4026
  metrics:
4027
  - type: main_score
4028
  value: 17.288999999999998
 
4306
  value: 12.232
4307
  - type: recall_at_5
4308
  value: 15.540999999999999
4309
+ - task:
4310
  type: Retrieval
4311
+ dataset:
 
4312
  name: MTEB DBPedia (default)
 
 
4313
  type: mteb/dbpedia
4314
+ config: default
4315
+ split: test
4316
+ revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659
4317
  metrics:
4318
  - type: main_score
4319
  value: 24.235
 
4597
  value: 8.698
4598
  - type: recall_at_5
4599
  value: 11.052
4600
+ - task:
4601
+ type: Classification
4602
+ dataset:
 
4603
  name: MTEB EmotionClassification (default)
 
 
4604
  type: mteb/emotion
4605
+ config: default
4606
+ split: test
4607
+ revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
4608
  metrics:
4609
  - type: accuracy
4610
  value: 45.09
 
4614
  value: 47.129694558751545
4615
  - type: main_score
4616
  value: 45.09
4617
+ - task:
4618
+ type: Retrieval
4619
+ dataset:
 
4620
  name: MTEB FEVER (default)
 
 
4621
  type: mteb/fever
4622
+ config: default
4623
+ split: test
4624
+ revision: bea83ef9e8fb933d90a2f1d5515737465d613e12
4625
  metrics:
4626
  - type: main_score
4627
  value: 30.267
 
4905
  value: 29.42
4906
  - type: recall_at_5
4907
  value: 36.112
4908
+ - task:
4909
  type: Retrieval
4910
+ dataset:
 
4911
  name: MTEB FiQA2018 (default)
 
 
4912
  type: mteb/fiqa
4913
+ config: default
4914
+ split: test
4915
+ revision: 27a168819829fe9bcd655c2df245fb19452e8e06
4916
  metrics:
4917
  - type: main_score
4918
  value: 16.619
 
5196
  value: 12.485
5197
  - type: recall_at_5
5198
  value: 15.862000000000002
5199
+ - task:
5200
  type: Retrieval
5201
+ dataset:
 
5202
  name: MTEB HotpotQA (default)
 
 
5203
  type: mteb/hotpotqa
5204
+ config: default
5205
+ split: test
5206
+ revision: ab518f4d6fcca38d87c25209f94beba119d02014
5207
  metrics:
5208
  - type: main_score
5209
  value: 36.217
 
5487
  value: 30.311
5488
  - type: recall_at_5
5489
  value: 34.092
5490
+ - task:
5491
+ type: Classification
5492
+ dataset:
 
5493
  name: MTEB ImdbClassification (default)
 
 
5494
  type: mteb/imdb
5495
+ config: default
5496
+ split: test
5497
+ revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
5498
  metrics:
5499
  - type: accuracy
5500
  value: 70.78
 
5508
  value: 70.56498155979408
5509
  - type: main_score
5510
  value: 70.78
5511
+ - task:
5512
+ type: Retrieval
5513
+ dataset:
 
5514
  name: MTEB MSMARCO (default)
 
 
5515
  type: mteb/msmarco
5516
+ config: default
5517
+ split: test
5518
+ revision: c5a29a104738b98a9e76336939199e264163d4a0
5519
  metrics:
5520
  - type: main_score
5521
  value: 34.981
 
5799
  value: 2.8240000000000003
5800
  - type: recall_at_5
5801
  value: 4.476999999999999
5802
+ - task:
5803
+ type: Classification
5804
+ dataset:
 
5805
  name: MTEB MTOPDomainClassification (en)
 
 
5806
  type: mteb/mtop_domain
5807
+ config: en
5808
+ split: test
5809
+ revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
5810
  metrics:
5811
  - type: accuracy
5812
  value: 88.95576835385319
 
5816
  value: 89.00721562093213
5817
  - type: main_score
5818
  value: 88.95576835385319
5819
+ - task:
5820
  type: Classification
5821
+ dataset:
 
5822
  name: MTEB MTOPIntentClassification (en)
 
 
5823
  type: mteb/mtop_intent
5824
+ config: en
5825
+ split: test
5826
+ revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
5827
  metrics:
5828
  - type: accuracy
5829
  value: 56.99726402188783
 
5833
  value: 59.96788951671549
5834
  - type: main_score
5835
  value: 56.99726402188783
5836
+ - task:
5837
  type: Classification
5838
+ dataset:
 
5839
  name: MTEB MassiveIntentClassification (en)
 
 
5840
  type: mteb/amazon_massive_intent
5841
+ config: en
5842
+ split: test
5843
+ revision: 4672e20407010da34463acc759c162ca9734bca6
5844
  metrics:
5845
  - type: accuracy
5846
  value: 63.79287155346336
 
5850
  value: 62.567311481126055
5851
  - type: main_score
5852
  value: 63.79287155346336
5853
+ - task:
5854
  type: Classification
5855
+ dataset:
 
5856
  name: MTEB MassiveScenarioClassification (en)
 
 
5857
  type: mteb/amazon_massive_scenario
5858
+ config: en
5859
+ split: test
5860
+ revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8
5861
  metrics:
5862
  - type: accuracy
5863
  value: 70.30934767989241
 
5867
  value: 70.1128179307388
5868
  - type: main_score
5869
  value: 70.30934767989241
5870
+ - task:
5871
+ type: Clustering
5872
+ dataset:
 
5873
  name: MTEB MedrxivClusteringP2P (default)
 
 
5874
  type: mteb/medrxiv-clustering-p2p
5875
+ config: default
5876
+ split: test
5877
+ revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
5878
  metrics:
5879
  - type: main_score
5880
  value: 27.61734940907637
 
5882
  value: 27.61734940907637
5883
  - type: v_measure_std
5884
  value: 1.2248100208316097
5885
+ - task:
5886
  type: Clustering
5887
+ dataset:
 
5888
  name: MTEB MedrxivClusteringS2S (default)
 
 
5889
  type: mteb/medrxiv-clustering-s2s
5890
+ config: default
5891
+ split: test
5892
+ revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
5893
  metrics:
5894
  - type: main_score
5895
  value: 23.802943866708308
 
5897
  value: 23.802943866708308
5898
  - type: v_measure_std
5899
  value: 1.4975518910969763
5900
+ - task:
5901
+ type: Reranking
5902
+ dataset:
 
5903
  name: MTEB MindSmallReranking (default)
 
 
5904
  type: mteb/mind_small
5905
+ config: default
5906
+ split: test
5907
+ revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
5908
  metrics:
5909
  - type: main_score
5910
  value: 29.431722284942175
 
5924
  value: -17.42779158552557
5925
  - type: nAUC_mrr_std
5926
  value: -5.997215692334967
5927
+ - task:
5928
+ type: Retrieval
5929
+ dataset:
 
5930
  name: MTEB NFCorpus (default)
 
 
5931
  type: mteb/nfcorpus
5932
+ config: default
5933
+ split: test
5934
+ revision: ec0fa4fe99da2ff19ca1214b7966684033a58814
5935
  metrics:
5936
  - type: main_score
5937
  value: 24.267
 
6215
  value: 7.236
6216
  - type: recall_at_5
6217
  value: 9.003
6218
+ - task:
6219
  type: Retrieval
6220
+ dataset:
 
6221
  name: MTEB NQ (default)
 
 
6222
  type: mteb/nq
6223
+ config: default
6224
+ split: test
6225
+ revision: b774495ed302d8c44a3a7ea25c90dbce03968f31
6226
  metrics:
6227
  - type: main_score
6228
  value: 19.543
 
6506
  value: 17.669
6507
  - type: recall_at_5
6508
  value: 23.488999999999997
6509
+ - task:
6510
  type: Retrieval
6511
+ dataset:
 
6512
  name: MTEB QuoraRetrieval (default)
 
 
6513
  type: mteb/quora
6514
+ config: default
6515
+ split: test
6516
+ revision: e4e08e0b7dbe3c8700f0daef558ff32256715259
6517
  metrics:
6518
  - type: main_score
6519
  value: 80.598
 
6797
  value: 79.074
6798
  - type: recall_at_5
6799
  value: 84.143
6800
+ - task:
6801
+ type: Clustering
6802
+ dataset:
 
6803
  name: MTEB RedditClustering (default)
 
 
6804
  type: mteb/reddit-clustering
6805
+ config: default
6806
+ split: test
6807
+ revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
6808
  metrics:
6809
  - type: main_score
6810
  value: 39.5465127563479
 
6812
  value: 39.5465127563479
6813
  - type: v_measure_std
6814
  value: 5.038703300031419
6815
+ - task:
6816
  type: Clustering
6817
+ dataset:
 
6818
  name: MTEB RedditClusteringP2P (default)
 
 
6819
  type: mteb/reddit-clustering-p2p
6820
+ config: default
6821
+ split: test
6822
+ revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
6823
  metrics:
6824
  - type: main_score
6825
  value: 47.07911795189491
 
6827
  value: 47.07911795189491
6828
  - type: v_measure_std
6829
  value: 11.546436135362846
6830
+ - task:
6831
+ type: Retrieval
6832
+ dataset:
 
6833
  name: MTEB SCIDOCS (default)
 
 
6834
  type: mteb/scidocs
6835
+ config: default
6836
+ split: test
6837
+ revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88
6838
  metrics:
6839
  - type: main_score
6840
  value: 12.386999999999999
 
7118
  value: 6.553000000000001
7119
  - type: recall_at_5
7120
  value: 9.013
7121
+ - task:
7122
+ type: STS
7123
+ dataset:
 
7124
  name: MTEB SICK-R (default)
 
 
7125
  type: mteb/sickr-sts
7126
+ config: default
7127
+ split: test
7128
+ revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
7129
  metrics:
7130
  - type: cosine_pearson
7131
  value: 75.67336823619708
 
7145
  value: 75.67336823619708
7146
  - type: spearman
7147
  value: 64.6753400763881
7148
+ - task:
7149
  type: STS
7150
+ dataset:
 
7151
  name: MTEB STS12 (default)
 
 
7152
  type: mteb/sts12-sts
7153
+ config: default
7154
+ split: test
7155
+ revision: a0d554a64d88156834ff5ae9920b964011b16384
7156
  metrics:
7157
  - type: cosine_pearson
7158
  value: 72.06681927996405
 
7172
  value: 72.06681927996405
7173
  - type: spearman
7174
  value: 62.248985055530525
7175
+ - task:
7176
  type: STS
7177
+ dataset:
 
7178
  name: MTEB STS13 (default)
 
 
7179
  type: mteb/sts13-sts
7180
+ config: default
7181
+ split: test
7182
+ revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
7183
  metrics:
7184
  - type: cosine_pearson
7185
  value: 76.53983680018591
 
7199
  value: 76.53983680018591
7200
  - type: spearman
7201
  value: 77.27600787572996
7202
+ - task:
7203
  type: STS
7204
+ dataset:
 
7205
  name: MTEB STS14 (default)
 
 
7206
  type: mteb/sts14-sts
7207
+ config: default
7208
+ split: test
7209
+ revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
7210
  metrics:
7211
  - type: cosine_pearson
7212
  value: 76.20854411766629
 
7226
  value: 76.20854411766629
7227
  - type: spearman
7228
  value: 71.914099628002
7229
+ - task:
7230
  type: STS
7231
+ dataset:
 
7232
  name: MTEB STS15 (default)
 
 
7233
  type: mteb/sts15-sts
7234
+ config: default
7235
+ split: test
7236
+ revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
7237
  metrics:
7238
  - type: cosine_pearson
7239
  value: 79.24273419832653
 
7253
  value: 79.24273419832653
7254
  - type: spearman
7255
  value: 79.75345871163103
7256
+ - task:
7257
  type: STS
7258
+ dataset:
 
7259
  name: MTEB STS16 (default)
 
 
7260
  type: mteb/sts16-sts
7261
+ config: default
7262
+ split: test
7263
+ revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
7264
  metrics:
7265
  - type: cosine_pearson
7266
  value: 75.64452330127995
 
7280
  value: 75.64452330127995
7281
  - type: spearman
7282
  value: 76.26343823222666
7283
+ - task:
7284
  type: STS
7285
+ dataset:
 
7286
  name: MTEB STS17 (es-en)
 
 
7287
  type: mteb/sts17-crosslingual-sts
7288
+ config: es-en
7289
+ split: test
7290
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7291
  metrics:
7292
  - type: cosine_pearson
7293
  value: 17.52217310066287
 
7307
  value: 17.52217310066287
7308
  - type: spearman
7309
  value: 14.729958484232528
7310
+ - task:
7311
  type: STS
7312
+ dataset:
 
7313
  name: MTEB STS17 (en-en)
 
 
7314
  type: mteb/sts17-crosslingual-sts
7315
+ config: en-en
7316
+ split: test
7317
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7318
  metrics:
7319
  - type: cosine_pearson
7320
  value: 84.67406984717113
 
7334
  value: 84.67406984717113
7335
  - type: spearman
7336
  value: 85.96709815630739
7337
+ - task:
7338
  type: STS
7339
+ dataset:
 
7340
  name: MTEB STS17 (fr-en)
 
 
7341
  type: mteb/sts17-crosslingual-sts
7342
+ config: fr-en
7343
+ split: test
7344
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7345
  metrics:
7346
  - type: cosine_pearson
7347
  value: 36.02331692863771
 
7361
  value: 36.02331692863771
7362
  - type: spearman
7363
  value: 34.28540470062557
7364
+ - task:
7365
  type: STS
7366
+ dataset:
 
7367
  name: MTEB STS17 (en-tr)
 
 
7368
  type: mteb/sts17-crosslingual-sts
7369
+ config: en-tr
7370
+ split: test
7371
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7372
  metrics:
7373
  - type: cosine_pearson
7374
  value: 13.925983981770388
 
7388
  value: 13.925983981770388
7389
  - type: spearman
7390
  value: 11.193291331109325
7391
+ - task:
7392
  type: STS
7393
+ dataset:
 
7394
  name: MTEB STS17 (en-de)
 
 
7395
  type: mteb/sts17-crosslingual-sts
7396
+ config: en-de
7397
+ split: test
7398
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7399
  metrics:
7400
  - type: cosine_pearson
7401
  value: 26.77839285232968
 
7415
  value: 26.77839285232968
7416
  - type: spearman
7417
  value: 23.010015986939717
7418
+ - task:
7419
  type: STS
7420
+ dataset:
 
7421
  name: MTEB STS17 (it-en)
 
 
7422
  type: mteb/sts17-crosslingual-sts
7423
+ config: it-en
7424
+ split: test
7425
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7426
  metrics:
7427
  - type: cosine_pearson
7428
  value: 25.330935194314364
 
7442
  value: 25.330935194314364
7443
  - type: spearman
7444
  value: 23.143555348782797
7445
+ - task:
7446
  type: STS
7447
+ dataset:
 
7448
  name: MTEB STS17 (en-ar)
 
 
7449
  type: mteb/sts17-crosslingual-sts
7450
+ config: en-ar
7451
+ split: test
7452
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7453
  metrics:
7454
  - type: cosine_pearson
7455
  value: 6.61651078645899
 
7469
  value: 6.61651078645899
7470
  - type: spearman
7471
  value: 5.415104433010482
7472
+ - task:
7473
  type: STS
7474
+ dataset:
 
7475
  name: MTEB STS17 (nl-en)
 
 
7476
  type: mteb/sts17-crosslingual-sts
7477
+ config: nl-en
7478
+ split: test
7479
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7480
  metrics:
7481
  - type: cosine_pearson
7482
  value: 32.718045784523184
 
7496
  value: 32.718045784523184
7497
  - type: spearman
7498
  value: 27.52844368619317
7499
+ - task:
7500
  type: STS
7501
+ dataset:
 
7502
  name: MTEB STS22 (es-en)
 
 
7503
  type: mteb/sts22-crosslingual-sts
7504
+ config: es-en
7505
+ split: test
7506
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7507
  metrics:
7508
  - type: cosine_pearson
7509
  value: 9.98410299881163
 
7523
  value: 9.98410299881163
7524
  - type: spearman
7525
  value: 10.98684405086525
7526
+ - task:
7527
  type: STS
7528
+ dataset:
 
7529
  name: MTEB STS22 (en)
 
 
7530
  type: mteb/sts22-crosslingual-sts
7531
+ config: en
7532
+ split: test
7533
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7534
  metrics:
7535
  - type: cosine_pearson
7536
  value: 59.3180680265132
 
7550
  value: 59.3180680265132
7551
  - type: spearman
7552
  value: 63.07956002739231
7553
+ - task:
7554
  type: STS
7555
+ dataset:
 
7556
  name: MTEB STS22 (de-en)
 
 
7557
  type: mteb/sts22-crosslingual-sts
7558
+ config: de-en
7559
+ split: test
7560
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7561
  metrics:
7562
  - type: cosine_pearson
7563
  value: 29.061215770374826
 
7577
  value: 29.061215770374826
7578
  - type: spearman
7579
  value: 36.21441725938738
7580
+ - task:
7581
  type: STS
7582
+ dataset:
 
7583
  name: MTEB STS22 (zh-en)
 
 
7584
  type: mteb/sts22-crosslingual-sts
7585
+ config: zh-en
7586
+ split: test
7587
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7588
  metrics:
7589
  - type: cosine_pearson
7590
  value: 11.266385865086239
 
7604
  value: 11.266385865086239
7605
  - type: spearman
7606
  value: 17.291293843893733
7607
+ - task:
7608
  type: STS
7609
+ dataset:
 
7610
  name: MTEB STS22 (pl-en)
 
 
7611
  type: mteb/sts22-crosslingual-sts
7612
+ config: pl-en
7613
+ split: test
7614
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7615
  metrics:
7616
  - type: cosine_pearson
7617
  value: 9.647587208410648
 
7631
  value: 9.647587208410648
7632
  - type: spearman
7633
  value: 21.33739699413266
7634
+ - task:
7635
  type: STS
7636
+ dataset:
 
7637
  name: MTEB STSBenchmark (default)
 
 
7638
  type: mteb/stsbenchmark-sts
7639
+ config: default
7640
+ split: test
7641
+ revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
7642
  metrics:
7643
  - type: cosine_pearson
7644
  value: 77.2598255013409
 
7658
  value: 77.2598255013409
7659
  - type: spearman
7660
  value: 75.40519061413276
7661
+ - task:
7662
+ type: Reranking
7663
+ dataset:
 
7664
  name: MTEB SciDocsRR (default)
 
 
7665
  type: mteb/scidocs-reranking
7666
+ config: default
7667
+ split: test
7668
+ revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
7669
  metrics:
7670
  - type: main_score
7671
  value: 72.10127087089839
 
7685
  value: 74.90946175462605
7686
  - type: nAUC_mrr_std
7687
  value: 71.81332269641806
7688
+ - task:
7689
+ type: Retrieval
7690
+ dataset:
 
7691
  name: MTEB SciFact (default)
 
 
7692
  type: mteb/scifact
7693
+ config: default
7694
+ split: test
7695
+ revision: 0228b52cf27578f30900b9e5271d331663a030d7
7696
  metrics:
7697
  - type: main_score
7698
  value: 50.63999999999999
 
7976
  value: 50.306
7977
  - type: recall_at_5
7978
  value: 57.443999999999996
7979
+ - task:
7980
+ type: PairClassification
7981
+ dataset:
 
7982
  name: MTEB SprintDuplicateQuestions (default)
 
 
7983
  type: mteb/sprintduplicatequestions-pairclassification
7984
+ config: default
7985
+ split: test
7986
+ revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
7987
  metrics:
7988
  - type: cosine_accuracy
7989
  value: 99.71386138613862
 
8067
  value: 85.7727737973388
8068
  - type: similarity_recall
8069
  value: 83.8
8070
+ - task:
8071
+ type: Clustering
8072
+ dataset:
 
8073
  name: MTEB StackExchangeClustering (default)
 
 
8074
  type: mteb/stackexchange-clustering
8075
+ config: default
8076
+ split: test
8077
+ revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
8078
  metrics:
8079
  - type: main_score
8080
  value: 48.18939518021159
 
8082
  value: 48.18939518021159
8083
  - type: v_measure_std
8084
  value: 4.6189444340187995
8085
+ - task:
8086
  type: Clustering
8087
+ dataset:
 
8088
  name: MTEB StackExchangeClusteringP2P (default)
 
 
8089
  type: mteb/stackexchange-clustering-p2p
8090
+ config: default
8091
+ split: test
8092
+ revision: 815ca46b2622cec33ccafc3735d572c266efdb44
8093
  metrics:
8094
  - type: main_score
8095
  value: 30.743938802421265
 
8097
  value: 30.743938802421265
8098
  - type: v_measure_std
8099
  value: 1.4645401677053824
8100
+ - task:
8101
+ type: Reranking
8102
+ dataset:
 
8103
  name: MTEB StackOverflowDupQuestions (default)
 
 
8104
  type: mteb/stackoverflowdupquestions-reranking
8105
+ config: default
8106
+ split: test
8107
+ revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
8108
  metrics:
8109
  - type: main_score
8110
  value: 43.254152892780986
 
8124
  value: 14.093758435205075
8125
  - type: nAUC_mrr_std
8126
  value: 4.198791420159734
8127
+ - task:
8128
+ type: Summarization
8129
+ dataset:
 
8130
  name: MTEB SummEval (default)
 
 
8131
  type: mteb/summeval
8132
+ config: default
8133
+ split: test
8134
+ revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
8135
  metrics:
8136
  - type: cosine_pearson
8137
  value: 29.88360050203766
 
8147
  value: 29.88360050203766
8148
  - type: spearman
8149
  value: 29.275185932109494
8150
+ - task:
8151
+ type: Retrieval
8152
+ dataset:
 
8153
  name: MTEB TRECCOVID (default)
 
 
8154
  type: mteb/trec-covid
8155
+ config: default
8156
+ split: test
8157
+ revision: bb9466bac8153a0349341eb1b22e06409e78ef4e
8158
  metrics:
8159
  - type: main_score
8160
  value: 45.747
 
8438
  value: 0.395
8439
  - type: recall_at_5
8440
  value: 0.628
8441
+ - task:
8442
  type: Retrieval
8443
+ dataset:
 
8444
  name: MTEB Touche2020 (default)
 
 
8445
  type: mteb/touche2020
8446
+ config: default
8447
+ split: test
8448
+ revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f
8449
  metrics:
8450
  - type: main_score
8451
  value: 18.618000000000002
 
8729
  value: 4.2540000000000004
8730
  - type: recall_at_5
8731
  value: 7.142999999999999
8732
+ - task:
8733
+ type: Classification
8734
+ dataset:
 
8735
  name: MTEB ToxicConversationsClassification (default)
 
 
8736
  type: mteb/toxic_conversations_50k
8737
+ config: default
8738
+ split: test
8739
+ revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
8740
  metrics:
8741
  - type: accuracy
8742
  value: 69.0625
 
8750
  value: 75.91632061778698
8751
  - type: main_score
8752
  value: 69.0625
8753
+ - task:
8754
  type: Classification
8755
+ dataset:
 
8756
  name: MTEB TweetSentimentExtractionClassification (default)
 
 
8757
  type: mteb/tweet_sentiment_extraction
8758
+ config: default
8759
+ split: test
8760
+ revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
8761
  metrics:
8762
  - type: accuracy
8763
  value: 55.387662705149964
 
8767
  value: 55.01561915660653
8768
  - type: main_score
8769
  value: 55.387662705149964
8770
+ - task:
8771
+ type: Clustering
8772
+ dataset:
 
8773
  name: MTEB TwentyNewsgroupsClustering (default)
 
 
8774
  type: mteb/twentynewsgroups-clustering
8775
+ config: default
8776
+ split: test
8777
+ revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
8778
  metrics:
8779
  - type: main_score
8780
  value: 33.535908963951435
 
8782
  value: 33.535908963951435
8783
  - type: v_measure_std
8784
  value: 1.8862804680454297
8785
+ - task:
8786
+ type: PairClassification
8787
+ dataset:
 
8788
  name: MTEB TwitterSemEval2015 (default)
 
 
8789
  type: mteb/twittersemeval2015-pairclassification
8790
+ config: default
8791
+ split: test
8792
+ revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
8793
  metrics:
8794
  - type: cosine_accuracy
8795
  value: 81.57000655659535
 
8873
  value: 47.93632075471698
8874
  - type: similarity_recall
8875
  value: 64.35356200527704
8876
+ - task:
8877
  type: PairClassification
8878
+ dataset:
 
8879
  name: MTEB TwitterURLCorpus (default)
 
 
8880
  type: mteb/twitterurlcorpus-pairclassification
8881
+ config: default
8882
+ split: test
8883
+ revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
8884
  metrics:
8885
  - type: cosine_accuracy
8886
  value: 87.09977878682035
 
8964
  value: 70.7599466704091
8965
  - type: similarity_recall
8966
  value: 77.64089929165382
 
 
 
 
 
 
 
8967
  ---
8968
 
8969
  # potion-base-8M Model Card