SentenceTransformer based on Snowflake/snowflake-arctic-embed-m
This is a sentence-transformers model finetuned from Snowflake/snowflake-arctic-embed-m. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
Model Details
Model Description
- Model Type: Sentence Transformer
- Base model: Snowflake/snowflake-arctic-embed-m
- Maximum Sequence Length: 512 tokens
- Output Dimensionality: 768 tokens
- Similarity Function: Cosine Similarity
Model Sources
- Documentation: Sentence Transformers Documentation
- Repository: Sentence Transformers on GitHub
- Hugging Face: Sentence Transformers on Hugging Face
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Normalize()
)
Usage
Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("federicovolponi/Snowflake-snowflake-arctic-embed-m-space-sup")
# Run inference
sentences = [
': Block diagram of the 7-band CCD-in-CMOS TDI sensor. Each TX slice has two serializers and its own PLL.\nThe CCD bands operate continuously and time interleaved. The output stages for the CCD arrays are implemented both at the top and bottom of each band to support the bi-directional operation. All 14 output stages in one column are connected to one delta-sigma column-level ADC with digital CDS implemented in the digital decimator. The outputs of every 128 ADCs are serialized to one of 32 LVDS outputs. Two clock signals are also provided via LVDS to synchronize the channels. These outputs are capable of running at an aggregate data rate of >50Gb/s using on-chip PLLs.\nThe sensor has been processed for Back-Side Illumination and it has been packaged in a custom ceramic PGA package. Figure 15 shows a picture of the sensor with its 7 bands. The figure shows the front-side and back-side versions of the chip side by side.\n(a) (b) Figure 15: 7-band CCD-in-CMOS TDI chip photograph. FSI shown only for reference (a) and BSI version (b).\nAs a proof-of-concept, an RGB butcher-brick filter has been used as glass lid for the sensor, to enable multicolor TDI, although filters may be processed directly on the wafer as well [9]. The sensor,\ncamera system and a color image captured from the setup are depicted in Figure 16, providing evidence that multispectral TDI is viable with the sensor.\nFigure 16: Colour TDI image captured from the sensor, sensor with RGB color filter and camera set-up.\nTable 3 below shows a comparison of different TDI sensors, including the first iteration of our sensor.\nIntegrated drivers\nThe measurements on the first iteration of the SoC verified',
'What is the aggregate data rate of the outputs of the 7-band CCD-in-CMOS TDI sensor?\n\n',
'What is the primary objective of the Zodiac Pioneer Mission?\n\n',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 768]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
Evaluation
Metrics
Information Retrieval
- Dataset:
dim_768
- Evaluated with
InformationRetrievalEvaluator
Metric | Value |
---|---|
cosine_accuracy@5 | 0.8408 |
cosine_accuracy@10 | 0.8843 |
cosine_precision@5 | 0.1682 |
cosine_precision@10 | 0.0884 |
cosine_recall@5 | 0.8408 |
cosine_recall@10 | 0.8843 |
cosine_ndcg@5 | 0.7496 |
cosine_ndcg@10 | 0.7639 |
cosine_mrr@5 | 0.719 |
cosine_mrr@10 | 0.725 |
cosine_map@5 | 0.719 |
cosine_map@10 | 0.725 |
Information Retrieval
- Dataset:
dim_512
- Evaluated with
InformationRetrievalEvaluator
Metric | Value |
---|---|
cosine_accuracy@5 | 0.8346 |
cosine_accuracy@10 | 0.8781 |
cosine_precision@5 | 0.1669 |
cosine_precision@10 | 0.0878 |
cosine_recall@5 | 0.8346 |
cosine_recall@10 | 0.8781 |
cosine_ndcg@5 | 0.7384 |
cosine_ndcg@10 | 0.7524 |
cosine_mrr@5 | 0.7061 |
cosine_mrr@10 | 0.7118 |
cosine_map@5 | 0.7061 |
cosine_map@10 | 0.7118 |
Information Retrieval
- Dataset:
dim_256
- Evaluated with
InformationRetrievalEvaluator
Metric | Value |
---|---|
cosine_accuracy@5 | 0.8147 |
cosine_accuracy@10 | 0.8632 |
cosine_precision@5 | 0.1629 |
cosine_precision@10 | 0.0863 |
cosine_recall@5 | 0.8147 |
cosine_recall@10 | 0.8632 |
cosine_ndcg@5 | 0.7159 |
cosine_ndcg@10 | 0.7318 |
cosine_mrr@5 | 0.6827 |
cosine_mrr@10 | 0.6894 |
cosine_map@5 | 0.6827 |
cosine_map@10 | 0.6894 |
Information Retrieval
- Dataset:
dim_768
- Evaluated with
InformationRetrievalEvaluator
Metric | Value |
---|---|
cosine_accuracy@5 | 0.9199 |
cosine_accuracy@10 | 0.9551 |
cosine_precision@5 | 0.184 |
cosine_precision@10 | 0.0955 |
cosine_recall@5 | 0.9199 |
cosine_recall@10 | 0.9551 |
cosine_ndcg@5 | 0.786 |
cosine_ndcg@10 | 0.7975 |
cosine_mrr@5 | 0.7408 |
cosine_mrr@10 | 0.7455 |
cosine_map@5 | 0.7408 |
cosine_map@10 | 0.7455 |
Information Retrieval
- Dataset:
dim_512
- Evaluated with
InformationRetrievalEvaluator
Metric | Value |
---|---|
cosine_accuracy@5 | 0.9071 |
cosine_accuracy@10 | 0.9519 |
cosine_precision@5 | 0.1814 |
cosine_precision@10 | 0.0952 |
cosine_recall@5 | 0.9071 |
cosine_recall@10 | 0.9519 |
cosine_ndcg@5 | 0.7794 |
cosine_ndcg@10 | 0.7943 |
cosine_mrr@5 | 0.7363 |
cosine_mrr@10 | 0.7427 |
cosine_map@5 | 0.7363 |
cosine_map@10 | 0.7427 |
Information Retrieval
- Dataset:
dim_256
- Evaluated with
InformationRetrievalEvaluator
Metric | Value |
---|---|
cosine_accuracy@5 | 0.8846 |
cosine_accuracy@10 | 0.9455 |
cosine_precision@5 | 0.1769 |
cosine_precision@10 | 0.0946 |
cosine_recall@5 | 0.8846 |
cosine_recall@10 | 0.9455 |
cosine_ndcg@5 | 0.7548 |
cosine_ndcg@10 | 0.7748 |
cosine_mrr@5 | 0.7108 |
cosine_mrr@10 | 0.7193 |
cosine_map@5 | 0.7108 |
cosine_map@10 | 0.7193 |
Training Details
Training Dataset
Unnamed Dataset
- Size: 7,232 training samples
- Columns:
positive
andanchor
- Approximate statistics based on the first 1000 samples:
positive anchor type string string details - min: 5 tokens
- mean: 354.69 tokens
- max: 512 tokens
- min: 9 tokens
- mean: 19.21 tokens
- max: 40 tokens
- Samples:
positive anchor , using diverse software or hardware designs may double design and verification costs due to having to build two different components for the same functionality. Hence, although DCLS execution also halves performance efficiency (the corresponding functionality is executed twice), it allows reusing the same design (e.g. the same core design) for the primary and the redundant paths (e.g. with staggered execution), thus containing design and verification costs.
Redundancy can be applied at different granularities accord- ing to the sphere of replication (SoR). Choosing the right SoR depends on several tradeoffs like area overheads, re- design costs, fault detection time, and overall system costs. In the context of DCLS, the SoR is placed at the level of the CPU (core), as done for the AURIX processors. This requires including two replicas of the same core and compare their memory transactions, which requires roughly duplicating com- putational resources in the chip and being able to ensure that replicas can provide independent behavior. On the other hand, storage (memories, caches) and communication means (buses, crossbars) do not need to be fully replicated and can build upon Error Correction Codes (ECC) and Cyclic Redundancy Check (CRC) as a form of lightweight redundancy with diversity.
HPC ASIL-D capable platforms typically combine a low- performance microcontroller amenable for the automotive do- main (i.e. ASIL-D capable) and an HPC accelerator deliv- ering high computation throughput, but whose adherence to ISO26262 requirements is unknown, so its appropriate use for ASIL-C/D systems needs to be investigated. Without loss of generality, we consider an NVIDIA GPU accelerator, thus analogous to those in NVIDIA Drive and Xavier families for the automotive domain. However, the findings in this paper can easily be extrapolated to other products.
Software faults and some hardware faults are regarded as systematic, and it must be proven that their failure risk is residual. However, random hardware faults cannot be avoided, and means are required to prevent them from causing hazards. Those faults can be caused by, for example, voltage droopsWhat are the advantages of using the same design for the primary and redundant paths in DCLS execution?
: First, the TT&C spectrum requirements of the new satellites shall be assessed. Second, the utilization of existing TT&C frequency allocations and their potential to incorporate the future number of satellites is studied. Only for the case that this study results in the need for new spectrum, the study groups were asked to investigate new potential TT&C frequency allocations in the frequency ranges 150.05-174 MHz and 400.15-420 MHz. The studies shall be completed for WRC-19.
This paper presents the intermediate results of the study groups. A study of the spectrum requirements of small satellites has been completed. The required spectrum for TT&C is expected to be less than 2.5 MHz for downlink and less than 1 MHz for uplink. Consequently, the study groups conducted sharing studies in various bands which will be summarized and evaluated from a satellite developer’s perspective.
After the Cubesat design standard was introduced in 1999 and first satellites of this new class have been launched in the subsequent years, small satellites have become increasingly popular in the past five years. Today not only universities use small satellite platforms for education and technology demonstration, but also commercial operators started to develop and deploy satellites with masses of typically less than 50 kg and reasonably short development times. Currently more than hundred new satellites are currently launched into space per year. The increase of launches was recognized by the International Telecommunication Union (ITU) which is responsible for the coordination of the shared use of frequencies. As the first Cubesats were mainly launched by new entrants into the space sector, mandatory regulatory procedures like frequency coordination were omitted or underestimated by the developers. Additionally, the new developers complaint that the existing regulatory procedures are too complicated and time-consuming for satellites with short development times. The ITU therefore decided at the WRC-12 to study the characteristics of picosatellites and nanosatellites and their current practice in filing satellites to the ITU. The studies were concluded in 2015 with two reports on the characteristics [1] and current filing practice [2]. In these reports it was identified that the characteristics that define small satellites (low mass, small dimensions, low power, …) are not relevant from a frequency coordination perspective and that the short development times are still long enough to properly file the systems to the ITU. As a resultWhat are the spectrum requirements for TT&C of small satellites?
:287–299, Dec 2019.
[20] Tam´as Vink´o and Dario Izzo. Global optimi- sation heuristics and test problems for prelimi- nary spacecraft trajectory design. Technical re- port, 2008.
[21] Matej Petkovic, Luke Lucas, Dragi Kocev, Saˇso Dˇzeroski, Redouane Boumghar, and Nikola Simidjievski. Quantifying the effects of gyro- less flying of the mars express spacecraft with machine learning. In 2019 IEEE International
[22] Janhavi H. Borse, Dipti D. Patil, Vinod Kumar, and Sudhir Kumar. Soft landing parameter measurements for candidate navigation trajec- tories using deep learning and ai-enabled plan- etary descent. Mathematical Problems in Engi- neering, 2022What are some of the research topics and methods explored in the provided references?
- Loss:
losses.WeightedMultipleNegativesRankingLoss
with these parameters:{ "scale": 20, "similarity_fct": "cos_sim" }
Evaluation Dataset
Unnamed Dataset
- Size: 804 evaluation samples
- Columns:
positive
andanchor
- Approximate statistics based on the first 1000 samples:
positive anchor type string string details - min: 4 tokens
- mean: 351.15 tokens
- max: 512 tokens
- min: 8 tokens
- mean: 19.36 tokens
- max: 45 tokens
- Samples:
positive anchor , the total number of test thermocouples has been rationalized taking into account redundancy needs, accommodation constraints and hardware passivation needs for flight. The test is subdivided into 19 phases (see Figure 12) with two phases before and after the test for the health check functional tests under room conditions. Functional tests demonstrate anomalies such as the PCDU Reset and operational malfunctions of the RAX instrument at its high temperatures. The PCDU Reset anomaly was solved during the test by a software patch and validated during the final hot and cold plateaus. To address the RAX anomaly at hot, various test configurations were simulated using the thermal numerical model during the test to actually perform RAX functional test at an intermediate plateau facilitating mission operational constraints for flight. Data collected from hot and cold thermal balance test phases, as well as the rover OFF transition from hot to cold, are the inputs for correlation activities conducted post-TV/TB test. The thermal numerical model updates mainly focus on conductive couplings
What was the solution to the PCDU Reset anomaly during the test?
, where +Z axis orients to the earth, and sun pointing attitude mode during day time
orienting -Z axis to the sun. Therefore, attitude control subsystem is required to maneuver the satellite attitude twice per revolution around its pitch axis. Figure 6 shows concept of the attitude maneuverer. Another attitude maneuverer is necessary to perform SAR observation and SAR data download to a to ground station, because X-band transmit antenna is oriented to +Z, so the satellite has to offset its attitude to orient the X-band transmit antenna toward the ground station.
3.4 High pointing accuracy
Disturbance torque and system momentum profiles during few revolutions were estimated as shown in Figure 7 and 8. Four micro reaction wheels, which can respond to these profiles were selected which enable attitude maneuvers within a short period of time. In order to perform a pitch attitude maneuver quickly, two wheels are located on pitch axis while one wheel was located on each of the remaining roll and yaw axes. Figure 9 shows the satellite attitudes during SAR observation. There are three kinds of attitude, strip map mode, sliding spot light mode, and spotlight mode. Large change of momentum is required for pitch axis when the satellite is in spotlight mode. However, two pitch reaction wheels do not generate enough momentum to execute spotlight mode. So, sliding spotlight mode was selected for high resolution SAR observation mode instead of spotlight mode, in order to relax the torque and momentum requirements to the pitch wheels. In addition, two pitch
Figure 7. Disturbance torque profile Figure 8. System momentum profile
reaction wheels are accelerated to plus direction or minus direction by using magnet torque before observation. In order to obtain a high resolution SAR data, high attitude control accuracy is required for spotlight mode observation. To achieve high pointing accuracy against a defined ground target point, the attitude control loop applied feed forward compensation with estimated attitude angle and rate. Figure 10 shows an example of dynamic error during a spotlight mode observation maneuver.[4]
Equipment for SAR mission consumes total large power more than 1300W, therefore PCDU has a risk of causing electrical and RF influence to the bus power and signal line. In order to research the system, electrical interface check was performed using bread board model of PCDU, batteryWhat is the reason for selecting sliding spotlight mode instead of spotlight mode for high resolution SAR observation?
, body shape and motion assumptions. Then, ORSAT uses DCA to determine the reentry risk posed to the Earth’s
population based on the year of reentry and orbit inclination. It also predicts impact kinetic energy (impact velocity and impact mass) of objects that survive reentry[18]. ORSAT has been in use for the last decade and currently in its 6.0 version. However, unlike DAS, OR-
SAT is not readily available. Only personnel at the Johnson Space Center, Orbital Debris Program Office run ORSAT. ORSAT is limited to ballistic reentry, only tumbling motions or
stable orientations of objects are allowed which produce no lift. Partial melting of objects is considered by a demise factor and almost all materials in the database are temperature de- pendent. Heating by oxidation is also considered [20]. Therefore, ORSAT determines when
and if a reentry object demises by using integrated trajectory, atmospheric, aerodynamic, aero-thermodynamic, and thermal models as outlined in section 3.1 [17, 18, 20].
Reentry demisability analysis using DAS requires the spacecraft to be defined to the level of each individual hardware part constituting the spacecraft. This step facilitates population
of the DAS Spacecraft Definition Module . Section 3.2.1 illustrates a generic spacecraft subdivision approach that can be followed to itemize the individual parts spacecraft parts.
Subsequently, non-demisable parts are identified before or by the actual reentry analysis as explained in section 3.2.2.
Itemization of the demisable spacecraft basic parts can be best approached by decompos- ing the spacecraft according to the Hierarchical System Terminology defined in the NASA Systems Engineering Handbook [14]. Tables 3.2, 3.3 and 3.4 illustrate a generic approach
to decompose a spacecraft into basic parts [29, 30, 9] excluding the payload. Description of the specific product for the basic part identified completes the process. Though slight vari-
ations are likely to occur in the decomposition of different missions, the Generic Spacecraft Subsystems Hierarchical Subdivision approach is robust, henceWhat is the limitation of ORSAT in terms of object motion?
- Loss:
losses.WeightedMultipleNegativesRankingLoss
with these parameters:{ "scale": 20, "similarity_fct": "cos_sim" }
Training Hyperparameters
Non-Default Hyperparameters
eval_strategy
: stepsper_device_train_batch_size
: 32per_device_eval_batch_size
: 32learning_rate
: 3e-06weight_decay
: 0.001num_train_epochs
: 20bf16
: Truetf32
: Falseload_best_model_at_end
: Truebatch_sampler
: no_duplicates
All Hyperparameters
Click to expand
overwrite_output_dir
: Falsedo_predict
: Falseeval_strategy
: stepsprediction_loss_only
: Trueper_device_train_batch_size
: 32per_device_eval_batch_size
: 32per_gpu_train_batch_size
: Noneper_gpu_eval_batch_size
: Nonegradient_accumulation_steps
: 1eval_accumulation_steps
: Nonelearning_rate
: 3e-06weight_decay
: 0.001adam_beta1
: 0.9adam_beta2
: 0.999adam_epsilon
: 1e-08max_grad_norm
: 1.0num_train_epochs
: 20max_steps
: -1lr_scheduler_type
: linearlr_scheduler_kwargs
: {}warmup_ratio
: 0.0warmup_steps
: 0log_level
: passivelog_level_replica
: warninglog_on_each_node
: Truelogging_nan_inf_filter
: Truesave_safetensors
: Truesave_on_each_node
: Falsesave_only_model
: Falserestore_callback_states_from_checkpoint
: Falseno_cuda
: Falseuse_cpu
: Falseuse_mps_device
: Falseseed
: 42data_seed
: Nonejit_mode_eval
: Falseuse_ipex
: Falsebf16
: Truefp16
: Falsefp16_opt_level
: O1half_precision_backend
: autobf16_full_eval
: Falsefp16_full_eval
: Falsetf32
: Falselocal_rank
: 0ddp_backend
: Nonetpu_num_cores
: Nonetpu_metrics_debug
: Falsedebug
: []dataloader_drop_last
: Falsedataloader_num_workers
: 0dataloader_prefetch_factor
: Nonepast_index
: -1disable_tqdm
: Falseremove_unused_columns
: Truelabel_names
: Noneload_best_model_at_end
: Trueignore_data_skip
: Falsefsdp
: []fsdp_min_num_params
: 0fsdp_config
: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}fsdp_transformer_layer_cls_to_wrap
: Noneaccelerator_config
: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}deepspeed
: Nonelabel_smoothing_factor
: 0.0optim
: adamw_torchoptim_args
: Noneadafactor
: Falsegroup_by_length
: Falselength_column_name
: lengthddp_find_unused_parameters
: Noneddp_bucket_cap_mb
: Noneddp_broadcast_buffers
: Falsedataloader_pin_memory
: Truedataloader_persistent_workers
: Falseskip_memory_metrics
: Trueuse_legacy_prediction_loop
: Falsepush_to_hub
: Falseresume_from_checkpoint
: Nonehub_model_id
: Nonehub_strategy
: every_savehub_private_repo
: Falsehub_always_push
: Falsegradient_checkpointing
: Falsegradient_checkpointing_kwargs
: Noneinclude_inputs_for_metrics
: Falseeval_do_concat_batches
: Truefp16_backend
: autopush_to_hub_model_id
: Nonepush_to_hub_organization
: Nonemp_parameters
:auto_find_batch_size
: Falsefull_determinism
: Falsetorchdynamo
: Noneray_scope
: lastddp_timeout
: 1800torch_compile
: Falsetorch_compile_backend
: Nonetorch_compile_mode
: Nonedispatch_batches
: Nonesplit_batches
: Noneinclude_tokens_per_second
: Falseinclude_num_input_tokens_seen
: Falseneftune_noise_alpha
: Noneoptim_target_modules
: Nonebatch_eval_metrics
: Falsebatch_sampler
: no_duplicatesmulti_dataset_batch_sampler
: proportional
Training Logs
Epoch | Step | Training Loss | loss | dim_256_cosine_map@10 | dim_512_cosine_map@10 | dim_768_cosine_map@10 |
---|---|---|---|---|---|---|
0.4425 | 100 | 0.5883 | - | - | - | - |
0.8850 | 200 | 0.2765 | - | - | - | - |
1.3274 | 300 | 0.2047 | - | - | - | - |
1.7699 | 400 | 0.1628 | - | - | - | - |
2.2124 | 500 | 0.1519 | 0.1204 | 0.7094 | 0.7271 | 0.7266 |
2.6549 | 600 | 0.1309 | - | - | - | - |
3.0973 | 700 | 0.1228 | - | - | - | - |
3.5398 | 800 | 0.1062 | - | - | - | - |
3.9823 | 900 | 0.097 | - | - | - | - |
4.4248 | 1000 | 0.0853 | 0.1026 | 0.7281 | 0.7409 | 0.7468 |
4.8673 | 1100 | 0.086 | - | - | - | - |
5.3097 | 1200 | 0.0723 | - | - | - | - |
5.7522 | 1300 | 0.0678 | - | - | - | - |
6.1947 | 1400 | 0.0655 | - | - | - | - |
6.6372 | 1500 | 0.0583 | 0.0970 | 0.7252 | 0.7479 | 0.7502 |
7.0796 | 1600 | 0.0586 | - | - | - | - |
7.5221 | 1700 | 0.0521 | - | - | - | - |
7.9646 | 1800 | 0.049 | - | - | - | - |
8.4071 | 1900 | 0.0437 | - | - | - | - |
8.8496 | 2000 | 0.0443 | 0.0974 | 0.7193 | 0.7427 | 0.7455 |
Framework Versions
- Python: 3.12.0
- Sentence Transformers: 3.0.1
- Transformers: 4.41.2
- PyTorch: 2.3.1+cu118
- Accelerate: 0.31.0
- Datasets: 2.20.0
- Tokenizers: 0.19.1
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
WeightedMultipleNegativesRankingLoss
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
- Downloads last month
- 9
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for federicovolponi/arctic-embed-m-space-sup
Base model
Snowflake/snowflake-arctic-embed-mEvaluation results
- Cosine Accuracy@5 on dim 768self-reported0.841
- Cosine Accuracy@10 on dim 768self-reported0.884
- Cosine Precision@5 on dim 768self-reported0.168
- Cosine Precision@10 on dim 768self-reported0.088
- Cosine Recall@5 on dim 768self-reported0.841
- Cosine Recall@10 on dim 768self-reported0.884
- Cosine Ndcg@5 on dim 768self-reported0.750
- Cosine Ndcg@10 on dim 768self-reported0.764
- Cosine Mrr@5 on dim 768self-reported0.719
- Cosine Mrr@10 on dim 768self-reported0.725