metadata
base_model: sentence-transformers/all-mpnet-base-v2
datasets: []
language: []
library_name: sentence-transformers
pipeline_tag: sentence-similarity
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- generated_from_trainer
- dataset_size:178829
- loss:MultipleNegativesRankingLoss
widget:
- source_sentence: who was actor larry parks
sentences:
- >-
American stage and movie actor.e eventually did so in tears, only to be
blacklisted anyway.
- >-
A possum (plural form: possums) is any of about 70 small-to medium-sized
arboreal marsupial species native to Australia, New Guinea, and Sulawesi
(and introduced to New Zealand and China). The common brushtail possum
was introduced to New Zealand by European settlers in an attempt to
establish a fur industry. There are no native predators of the possum in
New Zealand, so its numbers in New Zealand have risen to the point where
it is considered a serious pest.
- >-
A document used to change one or more minor provisions of a living trust
or joint living trust as an alternative to preparing a new living trust.
- source_sentence: what is the salary of a person with a biology degree
sentences:
- $10 to $25 per hour.
- $25,290 (2014-2015 academic year)
- >-
Biology majors who don’t attend a graduate program make a median salary
of $51,000 per year, which is a little below the median salary for
graduates from all other majors combined. Don’t let that fact stop you
from pursuing a degree in biology if it’s what you’re passionate about,
though. Career Options for Biology Majors. Below is a list of common
career options for biology majors. This isn’t a comprehensive list, as
students who major in biology go on to do many interesting things.
However, this list should give you an idea of the types of work that
would be available to you with a degree in biology.
- source_sentence: definition of pretext
sentences:
- >-
Peanut butter is an excellent source of nutrition. Required to contain
at least 90 percent peanuts, it includes more than 30 vitamins and
minerals. Peanut butter contains no cholesterol or trans fats, according
to the National Peanut Board. In fact, studies show that peanut butter
may even improve your levels of good cholesterol.
- >-
Pretext generally refers to a reason for an action which is false, and
offered to cover up true motives or intentions. It is a concept
sometimes brought up in the context of employment discrimination.
- 20.5 degrees Celsius (68.8 degrees Fahrenheit).
- source_sentence: what is cyber spoofing
sentences:
- >-
Once your question has been posted for at least 1 hour and has at least
one answer, click on 'Award Best Answer' button next to your chosen
answer. 1 Upload failed. 2 Please upload a file larger than 100x100
pixels. 3 We are experiencing some problems, please try again.
- >-
Though some vegetable sources of protein contain sufficient values of
all essential amino acids, many are lower in one or more essential amino
acids than animal sources, especially lysine, and to a lesser extent
methionine and threonine. 1 Proteins derived from plant foods (legumes,
seeds, grains, and vegetables) can be complete as well (examples include
chickpeas, black beans, pumpkin seeds, cashews, cauliflower, quinoa,
pistachios, turnip greens, black-eyed peas, and soy). 2 Most plant
foods tend to have less of one or more essential amino acid
- >-
A spoofing attack is a situation in which one person or program
successfully masquerades as another by falsifying data and thereby
gaining an illegitimate advantage.
- source_sentence: what type of reaction is iron plus oxygen
sentences:
- Pearl
- 'Yes'
- >-
When a metal undergos a combination reaction with oxygen, a metal oxide
is formed (similarily, a metal halide is formed if reacted with one of
the halogens). You see the products of this type of reaction whenever
you see rust. Rust is the product of a combination reaction of iron and
oxygen:
SentenceTransformer based on sentence-transformers/all-mpnet-base-v2
This is a sentence-transformers model finetuned from sentence-transformers/all-mpnet-base-v2. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
Model Details
Model Description
- Model Type: Sentence Transformer
- Base model: sentence-transformers/all-mpnet-base-v2
- Maximum Sequence Length: 384 tokens
- Output Dimensionality: 768 tokens
- Similarity Function: Cosine Similarity
Model Sources
- Documentation: Sentence Transformers Documentation
- Repository: Sentence Transformers on GitHub
- Hugging Face: Sentence Transformers on Hugging Face
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 384, 'do_lower_case': False}) with Transformer model: MPNetModel
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Normalize()
)
Usage
Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("DashReza7/all-mpnet-base-v2_FINETUNED")
# Run inference
sentences = [
'what type of reaction is iron plus oxygen',
'When a metal undergos a combination reaction with oxygen, a metal oxide is formed (similarily, a metal halide is formed if reacted with one of the halogens). You see the products of this type of reaction whenever you see rust. Rust is the product of a combination reaction of iron and oxygen: ',
'Pearl',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 768]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
Training Details
Training Dataset
Unnamed Dataset
- Size: 178,829 training samples
- Columns:
anchor
andpositive
- Approximate statistics based on the first 1000 samples:
anchor positive type string string details - min: 5 tokens
- mean: 9.37 tokens
- max: 22 tokens
- min: 3 tokens
- mean: 60.48 tokens
- max: 197 tokens
- Samples:
anchor positive what is rba
Results-Based Accountability is a disciplined way of thinking and taking action that communities can use to improve the lives of children, youth, families, adults and the community as a whole.
what is rba
Results-Based Accountability® (also known as RBA) is a disciplined way of thinking and taking action that communities can use to improve the lives of children, youth, families, adults and the community as a whole. RBA is also used by organizations to improve the performance of their programs. Creating Community Impact with RBA. Community impact focuses on conditions of well-being for children, families and the community as a whole that a group of leaders is working collectively to improve. For example: “Residents with good jobs,” “Children ready for school,” or “A safe and clean neighborhood”.
was ronald reagan a democrat
Yes
- Loss:
MultipleNegativesRankingLoss
with these parameters:{ "scale": 20.0, "similarity_fct": "cos_sim" }
Training Hyperparameters
Non-Default Hyperparameters
per_device_train_batch_size
: 64per_device_eval_batch_size
: 64learning_rate
: 2e-05num_train_epochs
: 1warmup_ratio
: 0.1fp16
: True
All Hyperparameters
Click to expand
overwrite_output_dir
: Falsedo_predict
: Falseeval_strategy
: noprediction_loss_only
: Trueper_device_train_batch_size
: 64per_device_eval_batch_size
: 64per_gpu_train_batch_size
: Noneper_gpu_eval_batch_size
: Nonegradient_accumulation_steps
: 1eval_accumulation_steps
: Nonelearning_rate
: 2e-05weight_decay
: 0.0adam_beta1
: 0.9adam_beta2
: 0.999adam_epsilon
: 1e-08max_grad_norm
: 1.0num_train_epochs
: 1max_steps
: -1lr_scheduler_type
: linearlr_scheduler_kwargs
: {}warmup_ratio
: 0.1warmup_steps
: 0log_level
: passivelog_level_replica
: warninglog_on_each_node
: Truelogging_nan_inf_filter
: Truesave_safetensors
: Truesave_on_each_node
: Falsesave_only_model
: Falserestore_callback_states_from_checkpoint
: Falseno_cuda
: Falseuse_cpu
: Falseuse_mps_device
: Falseseed
: 42data_seed
: Nonejit_mode_eval
: Falseuse_ipex
: Falsebf16
: Falsefp16
: Truefp16_opt_level
: O1half_precision_backend
: autobf16_full_eval
: Falsefp16_full_eval
: Falsetf32
: Nonelocal_rank
: 0ddp_backend
: Nonetpu_num_cores
: Nonetpu_metrics_debug
: Falsedebug
: []dataloader_drop_last
: Falsedataloader_num_workers
: 0dataloader_prefetch_factor
: Nonepast_index
: -1disable_tqdm
: Falseremove_unused_columns
: Truelabel_names
: Noneload_best_model_at_end
: Falseignore_data_skip
: Falsefsdp
: []fsdp_min_num_params
: 0fsdp_config
: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}fsdp_transformer_layer_cls_to_wrap
: Noneaccelerator_config
: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}deepspeed
: Nonelabel_smoothing_factor
: 0.0optim
: adamw_torchoptim_args
: Noneadafactor
: Falsegroup_by_length
: Falselength_column_name
: lengthddp_find_unused_parameters
: Noneddp_bucket_cap_mb
: Noneddp_broadcast_buffers
: Falsedataloader_pin_memory
: Truedataloader_persistent_workers
: Falseskip_memory_metrics
: Trueuse_legacy_prediction_loop
: Falsepush_to_hub
: Falseresume_from_checkpoint
: Nonehub_model_id
: Nonehub_strategy
: every_savehub_private_repo
: Falsehub_always_push
: Falsegradient_checkpointing
: Falsegradient_checkpointing_kwargs
: Noneinclude_inputs_for_metrics
: Falseeval_do_concat_batches
: Truefp16_backend
: autopush_to_hub_model_id
: Nonepush_to_hub_organization
: Nonemp_parameters
:auto_find_batch_size
: Falsefull_determinism
: Falsetorchdynamo
: Noneray_scope
: lastddp_timeout
: 1800torch_compile
: Falsetorch_compile_backend
: Nonetorch_compile_mode
: Nonedispatch_batches
: Nonesplit_batches
: Noneinclude_tokens_per_second
: Falseinclude_num_input_tokens_seen
: Falseneftune_noise_alpha
: Noneoptim_target_modules
: Nonebatch_eval_metrics
: Falseeval_on_start
: Falsebatch_sampler
: batch_samplermulti_dataset_batch_sampler
: proportional
Training Logs
Epoch | Step | Training Loss |
---|---|---|
0.1789 | 500 | 0.279 |
0.3578 | 1000 | 0.2194 |
0.5367 | 1500 | 0.21 |
0.7156 | 2000 | 0.207 |
0.8945 | 2500 | 0.198 |
Framework Versions
- Python: 3.10.12
- Sentence Transformers: 3.0.1
- Transformers: 4.42.4
- PyTorch: 2.3.1+cu121
- Accelerate: 0.32.1
- Datasets: 2.21.0
- Tokenizers: 0.19.1
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
MultipleNegativesRankingLoss
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}