Add SetFit model
Browse files- 1_Pooling/config.json +7 -0
- README.md +243 -0
- config.json +26 -0
- config_sentence_transformers.json +7 -0
- config_setfit.json +18 -0
- model.safetensors +3 -0
- model_head.pkl +3 -0
- modules.json +14 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +64 -0
- vocab.txt +0 -0
1_Pooling/config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"word_embedding_dimension": 384,
|
3 |
+
"pooling_mode_cls_token": false,
|
4 |
+
"pooling_mode_mean_tokens": true,
|
5 |
+
"pooling_mode_max_tokens": false,
|
6 |
+
"pooling_mode_mean_sqrt_len_tokens": false
|
7 |
+
}
|
README.md
ADDED
@@ -0,0 +1,243 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: setfit
|
3 |
+
tags:
|
4 |
+
- setfit
|
5 |
+
- sentence-transformers
|
6 |
+
- text-classification
|
7 |
+
- generated_from_setfit_trainer
|
8 |
+
metrics:
|
9 |
+
- accuracy
|
10 |
+
widget:
|
11 |
+
- text: This paper focuses on mining association rules between sets of items in large
|
12 |
+
databases, which can reveal interesting patterns and relationships among the data.
|
13 |
+
- text: In this paper, the authors explore the economic concepts of fairness and retaliation
|
14 |
+
within the context of reciprocity, demonstrating how these principles shape market
|
15 |
+
behaviors and interactions.
|
16 |
+
- text: Further research is needed to explore the applicability of the proposed model
|
17 |
+
to more complex multi-echelon inventory systems with additional features, such
|
18 |
+
as lead time variability and supplier reliability.
|
19 |
+
- text: The NCEP/NCAR 40-Year Reanalysis Project provides retrospective atmospheric
|
20 |
+
data sets by assimilating observational data into a model, resulting in improved
|
21 |
+
estimates of historical weather patterns for meteorological research and applications.
|
22 |
+
- text: This study aims to assess the accuracy of aerosol optical properties retrieved
|
23 |
+
from Aerosol Robotic Network (AERONET) Sun and sky radiance measurements using
|
24 |
+
ground-based reference data.
|
25 |
+
pipeline_tag: text-classification
|
26 |
+
inference: true
|
27 |
+
base_model: sentence-transformers/paraphrase-MiniLM-L3-v2
|
28 |
+
model-index:
|
29 |
+
- name: SetFit with sentence-transformers/paraphrase-MiniLM-L3-v2
|
30 |
+
results:
|
31 |
+
- task:
|
32 |
+
type: text-classification
|
33 |
+
name: Text Classification
|
34 |
+
dataset:
|
35 |
+
name: Unknown
|
36 |
+
type: unknown
|
37 |
+
split: test
|
38 |
+
metrics:
|
39 |
+
- type: accuracy
|
40 |
+
value: 0.7407692307692307
|
41 |
+
name: Accuracy
|
42 |
+
---
|
43 |
+
|
44 |
+
# SetFit with sentence-transformers/paraphrase-MiniLM-L3-v2
|
45 |
+
|
46 |
+
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-MiniLM-L3-v2](https://huggingface.co/sentence-transformers/paraphrase-MiniLM-L3-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
|
47 |
+
|
48 |
+
The model has been trained using an efficient few-shot learning technique that involves:
|
49 |
+
|
50 |
+
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
|
51 |
+
2. Training a classification head with features from the fine-tuned Sentence Transformer.
|
52 |
+
|
53 |
+
## Model Details
|
54 |
+
|
55 |
+
### Model Description
|
56 |
+
- **Model Type:** SetFit
|
57 |
+
- **Sentence Transformer body:** [sentence-transformers/paraphrase-MiniLM-L3-v2](https://huggingface.co/sentence-transformers/paraphrase-MiniLM-L3-v2)
|
58 |
+
- **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
|
59 |
+
- **Maximum Sequence Length:** 128 tokens
|
60 |
+
- **Number of Classes:** 13 classes
|
61 |
+
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
|
62 |
+
<!-- - **Language:** Unknown -->
|
63 |
+
<!-- - **License:** Unknown -->
|
64 |
+
|
65 |
+
### Model Sources
|
66 |
+
|
67 |
+
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
|
68 |
+
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
|
69 |
+
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
|
70 |
+
|
71 |
+
### Model Labels
|
72 |
+
| Label | Examples |
|
73 |
+
|:----------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
74 |
+
| Aims | <ul><li>'This study aims to provide an in-depth analysis of the impact of Coronavirus Disease 2019 (COVID-19) on Italy, focusing on the early stages of the outbreak and the subsequent government response.'</li><li>'In this paper, we propose SegNet, a deep convolutional encoder-decoder architecture for real-time image segmentation.'</li><li>'This study aims to develop a mathematical model for analyzing genetic variation using restriction endonucleases.'</li></ul> |
|
75 |
+
| Background | <ul><li>'Previous studies have demonstrated that statins, including pravastatin, can reduce the risk of coronary events in patients with elevated cholesterol levels. However, the efficacy of pravastatin in patients with average cholesterol levels is less clear.'</li><li>'Previous studies have shown that statins, including pravastatin, can reduce the risk of coronary events in patients with elevated cholesterol levels. However, this study investigates the effect of pravastatin on patients with average cholesterol levels.'</li><li>'Previous studies have shown that statins, including pravastatin, can reduce the risk of coronary events in patients with elevated cholesterol levels. However, this trial investigates the effect of pravastatin on patients with average cholesterol levels.'</li></ul> |
|
76 |
+
| Hypothesis | <ul><li>'Despite having average cholesterol levels, patients who received Pravastatin experienced a significant reduction in coronary events, suggesting a potential role for statins in preventing cardiovascular events beyond cholesterol level management in internal medicine.'</li><li>'This prospective observational study aimed to investigate the association between glycaemia levels and the risk of developing macrovascular and microvascular complications in individuals with type 2 diabetes, as previously identified in the UKPDS 35 study.'</li><li>'The results suggest that self-regulatory skills, particularly in the area of attention, significantly impact academic performance in elementary school students.'</li></ul> |
|
77 |
+
| Implications | <ul><li>'From 1995 to 1998, the UK Prospective Diabetes Study (UKPDS) 35 observed a significant association between higher glycaemia levels and increased risk of both macrovascular and microvascular complications in patients with type 2 diabetes.'</li><li>'The UKPDS 35 study provides robust evidence that every 1 mmol/L increase in HbA1c is associated with a 25% increased risk of macrovascular events and a 37% increased risk of microvascular complications in patients with type 2 diabetes, highlighting the importance of strict glycaemic control in internal medicine.'</li><li>"This study provides valuable insights into the early dynamics of the COVID-19 outbreak in Italy, contributing to the understanding of the disease's transmission patterns and impact on public health."</li></ul> |
|
78 |
+
| Importance | <ul><li>'Stroke and transient ischemic attack (TIA) are leading causes of long-term disability and mortality in internal medicine, with an estimated 15 million survivors worldwide.'</li><li>'The accurate assessment of insulin resistance and beta-cell function is crucial in the diagnosis and management of various metabolic disorders, including type 2 diabetes and metabolic syndrome.'</li><li>'The COVID-19 outbreak in Italy, which began in late February 2020, quickly became one of the most severe epidemic hotspots in Europe.'</li></ul> |
|
79 |
+
| Keywords | <ul><li>'Pravastatin is a statin drug commonly used in the treatment of hypercholesterolemia, specifically to lower low-density lipoprotein (LDL) cholesterol levels and reduce the risk of cardiovascular events in internal medicine.'</li><li>'Self-regulation refers to the ability of students to manage their emotions, behavior, and cognitive processes to achieve optimal learning (Zimmerman & Kitsantas, 2005).'</li><li>'The proposed method utilizes deep convolutional neural networks to extract rich features from input images, enabling both object detection and semantic segmentation with high accuracy in the field of artificial intelligence.'</li></ul> |
|
80 |
+
| Limitations | <ul><li>'However, it is important to note that the Homeostasis Model Assessment (HOMA) index does not directly measure insulin sensitivity or β-cell function, but rather provides an estimate based on fasting plasma glucose and insulin concentrations.'</li><li>'Despite providing a useful estimate of insulin resistance and beta-cell function, the Homeostasis Model Assessment has limitations in its applicability to individuals with extreme glucose or insulin levels, as well as those with certain diseases such as liver disease or pregnancy.'</li><li>'Despite the large sample size and long follow-up period, the observational nature of the study limits the ability to establish causality between glycaemia and the observed complications in type 2 diabetes.'</li></ul> |
|
81 |
+
| Method | <ul><li>'The study employed a randomized, double-blind, placebo-controlled design to investigate the effect of Pravastatin on coronary events in patients with average cholesterol levels.'</li><li>'Patients with a history of myocardial infarction and an average cholesterol level between 180 and 240 mg/dL were included in the study.'</li><li>'The study aimed to assess the impact of Pravastatin administration on the incidence of coronary events in internal medicine patients with average cholesterol levels.'</li></ul> |
|
82 |
+
| None | <ul><li>'The study enrolled patients with a recent myocardial infarction and an average cholesterol level, who were then randomly assigned to receive either pravastatin or placebo.'</li><li>'This systematic review and meta-analysis aimed to assess the efficacy and safety of dual antiplatelet therapy with aspirin and clopidogrel in the secondary prevention of stroke and transient ischemic attack in the field of internal medicine.'</li><li>'This study aims to evaluate the effectiveness of the Homeostasis Model Assessment (HOMA) in estimating insulin resistance and pancreatic beta-cell function in internal medicine, offering valuable insights for the diagnosis and management of metabolic disorders.'</li></ul> |
|
83 |
+
| Purpose | <ul><li>'This study investigates the impact of Pravastatin on reducing coronary events in internal medicine patients with average cholesterol levels after a myocardial infarction.'</li><li>'This systematic review and meta-analysis aimed to assess the efficacy and safety of dual antiplatelet therapy with aspirin and clopidogrel in the secondary prevention of stroke and transient ischemic attack in internal medicine.'</li><li>'This study aims to evaluate the effectiveness of the Homeostasis Model Assessment (HOMA) in estimating insulin resistance and beta-cell function in internal medicine patients, addressing the need for a simple and widely applicable method for diagnosing and monitoring these conditions.'</li></ul> |
|
84 |
+
| Reccomendations | <ul><li>'Further studies are needed to investigate the optimal duration of dual antiplatelet therapy in secondary prevention of stroke and transient ischemic attack, as well as the role of individual patient characteristics in determining the most effective treatment regimen.'</li><li>'Further research is warranted to explore the underlying mechanisms linking glycaemia to macrovascular and microvascular complications in type 2 diabetes, particularly in multi-ethnic populations.'</li><li>'Further studies are needed to investigate the potential role of IL-6 signaling in the prevention of bone loss in postmenopausal women.'</li></ul> |
|
85 |
+
| Result | <ul><li>'Despite having average cholesterol levels, patients treated with Pravastatin did not experience a significant reduction in coronary events compared to the placebo group.'</li><li>'In interviews with patients who experienced a reduction in coronary events after Pravastatin treatment, themes included improved energy levels and increased confidence in managing their heart health.'</li><li>'The study found that Pravastatin significantly reduced the risk of coronary events in patients with average cholesterol levels, consistent with previous research suggesting that statins benefit a wider population beyond those with hypercholesterolemia.'</li></ul> |
|
86 |
+
| Uncertainty | <ul><li>'Despite the widespread use of pravastatin in post-myocardial infarction patients with average cholesterol levels, the evidence regarding its impact on coronary events remains inconclusive and sometimes contradictory.'</li><li>'Despite the findings of this study showing a reduction in coronary events with Pravastatin use in patients with average cholesterol levels, contrasting evidence exists suggesting no significant benefit in similar patient populations (Miller et al., 2018).'</li><li>'Despite the proven benefits of dual antiplatelet therapy with aspirin and clopidogrel in the secondary prevention of cardiovascular events, particularly in coronary artery disease, there is a paucity of data specifically addressing its use in stroke or transient ischemic attack (TIA) patients.'</li></ul> |
|
87 |
+
|
88 |
+
## Evaluation
|
89 |
+
|
90 |
+
### Metrics
|
91 |
+
| Label | Accuracy |
|
92 |
+
|:--------|:---------|
|
93 |
+
| **all** | 0.7408 |
|
94 |
+
|
95 |
+
## Uses
|
96 |
+
|
97 |
+
### Direct Use for Inference
|
98 |
+
|
99 |
+
First install the SetFit library:
|
100 |
+
|
101 |
+
```bash
|
102 |
+
pip install setfit
|
103 |
+
```
|
104 |
+
|
105 |
+
Then you can load this model and run inference.
|
106 |
+
|
107 |
+
```python
|
108 |
+
from setfit import SetFitModel
|
109 |
+
|
110 |
+
# Download from the 🤗 Hub
|
111 |
+
model = SetFitModel.from_pretrained("Corran/SciGenSetfit2")
|
112 |
+
# Run inference
|
113 |
+
preds = model("This paper focuses on mining association rules between sets of items in large databases, which can reveal interesting patterns and relationships among the data.")
|
114 |
+
```
|
115 |
+
|
116 |
+
<!--
|
117 |
+
### Downstream Use
|
118 |
+
|
119 |
+
*List how someone could finetune this model on their own dataset.*
|
120 |
+
-->
|
121 |
+
|
122 |
+
<!--
|
123 |
+
### Out-of-Scope Use
|
124 |
+
|
125 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
126 |
+
-->
|
127 |
+
|
128 |
+
<!--
|
129 |
+
## Bias, Risks and Limitations
|
130 |
+
|
131 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
132 |
+
-->
|
133 |
+
|
134 |
+
<!--
|
135 |
+
### Recommendations
|
136 |
+
|
137 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
138 |
+
-->
|
139 |
+
|
140 |
+
## Training Details
|
141 |
+
|
142 |
+
### Training Set Metrics
|
143 |
+
| Training set | Min | Median | Max |
|
144 |
+
|:-------------|:----|:--------|:----|
|
145 |
+
| Word count | 11 | 28.3123 | 71 |
|
146 |
+
|
147 |
+
| Label | Training Sample Count |
|
148 |
+
|:----------------|:----------------------|
|
149 |
+
| Aims | 200 |
|
150 |
+
| Background | 200 |
|
151 |
+
| Hypothesis | 200 |
|
152 |
+
| Implications | 200 |
|
153 |
+
| Importance | 200 |
|
154 |
+
| Keywords | 200 |
|
155 |
+
| Limitations | 200 |
|
156 |
+
| Method | 200 |
|
157 |
+
| None | 200 |
|
158 |
+
| Purpose | 200 |
|
159 |
+
| Reccomendations | 200 |
|
160 |
+
| Result | 200 |
|
161 |
+
| Uncertainty | 200 |
|
162 |
+
|
163 |
+
### Training Hyperparameters
|
164 |
+
- batch_size: (256, 256)
|
165 |
+
- num_epochs: (1, 1)
|
166 |
+
- max_steps: -1
|
167 |
+
- sampling_strategy: oversampling
|
168 |
+
- num_iterations: 40
|
169 |
+
- body_learning_rate: (2e-05, 1e-05)
|
170 |
+
- head_learning_rate: 0.01
|
171 |
+
- loss: CosineSimilarityLoss
|
172 |
+
- distance_metric: cosine_distance
|
173 |
+
- margin: 0.25
|
174 |
+
- end_to_end: False
|
175 |
+
- use_amp: False
|
176 |
+
- warmup_proportion: 0.1
|
177 |
+
- seed: 42
|
178 |
+
- eval_max_steps: -1
|
179 |
+
- load_best_model_at_end: False
|
180 |
+
|
181 |
+
### Training Results
|
182 |
+
| Epoch | Step | Training Loss | Validation Loss |
|
183 |
+
|:------:|:----:|:-------------:|:---------------:|
|
184 |
+
| 0.0012 | 1 | 0.4201 | - |
|
185 |
+
| 0.0615 | 50 | 0.2562 | - |
|
186 |
+
| 0.1230 | 100 | 0.2334 | - |
|
187 |
+
| 0.1845 | 150 | 0.1974 | - |
|
188 |
+
| 0.2460 | 200 | 0.195 | - |
|
189 |
+
| 0.3075 | 250 | 0.1768 | - |
|
190 |
+
| 0.3690 | 300 | 0.146 | - |
|
191 |
+
| 0.4305 | 350 | 0.1541 | - |
|
192 |
+
| 0.4920 | 400 | 0.1647 | - |
|
193 |
+
| 0.5535 | 450 | 0.154 | - |
|
194 |
+
| 0.6150 | 500 | 0.1568 | - |
|
195 |
+
| 0.6765 | 550 | 0.1494 | - |
|
196 |
+
| 0.7380 | 600 | 0.1554 | - |
|
197 |
+
| 0.7995 | 650 | 0.1456 | - |
|
198 |
+
| 0.8610 | 700 | 0.1527 | - |
|
199 |
+
| 0.9225 | 750 | 0.1488 | - |
|
200 |
+
| 0.9840 | 800 | 0.1312 | - |
|
201 |
+
|
202 |
+
### Framework Versions
|
203 |
+
- Python: 3.10.12
|
204 |
+
- SetFit: 1.0.3
|
205 |
+
- Sentence Transformers: 2.2.2
|
206 |
+
- Transformers: 4.36.2
|
207 |
+
- PyTorch: 2.1.0+cu121
|
208 |
+
- Datasets: 2.16.1
|
209 |
+
- Tokenizers: 0.15.0
|
210 |
+
|
211 |
+
## Citation
|
212 |
+
|
213 |
+
### BibTeX
|
214 |
+
```bibtex
|
215 |
+
@article{https://doi.org/10.48550/arxiv.2209.11055,
|
216 |
+
doi = {10.48550/ARXIV.2209.11055},
|
217 |
+
url = {https://arxiv.org/abs/2209.11055},
|
218 |
+
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
|
219 |
+
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
|
220 |
+
title = {Efficient Few-Shot Learning Without Prompts},
|
221 |
+
publisher = {arXiv},
|
222 |
+
year = {2022},
|
223 |
+
copyright = {Creative Commons Attribution 4.0 International}
|
224 |
+
}
|
225 |
+
```
|
226 |
+
|
227 |
+
<!--
|
228 |
+
## Glossary
|
229 |
+
|
230 |
+
*Clearly define terms in order to be accessible across audiences.*
|
231 |
+
-->
|
232 |
+
|
233 |
+
<!--
|
234 |
+
## Model Card Authors
|
235 |
+
|
236 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
237 |
+
-->
|
238 |
+
|
239 |
+
<!--
|
240 |
+
## Model Card Contact
|
241 |
+
|
242 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
243 |
+
-->
|
config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/root/.cache/torch/sentence_transformers/sentence-transformers_paraphrase-MiniLM-L3-v2/",
|
3 |
+
"architectures": [
|
4 |
+
"BertModel"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"classifier_dropout": null,
|
8 |
+
"gradient_checkpointing": false,
|
9 |
+
"hidden_act": "gelu",
|
10 |
+
"hidden_dropout_prob": 0.1,
|
11 |
+
"hidden_size": 384,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 1536,
|
14 |
+
"layer_norm_eps": 1e-12,
|
15 |
+
"max_position_embeddings": 512,
|
16 |
+
"model_type": "bert",
|
17 |
+
"num_attention_heads": 12,
|
18 |
+
"num_hidden_layers": 3,
|
19 |
+
"pad_token_id": 0,
|
20 |
+
"position_embedding_type": "absolute",
|
21 |
+
"torch_dtype": "float32",
|
22 |
+
"transformers_version": "4.36.2",
|
23 |
+
"type_vocab_size": 2,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 30522
|
26 |
+
}
|
config_sentence_transformers.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"__version__": {
|
3 |
+
"sentence_transformers": "2.0.0",
|
4 |
+
"transformers": "4.7.0",
|
5 |
+
"pytorch": "1.9.0+cu102"
|
6 |
+
}
|
7 |
+
}
|
config_setfit.json
ADDED
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"normalize_embeddings": false,
|
3 |
+
"labels": [
|
4 |
+
"Aims",
|
5 |
+
"Background",
|
6 |
+
"Hypothesis",
|
7 |
+
"Implications",
|
8 |
+
"Importance",
|
9 |
+
"Keywords",
|
10 |
+
"Limitations",
|
11 |
+
"Method",
|
12 |
+
"None",
|
13 |
+
"Purpose",
|
14 |
+
"Reccomendations",
|
15 |
+
"Result",
|
16 |
+
"Uncertainty"
|
17 |
+
]
|
18 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:782421e8a8f86650f5c4c24184bb8cde66eb095e4f2bce737ad3508d1c844bd8
|
3 |
+
size 69565312
|
model_head.pkl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6790c7fffe6c2ab476607806d7b8ab06f8b147b2dce5a6a6eba84ea624ba05b8
|
3 |
+
size 41647
|
modules.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"idx": 0,
|
4 |
+
"name": "0",
|
5 |
+
"path": "",
|
6 |
+
"type": "sentence_transformers.models.Transformer"
|
7 |
+
},
|
8 |
+
{
|
9 |
+
"idx": 1,
|
10 |
+
"name": "1",
|
11 |
+
"path": "1_Pooling",
|
12 |
+
"type": "sentence_transformers.models.Pooling"
|
13 |
+
}
|
14 |
+
]
|
sentence_bert_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"max_seq_length": 128,
|
3 |
+
"do_lower_case": false
|
4 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": {
|
3 |
+
"content": "[CLS]",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"mask_token": {
|
10 |
+
"content": "[MASK]",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "[PAD]",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"sep_token": {
|
24 |
+
"content": "[SEP]",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"unk_token": {
|
31 |
+
"content": "[UNK]",
|
32 |
+
"lstrip": false,
|
33 |
+
"normalized": false,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false
|
36 |
+
}
|
37 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,64 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": true,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_basic_tokenize": true,
|
47 |
+
"do_lower_case": true,
|
48 |
+
"mask_token": "[MASK]",
|
49 |
+
"max_length": 128,
|
50 |
+
"model_max_length": 512,
|
51 |
+
"never_split": null,
|
52 |
+
"pad_to_multiple_of": null,
|
53 |
+
"pad_token": "[PAD]",
|
54 |
+
"pad_token_type_id": 0,
|
55 |
+
"padding_side": "right",
|
56 |
+
"sep_token": "[SEP]",
|
57 |
+
"stride": 0,
|
58 |
+
"strip_accents": null,
|
59 |
+
"tokenize_chinese_chars": true,
|
60 |
+
"tokenizer_class": "BertTokenizer",
|
61 |
+
"truncation_side": "right",
|
62 |
+
"truncation_strategy": "longest_first",
|
63 |
+
"unk_token": "[UNK]"
|
64 |
+
}
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|