|
--- |
|
library_name: transformers |
|
tags: |
|
- natural-language-inference |
|
- nli |
|
license: mit |
|
datasets: |
|
- nyu-mll/multi_nli |
|
language: |
|
- en |
|
base_model: microsoft/deberta-v3-base |
|
--- |
|
|
|
# deberta v3 base - Natural Language Inference |
|
|
|
## Model overview |
|
This model is trained for the Natural Language Inference task. It takes two sentences as input (a premise and a hypothesis) and predicts the relationship between them by assigning one of three labels: "entailment," "neutral," or "contradiction." The model is based on the [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) model, fine-tuned on the [nyu-mll/multi_nli](https://huggingface.co/datasets/nyu-mll/multi_nli) dataset, and returns scores corresponding to the labels. |
|
|
|
## Results |
|
After fine-tuning on the dataset, the model achieved the following results: |
|
|
|
- Loss: 0.276 |
|
- Accuracy: 0.899 |
|
- F1-Score: 0.899 |
|
|
|
These metrics were evaluated on the `validation_mismatched` split of the dataset. |
|
|
|
## Usage |
|
```python |
|
from transformers import AutoTokenizer, AutoModelForSequenceClassification |
|
import torch |
|
|
|
model_name = "chincyk/deberta-v3-base-nli" |
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
model = AutoModelForSequenceClassification.from_pretrained(model_name) |
|
|
|
premise = "The flight arrived on time at the airport." |
|
hypothesis = "The flight was delayed by several hours." |
|
inputs = tokenizer(premise, hypothesis, return_tensors='pt') |
|
|
|
with torch.no_grad(): |
|
outputs = model(**inputs) |
|
|
|
logits = outputs.logits |
|
probs = torch.softmax(logits, dim=-1).squeeze() |
|
|
|
id2label = model.config.id2label |
|
|
|
for i, prob in enumerate(probs): |
|
print(f"{id2label[i]}: {prob:.4f}") |
|
``` |