Edit model card

Text Equivalence Classification

This model is a fine-tuned version of distilroberta-base specifically for the task of determining whether two sentences are equivalent in meaning. It was trained on a curated dataset combining multiple sources to cover a wide range of topics and styles.

Model Description

The model aims to capture the nuanced relationships between sentences, leveraging the powerful pre-trained DistilRoBERTa base for understanding contextual meanings. It has been fine-tuned with a specific focus on identifying semantic equivalence, making it suitable for applications such as paraphrase identification, duplicate question detection, and semantic search tasks.

Intended Uses & Limitations

This model is intended for use in applications requiring the determination of semantic equivalence between pairs of sentences. It can be utilized in creating more efficient search algorithms, enhancing chatbot understandings, and improving content moderation by identifying duplicate or similar content.

However, the model may exhibit limitations when dealing with sentences that contain highly domain-specific jargon, rare languages, or nuanced cultural references. It is recommended to test the model with your specific data and use case to ensure its effectiveness.

Training and Evaluation Data

The model was trained on a dataset named "DatasetX," which is a composite of several public datasets including portions of the Quora Question Pairs and the Microsoft Research Paraphrase Corpus. This combination was chosen to create a robust model capable of understanding a wide variety of sentence structures and meanings.

Training Procedure

Training Hyperparameters

The following hyperparameters were used during training:

  • Learning rate: 5e-05
  • Train batch size: 8
  • Eval batch size: 8
  • Seed: 42
  • Optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • LR scheduler type: linear
  • Number of epochs: 3

Training Results

Training Loss Epoch Step Validation Loss Accuracy F1
0.5396 1.09 500 0.5511 0.8309 0.8848
0.3722 2.18 1000 0.4973 0.8529 0.8944

Framework Versions

  • Transformers 4.37.2
  • Pytorch 2.2.0
  • Datasets 2.17.0
  • Tokenizers 0.15.2
Downloads last month
8
Safetensors
Model size
82.1M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for IgnaCodeIA/text-Classification

Finetuned
(522)
this model