How to Use the Model for Inference:
You can use the Hugging Face pipeline
for easy inference:
from transformers import pipeline
# Load the model
model_path = "venkatd/NCBI_NER"
pipe = pipeline(
task="token-classification",
model=model_path,
tokenizer=model_path,
aggregation_strategy="simple"
)
# Test the pipeline
text = ("A 48-year-old female presented with vaginal bleeding and abnormal Pap smears. "
"Upon diagnosis of invasive non-keratinizing SCC of the cervix, she underwent a radical "
"hysterectomy with salpingo-oophorectomy which demonstrated positive spread to the pelvic "
"lymph nodes and the parametrium.")
result = pipe(text)
print(result)
Output Example:
The output will be entity type of Disease, score, and start/end positions in the text. Here’s a sample output format:
[
{
"entity_group": "Disease",
"score": 0.98,
"word": "SCC of the cervix",
"start": 121,
"end": 139
},
...
]
Model Summary and Training Details
Model Architecture
- Base Model:
microsoft/deberta-v3-base
- Task: Token Classification for Named Entity Recognition (NER) with a focus on disease entities.
- Number of Labels: 3 (O, B-Disease, I-Disease)
Dataset
- Dataset: NCBI Disease Corpus
- Description: The NCBI Disease corpus is a specialized medical dataset that includes 793 PubMed abstracts. It is structured to help in identifying disease mentions within scientific literature, and each mention is annotated with disease concepts from the MeSH (Medical Subject Headings) or OMIM (Online Mendelian Inheritance in Man) databases.
- Split:
- Training Set: 593 abstracts
- Development (Validation) Set: 100 abstracts
- Test Set: 100 abstracts
Training Details
- Training Steps: The model was trained using a cross-entropy loss function for token classification tasks. To optimize performance, we used gradient accumulation to achieve a stable loss and improve resource efficiency.
- Gradient Accumulation: 2 steps
- Batch Size: 8
- Device: Trained on a GPU if available, using mixed-precision training for better performance.
Optimizer and Learning Rate Scheduler
Optimizer: AdamW
- Learning Rate: 1e-5
- Betas: (0.9, 0.999)
- Epsilon: 1e-8
Learning Rate Scheduler: Cosine Scheduler with Warmup
- Warmup Steps: 10% of total training steps
- Total Training Steps: Calculated as
len(train_loader) * num_epochs
Epochs and Validation
- Epochs: 5
- Training and Validation Loss: The model achieved a stable loss over 5 epochs, with the best validation loss recorded. The best model based on validation loss was saved for evaluation.
Evaluation and Performance
- Test Dataset F1 Score: 0.9772
- Evaluation Metric: F1 score, which indicates the balance between precision and recall, was used as the primary metric to assess the model’s performance.
- Downloads last month
- 6
Model tree for venkatd/NCBI_NER
Base model
microsoft/deberta-v3-base