Edit model card

Model Card for Finetuned NepBertA-NER

This model is a fine-tuned version of the NepBERTa model, specifically trained for Named Entity Recognition (NER) tasks in the Nepali language. It recognizes entities such as persons (PER), organizations (ORG), and locations (LOC) in Nepali text. The model has been trained on a custom dataset and supports token classification for the following entity tags:

  • O (Other)
  • B-PER (Beginning of a person’s name)
  • I-PER (Inside of a person’s name)
  • B-ORG (Beginning of an organization)
  • I-ORG (Inside of an organization)
  • B-LOC (Beginning of a location)
  • I-LOC (Inside of a location)

Model Details

Model Description

  • Developed by: Priyanshu Koirala (Synapse Technologies)
  • Model type: Token Classification (NER)
  • Language(s) (NLP): Nepali
  • License: Apache 2.0
  • Finetuned from model: NepBERTa

Uses

Direct Use

The model can be directly used to recognize and classify named entities in Nepali text, such as persons, organizations, and locations. This is useful for text analysis tasks like extracting important information from Nepali documents, news articles, and customer feedback.

Downstream Use

The model can be further fine-tuned on other similar datasets or integrated into applications for Nepali language processing.

Out-of-Scope Use

The model may not perform well for texts outside the scope of its training data, such as texts with unseen entity types or non-Nepali language texts.

Bias, Risks, and Limitations

As with any NER model, there may be biases in the data that influence how the model identifies and classifies entities. It may struggle with unseen entities, domain-specific jargon, or ambiguous contexts.

Recommendations

Users should evaluate the model in their specific use case, ensuring that the data fed into the model aligns with the training data, and understand that the model might require further fine-tuning for specialized tasks.

How to Get Started with the Model

Use the following code to start using the model:

import torch
from transformers import AutoTokenizer, AutoModelForTokenClassification

device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')

# Load model and tokenizer
model = AutoModelForTokenClassification.from_pretrained("SynapseHQ/Finetuned-NER-NepBertA")
tokenizer = AutoTokenizer.from_pretrained("SynapseHQ/Finetuned-NER-NepBertA")
model.to(device)

def predict_ner_chunked(text, model, tokenizer, device, max_length=512):
    model.eval()
    words = text.split()
    ner_results = []
    
    for i in range(0, len(words), max_length):
        chunk = ' '.join(words[i:i+max_length])
        tokens = tokenizer(chunk, return_tensors="pt", truncation=True, padding=True, max_length=max_length)
        tokens = {k: v.to(device) for k, v in tokens.items()}
        
        with torch.no_grad():
            outputs = model(**tokens)
        
        predictions = torch.argmax(outputs.logits, dim=2)
        predicted_labels = [model.config.id2label[p.item()] for p in predictions[0]]
        
        chunk_words = tokenizer.convert_ids_to_tokens(tokens["input_ids"][0])
        for word, label in zip(chunk_words, predicted_labels):
            if label in ["B-PER", "I-PER", "B-ORG"] and word not in ["[CLS]", "[SEP]", "[PAD]"]:
                ner_results.append((word, label))
    
    return ner_results

# Test the model
text = "सङ्घीय लोकतान्त्रिक गणतन्त्र नेपालको प्रधानमन्त्री शेरबहादुर देउवा हुन्।"
ner_results = predict_ner_chunked(text, model, tokenizer, device)
print(ner_results)

Training Details

Training Data

The model was trained on a custom-labeled dataset in Nepali, consisting of sentences annotated with named entities for People (PER), Organizations (ORG), and Locations (LOC).

Training Procedure

  • Optimizer: AdamW
  • Learning Rate: 5e-5
  • Batch Size: 16
  • Epochs: 5
  • Validation Split: 20% of the dataset was reserved for validation.
  • Hardware: Trained on a single GPU.

Training Hyperparameters

  • Number of labels: 7 (including O label)
  • Maximum sequence length: 128 tokens
  • Gradient accumulation: 1

Evaluation

Metrics

The model was evaluated using the seqeval metric, with the following results on the validation set:

  • F1 Score: 0.89
  • Precision: 0.86
  • Recall: 0.90

Citation for the Base Model

If you use this model or the base model in your work, please consider citing NepBERTa as follows:

@inproceedings{timilsina2022nepberta,
  title={NepBERTa: Nepali language model trained in a large corpus},
  author={Timilsina, Sulav and Gautam, Milan and Bhattarai, Binod},
  booktitle={Proceedings of the 2nd conference of the Asia-pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing},
  year={2022},
  organization={Association for Computational Linguistics (ACL)}
}

Citation

If you use this model in your research, please consider citing it:

@misc{nepali_ner,
  author = {Synapse Technologies},
  title = {Finetuned NepBertA-NER for Nepali},
  year = {2024},
  publisher = {Hugging Face},
  howpublished = {\url{https://huggingface.co/SynapseHQ/Finetuned-NER-NepBertA}},
}
Downloads last month
17
Safetensors
Model size
109M params
Tensor type
F32
·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for SynapseHQ/Finetuned-NER-NepBertA

Base model

NepBERTa/NepBERTa
Finetuned
(2)
this model