Edit model card

dev-ner-ontonote-bert-finetuned

This model is a fine-tuned version of bert-base-cased on the dev-ner-ontone dataset.

It achieves the following results on the evaluation set:

  • Loss: 0.0241
  • Precision: 0.9404
  • Recall: 0.9668
  • F1: 0.9535
  • Accuracy: 0.9937

Model description

NER Model to identify one or more of the following entities from any given sentence

['CARDINAL', 'DATE', 'EVENT', 'FAC', 'GPE', 'LANGUAGE', 'LAW', 'LOC', 'MONEY', 'NORP', 'ORDINAL', 'ORG', 'PERCENT', 'PERSON', 'PRODUCT', 'QUANTITY', 'TIME', 'WORK_OF_ART']

Intended uses & limitations

Can be used for named entity recognition for information extraction/retrieval. The limitation of this work is that since it uses BERT as it's pretrained backbone all the limitations and biases there get carried forward in this model.

Also, in the training data, we had only a very small number of examples of a few entities. To be precise this is the distribution of the data

{'GPE': 2268, 'PERSON': 2020, 'ORG': 1740, 'DATE': 1507, 'CARDINAL': 938, 'NORP': 847, 'MONEY': 274, 'ORDINAL': 232, 'TIME': 214, 'LOC': 204, 'PERCENT': 177, 'EVENT': 143, 'WORK_OF_ART': 142, 'FAC': 115, 'QUANTITY': 100, 'PRODUCT': 72, 'LAW': 40, 'LANGUAGE': 33}

Hence tokenwise, the model will underperform on entities related to language or law etc.

Training and evaluation data

Here is a snapshot of the dataset stats

DatasetDict({
    train: Dataset({
        features: ['text', 'entities', 'entities-suggestion', 'entities-suggestion-metadata', 'external_id', 'metadata'],
        num_rows: 8528
    })
    validation: Dataset({
        features: ['text', 'entities', 'entities-suggestion', 'entities-suggestion-metadata', 'external_id', 'metadata'],
        num_rows: 8528
    })
})

For more detailed information, refer this page

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Precision Recall F1 Accuracy
No log 1.0 267 0.1113 0.7576 0.7973 0.7769 0.9689
0.2811 2.0 534 0.0559 0.8732 0.9087 0.8906 0.9847
0.2811 3.0 801 0.0360 0.9147 0.9478 0.9309 0.9904
0.063 4.0 1068 0.0275 0.9333 0.9600 0.9465 0.9928
0.063 5.0 1335 0.0241 0.9404 0.9668 0.9535 0.9937

Framework versions

  • Transformers 4.27.4
  • Pytorch 2.3.1
  • Datasets 2.20.0
  • Tokenizers 0.13.3
Downloads last month
6
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ElisonSherton/dev-ner-ontonote-bert-finetuned

Finetuned
(1906)
this model

Dataset used to train ElisonSherton/dev-ner-ontonote-bert-finetuned