File size: 4,492 Bytes
b96c873 1782759 7e59583 b96c873 1782759 7d02523 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 1782759 7e59583 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 |
---
dataset_info:
features:
- name: Tokens
sequence: string
- name: Tags
sequence: int64
- name: Tags_string
sequence: string
- name: Original_source
dtype: string
splits:
- name: train
num_bytes: 276428315
num_examples: 471343
- name: test
num_bytes: 6419858
num_examples: 11136
- name: validation
num_bytes: 6345480
num_examples: 11456
download_size: 54821843
dataset_size: 289193653
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: validation
path: data/validation-*
task_categories:
- token-classification
language:
- es
size_categories:
- 100K<n<1M
license: apache-2.0
---
# Dataset Card for es-ner-massive
## Dataset Details
### Dataset Description
The es-ner-massive dataset is a combination of three datasets: tner/wikineural, conll2002, and polyglot_ner. It is designed for Named Entity Recognition (NER) tasks. Tags are curated to be span-based and encoded according to the following convention:
```python
encodings_dictionary = {
"O": 0,
"PER": 1,
'ORG': 2,
"LOC": 3,
"MISC": 4
}
```
## Dataset Details
### Dataset Description
The dataset was desing with the idea of combining middle size NER datasets in Spanish in order to perfom basic NER or to make Transfer Learning Operations with a solid knowledge base on the pretrained model.
- **Curated by:**
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
- **Language(s) (NLP):** [Spanish]
- **License:** [More Information Needed]
### Dataset Sources [optional]
Here the original sources:
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
## Uses
The intended use is to perform fine-tune of your pretrainned model into NER task.
## Dataset Structure
[More Information Needed]
## Dataset Creation
### Curation Rationale
refer to the original datasets of the compilation
[More Information Needed]
### Source Data
#### Data Collection and Processing
refer to the original datasets of the compilation
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
#### Who are the source data producers?
refer to the original datasets of the compilation
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
### Annotations [optional]
All the original NER tags that were in a BIO schema were passed to Span Schema
#### Annotation process
refer to the original datasets of the compilation
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
#### Who are the annotators?
refer to the original datasets of the compilation
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
#### Personal and Sensitive Information
refer to the original datasets of the compilation
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
## Bias, Risks, and Limitations
refer to the original datasets of the compilation
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
### Recommendations
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
refer to the original datasets of the compilation
[polyglot_ner](https://huggingface.co/datasets/polyglot_ner)
[conll2002](https://huggingface.co/datasets/conll2002)
[tner/wikineural](https://huggingface.co/datasets/tner/wikineural)
## Dataset Card Contact
You can email the author of this compilation at data_analitics_[email protected] |