DFKI
/

Bulgarian
embeddings
glove
cc100
glove-bg-cc100 / README.md
DGurgurov's picture
Upload README.md with huggingface_hub
99fa830 verified
|
raw
history blame
1.78 kB
---
language: bg
tags:
- embeddings
- glove
- cc100
license: cc-by-sa-4.0
---
# CC100 GloVe Embeddings for BG Language
## Model Description
- **Language:** bg
- **Embedding Algorithm:** GloVe (Global Vectors for Word Representation)
- **Vocabulary Size:** 3917287
- **Vector Dimensions:** 300
- **Training Data:** CC100 dataset
## Training Information
We trained GloVe embeddings using the original C code. The model was trained by stochastically sampling nonzero elements from the co-occurrence matrix, over 100 iterations, to produce 300-dimensional vectors. We used a context window of ten words to the left and ten words to the right. Words with fewer than 5 co-occurrences were excluded for languages with over 1 million tokens in the training data, and the threshold was set to 2 for languages with smaller datasets.
We used data from CC100 for training the static word embeddings. We set xmax = 100, α = 3/4, and used AdaGrad optimization with an initial learning rate of 0.05.
## Usage
These embeddings can be used for various NLP tasks such as text classification, named entity recognition, and as input features for neural networks.
## Citation
If you use these embeddings in your research, please cite:
```bibtex
@misc{gurgurov2024lowremrepositorywordembeddings,
title={LowREm: A Repository of Word Embeddings for 87 Low-Resource Languages Enhanced with Multilingual Graph Knowledge},
author={Daniil Gurgurov and Rishu Kumar and Simon Ostermann},
year={2024},
eprint={2409.18193},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2409.18193},
}
```
## License
These embeddings are released under the [CC-BY-SA 4.0 License](https://creativecommons.org/licenses/by-sa/4.0/).