The follwing information is gathered by Chris Ha for research purposes. No guarantee of correctness implied. | |
For specific information, consult original developers or refer to the Original license and `original_README.md`. | |
## Architecture | |
* likely sentencepiece trained tokenizer transfered imported into huggingface tokenizers. If so, original sentencepiece model not provided. | |
## Algorithm | |
* BPE | |
* UOV handled with byte_fallback infered by the absence of bytelevel() and spm origin | |
## Pretokenizers | |
* Whitespace : (Can be relevant for code or compression.) | |
* Metaspace() with ("▁" U+2581 LOWER ONE EIGHTH BLOCK) as meta symbol | |
* prefix applied (i.e. both "the" and "▁the" present) | |
* Whitespace reservation : None (no multiple whitespace nor metasymbols) | |
* Punctuation : (code & math) | |
* Split and Isolated | |
* Numbers : (code & math, compression ratios) | |
* Split and Isolated(invidiual_digits=true) | |
## Vocabulary | |
* Size : 100257 | |
* Around 96567 are single width character unicode characters (WIP) | |
* Average Length : WIP | |
* Median Length : 1 char | |
* Languages | |
* "[multiple languages including](https://huggingface.co/xverse/XVERSE-13B/discussions/5) ar, bg, ca, cs, da, de, el, en, es, et, fa, fi, fr, hi, hu, id, it, iw, ja, kk, ko, lt, lv, mr, ms, nl, no, pl, pt, ro, ru, sk, sl, sr, sv, ta, th, tr, uk, vi, and zh" | |
* Most multi-char tokens are english or chinese | |