BERT base Japanese (character-level tokenization with whole word masking, jawiki-20200831)
This pretrained model is almost the same as cl-tohoku/bert-base-japanese-char-v2 but do not need fugashi
or unidic_lite
.
The only difference is in word_tokenzer_type
property (specify basic
instead of mecab
) in tokenizer_config.json
.
- Downloads last month
- 53
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.