Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
xiaoou
/
am-4sentences
like
0
Token Classification
Transformers
Safetensors
roberta
Inference Endpoints
Model card
Files
Files and versions
Community
Train
Deploy
Use this model
main
am-4sentences
1 contributor
History:
3 commits
xiaoou
Upload tokenizer
2fefd9f
12 months ago
.gitattributes
1.52 kB
initial commit
12 months ago
config.json
1.01 kB
Upload RobertaForTokenClassification
12 months ago
merges.txt
456 kB
Upload tokenizer
12 months ago
model.safetensors
496 MB
LFS
Upload RobertaForTokenClassification
12 months ago
special_tokens_map.json
958 Bytes
Upload tokenizer
12 months ago
tokenizer.json
2.11 MB
Upload tokenizer
12 months ago
tokenizer_config.json
1.32 kB
Upload tokenizer
12 months ago
vocab.json
798 kB
Upload tokenizer
12 months ago