YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
This model provides a MobileBERT (Sun et al., 2020) fine-tuned on the SST data with three sentiments (0 -- negative, 1 -- neutral, and 2 -- positive).
Example Usage
Below, we provide illustrations on how to use this model to make sentiment predictions.
import torch
from transformers import AutoTokenizer, AutoConfig, MobileBertForSequenceClassification
# load model
model_name = r'cambridgeltl/sst_mobilebert-uncased'
tokenizer = AutoTokenizer.from_pretrained(model_name)
config = AutoConfig.from_pretrained(model_name)
model = MobileBertForSequenceClassification.from_pretrained(model_name, config=config)
model.eval()
'''
labels:
0 -- negative
1 -- neutral
2 -- positive
'''
# prepare exemplar sentences
batch_sentences = [
"in his first stab at the form , jacquot takes a slightly anarchic approach that works only sporadically .",
"a valueless kiddie paean to pro basketball underwritten by the nba .",
"a very well-made , funny and entertaining picture .",
]
# prepare input
inputs = tokenizer(batch_sentences, max_length=256, truncation=True, padding=True, return_tensors='pt')
input_ids, attention_mask = inputs.input_ids, inputs.attention_mask
# make predictions
outputs = model(input_ids=input_ids, attention_mask=attention_mask)
predictions = torch.argmax(outputs.logits, dim = -1)
print (predictions)
# tensor([1, 0, 2])
Citation:
If you find this model useful, please kindly cite our model as
@misc{susstmobilebert,
author = {Su, Yixuan},
title = {A MobileBERT Fine-tuned on SST},
howpublished = {\url{https://huggingface.co/cambridgeltl/sst_mobilebert-uncased}},
year = 2022
}
- Downloads last month
- 63
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.