Update README.md
Browse files
README.md
CHANGED
@@ -4,15 +4,17 @@ tags:
|
|
4 |
- sentiment
|
5 |
- twitter
|
6 |
- reviews
|
|
|
7 |
---
|
8 |
|
|
|
9 |
|
10 |
# Overview
|
11 |
-
This model is a fine-tuned checkpoint of [RoBERTa-large](https://huggingface.co/roberta-large) ([Liu et al. 2019](https://arxiv.org/pdf/1907.11692.pdf)). It enables reliable binary sentiment analysis for various types of English-language text. For each instance, it predicts either positive (1) or negative (0) sentiment. The model was fine-tuned and evaluated on 15 data sets from diverse text sources to enhance generalization across different types of texts (reviews, tweets, etc.). Consequently, it outperforms models trained on only one type of text (e.g., movie reviews from the popular SST-2 benchmark) when used on new data as shown below.
|
12 |
|
13 |
|
14 |
# Predictions on a data set
|
15 |
-
If you want to predict sentiment for your own data, we provide an example script via [Google Colab](https://colab.research.google.com/notebooks/intro.ipynb). You can load your data to a Google Drive and run the script for free on a Colab GPU. Set-up only takes a few minutes. We suggest that you manually label a subset of your data to evaluate performance for your use case. For performance benchmark values across various sentiment analysis contexts, please refer to our paper ([
|
16 |
|
17 |
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/chrsiebert/sentiment-roberta-large-english/blob/main/sentiment_roberta_prediction_example.ipynb)
|
18 |
|
@@ -63,13 +65,13 @@ To evaluate the performance of our general-purpose sentiment analysis model, we
|
|
63 |
Other values were left at their defaults as listed [here](https://huggingface.co/transformers/main_classes/trainer.html#transformers.TrainingArguments).
|
64 |
|
65 |
# Citation and contact
|
66 |
-
Please cite [this paper](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3489963) when you use our model. Feel free to reach out to [[email protected]](mailto:[email protected]) with any questions or feedback you may have.
|
67 |
|
68 |
```
|
69 |
-
@article{
|
70 |
-
title={More than a feeling:
|
71 |
-
author={
|
72 |
-
journal={
|
73 |
-
year={
|
74 |
}
|
75 |
```
|
|
|
4 |
- sentiment
|
5 |
- twitter
|
6 |
- reviews
|
7 |
+
- siebert
|
8 |
---
|
9 |
|
10 |
+
## SiEBERT - English-Language Sentiment Classification
|
11 |
|
12 |
# Overview
|
13 |
+
This model ("SiEBERT", prefix for "Sentiment in English") is a fine-tuned checkpoint of [RoBERTa-large](https://huggingface.co/roberta-large) ([Liu et al. 2019](https://arxiv.org/pdf/1907.11692.pdf)). It enables reliable binary sentiment analysis for various types of English-language text. For each instance, it predicts either positive (1) or negative (0) sentiment. The model was fine-tuned and evaluated on 15 data sets from diverse text sources to enhance generalization across different types of texts (reviews, tweets, etc.). Consequently, it outperforms models trained on only one type of text (e.g., movie reviews from the popular SST-2 benchmark) when used on new data as shown below.
|
14 |
|
15 |
|
16 |
# Predictions on a data set
|
17 |
+
If you want to predict sentiment for your own data, we provide an example script via [Google Colab](https://colab.research.google.com/notebooks/intro.ipynb). You can load your data to a Google Drive and run the script for free on a Colab GPU. Set-up only takes a few minutes. We suggest that you manually label a subset of your data to evaluate performance for your use case. For performance benchmark values across various sentiment analysis contexts, please refer to our paper ([Hartmann et al. 2022](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3489963)).
|
18 |
|
19 |
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/chrsiebert/sentiment-roberta-large-english/blob/main/sentiment_roberta_prediction_example.ipynb)
|
20 |
|
|
|
65 |
Other values were left at their defaults as listed [here](https://huggingface.co/transformers/main_classes/trainer.html#transformers.TrainingArguments).
|
66 |
|
67 |
# Citation and contact
|
68 |
+
Please cite [this paper](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3489963) (Forthcoming in the [IJRM](https://www.journals.elsevier.com/international-journal-of-research-in-marketing)) when you use our model. Feel free to reach out to [[email protected]](mailto:[email protected]) with any questions or feedback you may have.
|
69 |
|
70 |
```
|
71 |
+
@article{hartmann2022,
|
72 |
+
title={More than a feeling: Accuracy and Application of Sentiment Analysis},
|
73 |
+
author={Hartmann, Jochen and Heitmann, Mark and Siebert, Christian and Schamp, Christina},
|
74 |
+
journal={International Journal of Research in Marketing (Forthcoming)},
|
75 |
+
year={2022}
|
76 |
}
|
77 |
```
|