andreas122001
commited on
Commit
•
3c17cc3
1
Parent(s):
9f87925
Update README.md
Browse files
README.md
CHANGED
@@ -1,18 +1,75 @@
|
|
1 |
---
|
2 |
license: openrail
|
3 |
widget:
|
4 |
-
- text:
|
5 |
example_title: default
|
6 |
-
- text:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
example_title: real wikipedia
|
8 |
-
- text:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
example_title: generated wikipedia
|
10 |
-
- text:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
example_title: from ChatGPT
|
12 |
-
- text:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
example_title: GPT-3 paper
|
14 |
datasets:
|
15 |
- NicolaiSivesind/human-vs-machine
|
|
|
16 |
language:
|
17 |
- en
|
18 |
pipeline_tag: text-classification
|
@@ -20,4 +77,116 @@ tags:
|
|
20 |
- mgt-detection
|
21 |
- ai-detection
|
22 |
---
|
23 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: openrail
|
3 |
widget:
|
4 |
+
- text: I am totally a human, trust me bro.
|
5 |
example_title: default
|
6 |
+
- text: >-
|
7 |
+
In Finnish folklore, all places and things, and also human beings, have a
|
8 |
+
haltija (a genius, guardian spirit) of their own. One such haltija is called
|
9 |
+
etiäinen—an image, doppelgänger, or just an impression that goes ahead of a
|
10 |
+
person, doing things the person in question later does. For example, people
|
11 |
+
waiting at home might hear the door close or even see a shadow or a
|
12 |
+
silhouette, only to realize that no one has yet arrived. Etiäinen can also
|
13 |
+
refer to some kind of a feeling that something is going to happen. Sometimes
|
14 |
+
it could, for example, warn of a bad year coming. In modern Finnish, the
|
15 |
+
term has detached from its shamanistic origins and refers to premonition.
|
16 |
+
Unlike clairvoyance, divination, and similar practices, etiäiset (plural)
|
17 |
+
are spontaneous and can't be induced. Quite the opposite, they may be
|
18 |
+
unwanted and cause anxiety, like ghosts. Etiäiset need not be too dramatic
|
19 |
+
and may concern everyday events, although ones related to e.g. deaths are
|
20 |
+
common. As these phenomena are still reported today, they can be considered
|
21 |
+
a living tradition, as a way to explain the psychological experience of
|
22 |
+
premonition.
|
23 |
example_title: real wikipedia
|
24 |
+
- text: >-
|
25 |
+
In Finnish folklore, all places and things, animate or inanimate, have a
|
26 |
+
spirit or "etiäinen" that lives there. Etiäinen can manifest in many forms,
|
27 |
+
but is usually described as a kind, elderly woman with white hair. She is
|
28 |
+
the guardian of natural places and often helps people in need. Etiäinen has
|
29 |
+
been a part of Finnish culture for centuries and is still widely believed in
|
30 |
+
today. Folklorists study etiäinen to understand Finnish traditions and how
|
31 |
+
they have changed over time.
|
32 |
example_title: generated wikipedia
|
33 |
+
- text: >-
|
34 |
+
This paper presents a novel framework for sparsity-certifying graph
|
35 |
+
decompositions, which are important tools in various areas of computer
|
36 |
+
science, including algorithm design, complexity theory, and optimization.
|
37 |
+
Our approach is based on the concept of "cut sparsifiers," which are sparse
|
38 |
+
graphs that preserve the cut structure of the original graph up to a certain
|
39 |
+
error bound. We show that cut sparsifiers can be efficiently constructed
|
40 |
+
using a combination of spectral techniques and random sampling, and we use
|
41 |
+
them to develop new algorithms for decomposing graphs into sparse subgraphs.
|
42 |
example_title: from ChatGPT
|
43 |
+
- text: >-
|
44 |
+
Recent work has demonstrated substantial gains on many NLP tasks and
|
45 |
+
benchmarks by pre-training on a large corpus of text followed by fine-tuning
|
46 |
+
on a specific task. While typically task-agnostic in architecture, this
|
47 |
+
method still requires task-specific fine-tuning datasets of thousands or
|
48 |
+
tens of thousands of examples. By contrast, humans can generally perform a
|
49 |
+
new language task from only a few examples or from simple instructions -
|
50 |
+
something which current NLP systems still largely struggle to do. Here we
|
51 |
+
show that scaling up language models greatly improves task-agnostic,
|
52 |
+
few-shot performance, sometimes even reaching competitiveness with prior
|
53 |
+
state-of-the-art fine-tuning approaches. Specifically, we train GPT-3, an
|
54 |
+
autoregressive language model with 175 billion parameters, 10x more than any
|
55 |
+
previous non-sparse language model, and test its performance in the few-shot
|
56 |
+
setting. For all tasks, GPT-3 is applied without any gradient updates or
|
57 |
+
fine-tuning, with tasks and few-shot demonstrations specified purely via
|
58 |
+
text interaction with the model. GPT-3 achieves strong performance on many
|
59 |
+
NLP datasets, including translation, question-answering, and cloze tasks, as
|
60 |
+
well as several tasks that require on-the-fly reasoning or domain
|
61 |
+
adaptation, such as unscrambling words, using a novel word in a sentence, or
|
62 |
+
performing 3-digit arithmetic. At the same time, we also identify some
|
63 |
+
datasets where GPT-3's few-shot learning still struggles, as well as some
|
64 |
+
datasets where GPT-3 faces methodological issues related to training on
|
65 |
+
large web corpora. Finally, we find that GPT-3 can generate samples of news
|
66 |
+
articles which human evaluators have difficulty distinguishing from articles
|
67 |
+
written by humans. We discuss broader societal impacts of this finding and
|
68 |
+
of GPT-3 in general.
|
69 |
example_title: GPT-3 paper
|
70 |
datasets:
|
71 |
- NicolaiSivesind/human-vs-machine
|
72 |
+
- gfissore/arxiv-abstracts-2021
|
73 |
language:
|
74 |
- en
|
75 |
pipeline_tag: text-classification
|
|
|
77 |
- mgt-detection
|
78 |
- ai-detection
|
79 |
---
|
80 |
+
|
81 |
+
Machine-generated text-detection by fine-tuning of language models
|
82 |
+
===
|
83 |
+
|
84 |
+
This project is related to a bachelor's thesis with the title "*Turning Poachers into Gamekeepers: Detecting Machine-Generated Text in Academia using Large Language Models*" (not yet published) written by *Nicolai Thorer Sivesind* and *Andreas Bentzen Winje* at the *Department of Computer Science* at the *Norwegian University of Science and Technology*.
|
85 |
+
|
86 |
+
It contains text classification models trained to distinguish human-written text from text generated by language models like ChatGPT and GPT-3. The best models were able to achieve an accuracy of 100% on real and *GPT-3*-generated wikipedia articles (4500 samples), and an accuracy of 98.4% on real and *ChatGPT*-generated research abstracts (3000 samples).
|
87 |
+
|
88 |
+
The dataset card for the dataset that was created in relation to this project can be found [here](https://huggingface.co/datasets/NicolaiSivesind/human-vs-machine).
|
89 |
+
|
90 |
+
**NOTE**: the hosted inference on this site only works for the RoBERTa-models, and not for the Bloomz-models. The Bloomz-models otherwise can produce wrong predictions when not explicitly providing the attention mask from the tokenizer to the model for inference. To be sure, the [pipeline](https://huggingface.co/docs/transformers/main_classes/pipelines)-library seems to produce the most consistent results.
|
91 |
+
|
92 |
+
|
93 |
+
## Fine-tuned detectors
|
94 |
+
|
95 |
+
This project includes 12 fine-tuned models based on the RoBERTa-base model, and three sizes of the bloomz-models.
|
96 |
+
|
97 |
+
| Base-model | RoBERTa-base | Bloomz-560m | Bloomz-1b7 | Bloomz-3b |
|
98 |
+
|------------|--------------------------------------------------------------------------------|--------------------------------------------------------------------------------------------|------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------|
|
99 |
+
| Wiki | [roberta-wiki](https://huggingface.co/andreas122001/roberta-academic-detector) | [Bloomz-560m-wiki](https://huggingface.co/andreas122001/bloomz-560m-wiki-detector) | [Bloomz-1b7-wiki](https://huggingface.co/andreas122001/bloomz-1b7-wiki-detector) | [Bloomz-3b-wiki](https://huggingface.co/andreas122001/bloomz-3b-wiki-detector) |
|
100 |
+
| Academic | [roberta-academic](https://huggingface.co/andreas122001/roberta-wiki-detector) | [Bloomz-560m-academic](https://huggingface.co/andreas122001/bloomz-560m-academic-detector) | [Bloomz-1b7-academic](https://huggingface.co/andreas122001/bloomz-1b7-academic-detector) | [Bloomz-3b-academic](https://huggingface.co/andreas122001/bloomz-3b-academic-detector) |
|
101 |
+
| Mixed | [roberta-mixed](https://huggingface.co/andreas122001/roberta-mixed-detector) | [Bloomz-560m-mixed](https://huggingface.co/andreas122001/bloomz-560m-mixed-detector) | [Bloomz-1b7-mixed](https://huggingface.co/andreas122001/bloomz-1b7-mixed-detector) | [Bloomz-3b-mixed](https://huggingface.co/andreas122001/bloomz-3b-mixed-detector) |
|
102 |
+
|
103 |
+
|
104 |
+
### Datasets
|
105 |
+
|
106 |
+
The models were trained on selections from the [GPT-wiki-intros]() and [ChatGPT-Research-Abstracts](), and are separated into three types, **wiki**-detectors, **academic**-detectors and **mixed**-detectors, respectively.
|
107 |
+
|
108 |
+
- **Wiki-detectors**:
|
109 |
+
- Trained on 30'000 datapoints (10%) of GPT-wiki-intros.
|
110 |
+
- Best model (in-domain) is Bloomz-3b-wiki, with an accuracy of 100%.
|
111 |
+
- **Academic-detectors**:
|
112 |
+
- Trained on 20'000 datapoints (100%) of ChatGPT-Research-Abstracts.
|
113 |
+
- Best model (in-domain) is Bloomz-3b-academic, with an accuracy of 98.4%
|
114 |
+
- **Mixed-detectors**:
|
115 |
+
- Trained on 15'000 datapoints (5%) of GPT-wiki-intros and 10'000 datapoints (50%) of ChatGPT-Research-Abstracts.
|
116 |
+
- Best model (in-domain) is RoBERTa-mixed, with an F1-score of 99.3%.
|
117 |
+
|
118 |
+
|
119 |
+
### Hyperparameters
|
120 |
+
|
121 |
+
All models were trained using the same hyperparameters:
|
122 |
+
|
123 |
+
```python
|
124 |
+
{
|
125 |
+
"num_train_epochs": 1,
|
126 |
+
"adam_beta1": 0.9,
|
127 |
+
"adam_beta2": 0.999,
|
128 |
+
"batch_size": 8,
|
129 |
+
"adam_epsilon": 1e-08
|
130 |
+
"optim": "adamw_torch" # the optimizer (AdamW)
|
131 |
+
"learning_rate": 5e-05, # (LR)
|
132 |
+
"lr_scheduler_type": "linear", # scheduler type for LR
|
133 |
+
"seed": 42, # seed for PyTorch RNG-generator.
|
134 |
+
}
|
135 |
+
```
|
136 |
+
|
137 |
+
### Metrics
|
138 |
+
|
139 |
+
Metrics can be found at https://wandb.ai/idatt2900-072/IDATT2900-072.
|
140 |
+
|
141 |
+
|
142 |
+
In-domain performance of wiki-detectors:
|
143 |
+
|
144 |
+
| Base model | Accuracy | Precision | Recall | F1-score |
|
145 |
+
|-------------|----------|-----------|--------|----------|
|
146 |
+
| Bloomz-560m | 0.973 | *1.000 | 0.945 | 0.972 |
|
147 |
+
| Bloomz-1b7 | 0.972 | *1.000 | 0.945 | 0.972 |
|
148 |
+
| Bloomz-3b | *1.000 | *1.000 | *1.000 | *1.000 |
|
149 |
+
| RoBERTa | 0.998 | 0.999 | 0.997 | 0.998 |
|
150 |
+
|
151 |
+
|
152 |
+
In-domain peformance of academic-detectors:
|
153 |
+
|
154 |
+
| Base model | Accuracy | Precision | Recall | F1-score |
|
155 |
+
|-------------|----------|-----------|--------|----------|
|
156 |
+
| Bloomz-560m | 0.964 | 0.963 | 0.965 | 0.964 |
|
157 |
+
| Bloomz-1b7 | 0.946 | 0.941 | 0.951 | 0.946 |
|
158 |
+
| Bloomz-3b | *0.984 | *0.983 | 0.985 | *0.984 |
|
159 |
+
| RoBERTa | 0.982 | 0.968 | *0.997 | 0.982 |
|
160 |
+
|
161 |
+
|
162 |
+
F1-scores of the mixed-detectors on all three datasets:
|
163 |
+
|
164 |
+
| Base model | Mixed | Wiki | CRA |
|
165 |
+
|-------------|--------|--------|--------|
|
166 |
+
| Bloomz-560m | 0.948 | 0.972 | *0.848 |
|
167 |
+
| Bloomz-1b7 | 0.929 | 0.964 | 0.816 |
|
168 |
+
| Bloomz-3b | 0.988 | 0.996 | 0.772 |
|
169 |
+
| RoBERTa | *0.993 | *0.997 | 0.829 |
|
170 |
+
|
171 |
+
|
172 |
+
## Credits
|
173 |
+
|
174 |
+
- [GPT-wiki-intro](https://huggingface.co/datasets/aadityaubhat/GPT-wiki-intro), by Aaditya Bhat
|
175 |
+
- [arxiv-abstracts-2021](https://huggingface.co/datasets/gfissore/arxiv-abstracts-2021), by Giancarlo
|
176 |
+
- [Bloomz](bigscience/bloomz), by BigScience
|
177 |
+
- [RoBERTa](https://huggingface.co/roberta-base), by Liu et. al.
|
178 |
+
|
179 |
+
|
180 |
+
## Citation
|
181 |
+
|
182 |
+
Please use the following citation:
|
183 |
+
|
184 |
+
```
|
185 |
+
@misc {sivesind_2023,
|
186 |
+
author = { {Nicolai Thorer Sivesind} and {Andreas Bentzen Winje} },
|
187 |
+
title = { Machine-generated text-detection by fine-tuning of language models },
|
188 |
+
url = { https://huggingface.co/andreas122001/roberta-academic-detector }
|
189 |
+
year = 2023,
|
190 |
+
publisher = { Hugging Face }
|
191 |
+
}
|
192 |
+
```
|