|
--- |
|
library_name: transformers |
|
base_model: EleutherAI/pythia-14m |
|
tags: |
|
- generated_from_trainer |
|
model-index: |
|
- name: pythia-14m-finewebedu-sentences |
|
results: [] |
|
--- |
|
|
|
# pythia-14m-finewebedu-sentences |
|
|
|
- Generate half intelligible English sentences using a small GPT like model. |
|
- Will output one sentence at a time. |
|
|
|
This model is a fine-tuned version of [EleutherAI/pythia-14m](https://huggingface.co/EleutherAI/pythia-14m) on the [agentlans/finewebedu-sentences dataset](https://huggingface.co/datasets/agentlans/finewebedu-sentences). |
|
|
|
## Model description |
|
|
|
To generate 10 random sentences starting from an empty string on a CUDA device: |
|
|
|
```python |
|
from transformers import pipeline, set_seed |
|
|
|
generator = pipeline('text-generation', model='agentlans/pythia-14m-finewebedu-sentences', device='cuda') |
|
|
|
set_seed(1234) |
|
results = generator("", max_length=100, num_return_sequences=10, do_sample=True) |
|
|
|
for x in results: |
|
print(x['generated_text']) |
|
``` |
|
|
|
Output: |
|
```text |
|
The main outcome is that the group is associated with other domains. |
|
If you're planning to get a long-term answer, you can check the link "watch" and see that you can change your website. |
|
They are very difficult to make it easy to understand how it works as a healthy. |
|
In the most cases, the prevalence of DTP is reduced from 5-HT1. |
|
It is a significant difference between the risk of injury, and there is no need to be a single complication. |
|
The time of taking too high on the scale of the region is to begin, with a bit other type of view to the whole system. |
|
The total cost of your daily distribution is $24. The overall number of children is 0.5 times is 50. |
|
The more difficult time is to learn the basics of the work, but it is important to do the same job. |
|
It is now on the other hand, however, in the middle of the 19th century, the country must follow the law of the country and alliances between the countries and communities. |
|
This is why it is hard to do this. |
|
``` |
|
|
|
## Intended uses & limitations |
|
|
|
- For generating short lines of English text |
|
- Could be useful for |
|
- data augmentation |
|
- creative inspiration |
|
- entertainment |
|
- CAPTCHA |
|
- Can be further finetuned on other data such as: |
|
- prompts |
|
- famous quotes |
|
- news headlines |
|
- blog post titles |
|
|
|
Limitations include: |
|
|
|
- Not guaranteed to make sensible, coherent, or grammatically correct sentences |
|
- No regard for accuracy or truthfulness whatsoever |
|
- It's a bunch of words from a probability model, what do you expect? |
|
|
|
## Training and evaluation data |
|
|
|
Sentences from [HuggingFaceFW/fineweb-edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) |
|
|
|
## Training procedure |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- learning_rate: 5e-05 |
|
- train_batch_size: 8 |
|
- eval_batch_size: 8 |
|
- seed: 42 |
|
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 |
|
- lr_scheduler_type: linear |
|
- num_epochs: 15.0 |
|
|
|
### Framework versions |
|
|
|
- Transformers 4.44.2 |
|
- Pytorch 2.2.2+cu121 |
|
- Datasets 2.18.0 |
|
- Tokenizers 0.19.1 |
|
|