File size: 1,752 Bytes
b324b19
 
 
 
 
 
 
 
 
 
 
 
79de40a
 
 
 
 
8a0b276
 
 
79de40a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0776384
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
---
license: mit
datasets:
- agicorp/python_code_instructions_18k_alpaca
language:
- en
base_model:
- openai-community/gpt2
pipeline_tag: text-generation
library_name: transformers
tags:
- code
---
# Model Card

GPT2Coder is a language model that uses openAI's GPT2 model architecture,
the model was pre-trained on multiple code data focused on python and languages
​​​​such as Spanish and English.

- It is a pre-trained model in a medium amount of code, so it is not recommended to use it like this, but it is functional and serves uses such as fine tuning and other tasks.

## Model Details


- **Developed by:** BueormAI
- **Shared by:** BueormLLC
- **Model type:** Transformer
- **Language(s) (NLP):** English (en), Spanish (es)
- **License:** MiT
- **Finetuned from model:** GPT2 Architecture

## Bias, Risks, and Limitations

The model can generate unexpected code and output, in addition to offensive texts and non-functional code.


### Recommendations

We recommend using the model with caution and handling its outputs with discretion as they may turn out to be non-functional outputs and harmful and dangerous code.

## Training Details

### Training Hyperparameters

- **Training regime:** fp16 mixed precision
- **Max_lenght:** 1024 tokens
- **pretrain epochs:** 1 epochs
- **finetuning epochs:** 2 epochs

## Environmental Impact

- **Hardware Type:** GPU P100
- **Hours used:** 18 hours
- **Cloud Provider:** Kaggle


# By Bueorm
Thanks to all the people who download and support our projects 
and manage a vision towards the future with AI, we hope you will support
us to continue advancing and launching more followed models.

- [Paypal Donations](https://paypal.me/bueorm)
- [Patreon Subscription](https://patreon.com/bueorm)