Text Generation
Transformers
PyTorch
English
code
gpt_jx
text-generation-inference
custom_code
KnutJaegersberg commited on
Commit
1043c8a
1 Parent(s): 7f92913

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -0
README.md CHANGED
@@ -13,6 +13,12 @@ inference: false
13
  tags:
14
  - text-generation-inference
15
  ---
 
 
 
 
 
 
16
  ### **Model Description**
17
  ***GPT-JX*** is a **3 billion paramter** autoregressive Foundational Large Language Model pre-trained on *High Quality*, *Cleaned* and *Deduplicated* **1.1 trillion tokens** of english text and code. ***GPT-JX*** uses the base architecture of traditional *Transformers Decoder* with **slight changes** which is discussed later. ***GPT-JX*** was pre-trained on tokens for **English text** and **20 Programming Languages**. ***GPT-JX*** shows impressing performance when compared to **Large Language Models with 7 billion parameters** such as **LLaMa-7B-v2, Falcon-7B & MPT-7B**.
18
 
 
13
  tags:
14
  - text-generation-inference
15
  ---
16
+ Important Note:
17
+ I have not created this MIT licensed model, I discovered it and downloaded it. It was taken down by its creators, so I reupload it. More info:
18
+
19
+ https://github.com/huggingface/transformers/issues/25723
20
+
21
+
22
  ### **Model Description**
23
  ***GPT-JX*** is a **3 billion paramter** autoregressive Foundational Large Language Model pre-trained on *High Quality*, *Cleaned* and *Deduplicated* **1.1 trillion tokens** of english text and code. ***GPT-JX*** uses the base architecture of traditional *Transformers Decoder* with **slight changes** which is discussed later. ***GPT-JX*** was pre-trained on tokens for **English text** and **20 Programming Languages**. ***GPT-JX*** shows impressing performance when compared to **Large Language Models with 7 billion parameters** such as **LLaMa-7B-v2, Falcon-7B & MPT-7B**.
24