KnutJaegersberg
commited on
Commit
•
1043c8a
1
Parent(s):
7f92913
Update README.md
Browse files
README.md
CHANGED
@@ -13,6 +13,12 @@ inference: false
|
|
13 |
tags:
|
14 |
- text-generation-inference
|
15 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
### **Model Description**
|
17 |
***GPT-JX*** is a **3 billion paramter** autoregressive Foundational Large Language Model pre-trained on *High Quality*, *Cleaned* and *Deduplicated* **1.1 trillion tokens** of english text and code. ***GPT-JX*** uses the base architecture of traditional *Transformers Decoder* with **slight changes** which is discussed later. ***GPT-JX*** was pre-trained on tokens for **English text** and **20 Programming Languages**. ***GPT-JX*** shows impressing performance when compared to **Large Language Models with 7 billion parameters** such as **LLaMa-7B-v2, Falcon-7B & MPT-7B**.
|
18 |
|
|
|
13 |
tags:
|
14 |
- text-generation-inference
|
15 |
---
|
16 |
+
Important Note:
|
17 |
+
I have not created this MIT licensed model, I discovered it and downloaded it. It was taken down by its creators, so I reupload it. More info:
|
18 |
+
|
19 |
+
https://github.com/huggingface/transformers/issues/25723
|
20 |
+
|
21 |
+
|
22 |
### **Model Description**
|
23 |
***GPT-JX*** is a **3 billion paramter** autoregressive Foundational Large Language Model pre-trained on *High Quality*, *Cleaned* and *Deduplicated* **1.1 trillion tokens** of english text and code. ***GPT-JX*** uses the base architecture of traditional *Transformers Decoder* with **slight changes** which is discussed later. ***GPT-JX*** was pre-trained on tokens for **English text** and **20 Programming Languages**. ***GPT-JX*** shows impressing performance when compared to **Large Language Models with 7 billion parameters** such as **LLaMa-7B-v2, Falcon-7B & MPT-7B**.
|
24 |
|