jacobrenn commited on
Commit
6f90c05
1 Parent(s): 6c45204

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -4
README.md CHANGED
@@ -13,12 +13,11 @@ library_name: transformers
13
  <!-- Provide a quick summary of what the model is/does. -->
14
 
15
  AI Squared's `dlite-v1-124m` ([blog post](https://medium.com/ai-squared/introducing-dlite-a-lightweight-chatgpt-like-model-based-on-dolly-deaa49402a1f)) is a large language
16
- model which is derived from OpenAI's smallest [GPT-2](https://huggingface.co/gpt2) model and fine-tuned on a single T4 GPU on a corpos of 50k records
17
  ([Stanford Alpaca](https://crfm.stanford.edu/2023/03/13/alpaca.html)) to help it exhibit chat-based capabilities.
18
 
19
  While `dlite-v1-124m` is **not a state-of-the-art model**, we believe that the level of interactivity that can be achieved on such a small model that is trained so cheaply
20
- is important to showcase, as it continues to demonstrate that creating powerful AI capabilities is much more accessible than previously thought.
21
- ## Model Details
22
 
23
 
24
  ### Model Description
@@ -45,11 +44,16 @@ Just as with any other LLM, we advise users of this technology to exercise good
45
 
46
  ## Usage
47
 
 
 
 
48
  ### Load Model and Tokenizer from this Repository Using the `transformers` Package
49
 
50
  ```python
51
- from transfomrers import AutoModelForCausalLM, AutoTokenizer
 
52
  model_id = 'aisquared/dlite-v1-124m'
 
53
  tokenizer = AutoTokenizer.from_pretrained(model_id)
54
  model = AutoModelForCausalLM.from_pretrained(model_id)
55
  ```
 
13
  <!-- Provide a quick summary of what the model is/does. -->
14
 
15
  AI Squared's `dlite-v1-124m` ([blog post](https://medium.com/ai-squared/introducing-dlite-a-lightweight-chatgpt-like-model-based-on-dolly-deaa49402a1f)) is a large language
16
+ model which is derived from OpenAI's smallest [GPT-2](https://huggingface.co/gpt2) model and fine-tuned on a single T4 GPU on a corpus of 50k records
17
  ([Stanford Alpaca](https://crfm.stanford.edu/2023/03/13/alpaca.html)) to help it exhibit chat-based capabilities.
18
 
19
  While `dlite-v1-124m` is **not a state-of-the-art model**, we believe that the level of interactivity that can be achieved on such a small model that is trained so cheaply
20
+ is important to showcase, as it continues to demonstrate that creating powerful AI capabilities may be much more accessible than previously thought.
 
21
 
22
 
23
  ### Model Description
 
44
 
45
  ## Usage
46
 
47
+ The code below shows how to use `dlite-v1-124m` in the way which it was trained. While the model can be used "out of the box" using the
48
+ `transformers` library, using the function defined below to create a response from the model will achieve better results.
49
+
50
  ### Load Model and Tokenizer from this Repository Using the `transformers` Package
51
 
52
  ```python
53
+ from transformers import AutoModelForCausalLM, AutoTokenizer
54
+
55
  model_id = 'aisquared/dlite-v1-124m'
56
+
57
  tokenizer = AutoTokenizer.from_pretrained(model_id)
58
  model = AutoModelForCausalLM.from_pretrained(model_id)
59
  ```