Update app.py
Browse files
app.py
CHANGED
@@ -6,7 +6,7 @@ model = "PY007/TinyLlama-1.1B-Chat-v0.1"
|
|
6 |
tokenizer = transformers.LlamaTokenizer.from_pretrained(model)
|
7 |
pipeline = transformers.pipeline("text-generation",model=model,torch_dtype=torch.bfloat16)
|
8 |
|
9 |
-
prompt = "What is
|
10 |
formatted_prompt = (
|
11 |
f"### Human: {prompt}### Assistant:"
|
12 |
)
|
@@ -15,4 +15,4 @@ formatted_prompt = (
|
|
15 |
sequences = pipeline(formatted_prompt,do_sample=True,top_k=50,top_p = 0.7,num_return_sequences=1,repetition_penalty=1.1,max_new_tokens=500)
|
16 |
for seq in sequences:
|
17 |
st.write(f"Result: {seq['generated_text']}")
|
18 |
-
|
|
|
6 |
tokenizer = transformers.LlamaTokenizer.from_pretrained(model)
|
7 |
pipeline = transformers.pipeline("text-generation",model=model,torch_dtype=torch.bfloat16)
|
8 |
|
9 |
+
prompt = "What is 653+2343?"
|
10 |
formatted_prompt = (
|
11 |
f"### Human: {prompt}### Assistant:"
|
12 |
)
|
|
|
15 |
sequences = pipeline(formatted_prompt,do_sample=True,top_k=50,top_p = 0.7,num_return_sequences=1,repetition_penalty=1.1,max_new_tokens=500)
|
16 |
for seq in sequences:
|
17 |
st.write(f"Result: {seq['generated_text']}")
|
18 |
+
|