|
--- |
|
datasets: |
|
- WizardLM/WizardLM_evol_instruct_V2_196k |
|
- Open-Orca/OpenOrca |
|
language: |
|
- en |
|
tags: |
|
- chat |
|
- palmyra |
|
license: apache-2.0 |
|
--- |
|
|
|
|
|
**DEPRECATED MODEL NOTICE** |
|
========================== |
|
|
|
Please note that this model is no longer maintained or supported by our team. We strongly advise against using it in production or for any critical applications. |
|
|
|
Instead, we recommend using our latest and greatest models, which can be found at: |
|
|
|
https://huggingface.co/collections/Writer/palmyra-writer-license-66476fa8156169f8720a2c89 |
|
|
|
========================== |
|
|
|
|
|
|
|
|
|
# Writer/palmyra-20b-chat |
|
--- |
|
|
|
# Usage |
|
|
|
```py |
|
|
|
import torch |
|
from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer |
|
|
|
model_name = "Writer/palmyra-20b-chat" |
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
model_name, |
|
torch_dtype=torch.float16, |
|
device_map="auto", |
|
) |
|
|
|
prompt = "What is the meaning of life?" |
|
|
|
input_text = ( |
|
"A chat between a curious user and an artificial intelligence assistant. " |
|
"The assistant gives helpful, detailed, and polite answers to the user's questions. " |
|
"USER: {prompt} " |
|
"ASSISTANT:" |
|
) |
|
|
|
model_inputs = tokenizer(input_text.format(prompt=prompt), return_tensors="pt").to( |
|
"cuda" |
|
) |
|
|
|
gen_conf = { |
|
"top_k": 20, |
|
"max_new_tokens": 2048, |
|
"temperature": 0.6, |
|
"do_sample": True, |
|
"eos_token_id": tokenizer.eos_token_id, |
|
} |
|
|
|
streamer = TextStreamer(tokenizer) |
|
if "token_type_ids" in model_inputs: |
|
del model_inputs["token_type_ids"] |
|
|
|
all_inputs = {**model_inputs, **gen_conf} |
|
output = model.generate(**all_inputs, streamer=streamer) |
|
|
|
print("-"*20) |
|
print(output) |
|
|
|
``` |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Writer__palmyra-20b-chat) |
|
|
|
| Metric | Value | |
|
|-----------------------|---------------------------| |
|
| Avg. | 38.97 | |
|
| ARC (25-shot) | 43.52 | |
|
| HellaSwag (10-shot) | 72.83 | |
|
| MMLU (5-shot) | 35.18 | |
|
| TruthfulQA (0-shot) | 43.17 | |
|
| Winogrande (5-shot) | 66.46 | |
|
| GSM8K (5-shot) | 3.94 | |
|
| DROP (3-shot) | 7.7 | |