Edit model card

Instruct_Llama70B_Dolly15k

Fine-tuned from Llama-2-70B,used Dolly15k for the dataset. 80% for training, 15% validation, 5% test. Trained for 1.5 epochs using QLora. Trained with 1024 context window.

Model Details

  • Trained by: trained by Brillibits. See YouTube as well.
  • Model type: Instruct_Llama70B_Dolly15k is an auto-regressive language model based on the Llama 2 transformer architecture.
  • Language(s): English
  • License for Instruct_Llama70B_Dolly15ks: llama2 license

Prompting

Prompt Template With Context

Write a 10-line poem about a given topic

Input:

The topic is about racecars

Output:

Prompt Template Without Context

Who was the was the second president of the United States?

Output:

Professional Assistance

This model and other models like it are great, but where LLMs hold the most promise is when they are applied on custom data to automate a wide variety of tasks

If you have a dataset and want to see if you might be able to apply that data to automate some tasks, and you are looking for professional assistance, contact me here

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 60.97
ARC (25-shot) 68.34
HellaSwag (10-shot) 87.21
MMLU (5-shot) 69.52
TruthfulQA (0-shot) 46.46
Winogrande (5-shot) 84.29
GSM8K (5-shot) 42.68
DROP (3-shot) 28.26
Downloads last month
1,493
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train Brillibits/Instruct_Llama70B_Dolly15k