Files changed (1) hide show
  1. README.md +130 -14
README.md CHANGED
@@ -1,25 +1,128 @@
1
  ---
2
- extra_gated_heading: Access Llama 2 on Hugging Face
3
- extra_gated_description: >-
4
- This is a form to enable access to Llama 2 on Hugging Face after you have been
5
- granted access from Meta. Please visit the [Meta
6
- website](https://ai.meta.com/resources/models-and-libraries/llama-downloads)
7
- and accept our license terms and acceptable use policy before submitting this
8
- form. Requests will be processed in 1-2 days.
9
- extra_gated_button_content: Submit
10
- extra_gated_fields:
11
- I agree to share my name, email address and username with Meta and confirm that I have already been granted download access on the Meta website: checkbox
12
  language:
13
  - en
14
- pipeline_tag: text-generation
15
- inference: false
16
  tags:
17
  - facebook
18
  - meta
19
  - pytorch
20
  - llama
21
  - llama-2
22
- license: other
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  ---
24
  # Overview
25
  This model has been pruned to 10% sparsity using the [Wanda pruning method](https://arxiv.org/abs/2306.11695). This method requires no retraining or weight updates and still achieves competitive performance. A link to the base model can be found [here](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf).
@@ -137,4 +240,17 @@ Please report any software “bug,” or other problems with the models through
137
  |---|---|---|---|---|
138
  |7B| [Link](https://huggingface.co/llamaste/Llama-2-7b) | [Link](https://huggingface.co/llamaste/Llama-2-7b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-7b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-7b-chat-hf)|
139
  |13B| [Link](https://huggingface.co/llamaste/Llama-2-13b) | [Link](https://huggingface.co/llamaste/Llama-2-13b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-13b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-13b-hf)|
140
- |70B| [Link](https://huggingface.co/llamaste/Llama-2-70b) | [Link](https://huggingface.co/llamaste/Llama-2-70b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-70b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-70b-hf)|
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
 
 
 
 
 
 
 
 
 
2
  language:
3
  - en
4
+ license: other
 
5
  tags:
6
  - facebook
7
  - meta
8
  - pytorch
9
  - llama
10
  - llama-2
11
+ extra_gated_heading: Access Llama 2 on Hugging Face
12
+ extra_gated_description: This is a form to enable access to Llama 2 on Hugging Face
13
+ after you have been granted access from Meta. Please visit the [Meta website](https://ai.meta.com/resources/models-and-libraries/llama-downloads)
14
+ and accept our license terms and acceptable use policy before submitting this form.
15
+ Requests will be processed in 1-2 days.
16
+ extra_gated_button_content: Submit
17
+ extra_gated_fields:
18
+ ? I agree to share my name, email address and username with Meta and confirm that
19
+ I have already been granted download access on the Meta website
20
+ : checkbox
21
+ pipeline_tag: text-generation
22
+ inference: false
23
+ model-index:
24
+ - name: Llama-2-7b-chat-hf-10-sparsity
25
+ results:
26
+ - task:
27
+ type: text-generation
28
+ name: Text Generation
29
+ dataset:
30
+ name: AI2 Reasoning Challenge (25-Shot)
31
+ type: ai2_arc
32
+ config: ARC-Challenge
33
+ split: test
34
+ args:
35
+ num_few_shot: 25
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 53.16
39
+ name: normalized accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wang7776/Llama-2-7b-chat-hf-10-sparsity
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: HellaSwag (10-Shot)
48
+ type: hellaswag
49
+ split: validation
50
+ args:
51
+ num_few_shot: 10
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 78.26
55
+ name: normalized accuracy
56
+ source:
57
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wang7776/Llama-2-7b-chat-hf-10-sparsity
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: MMLU (5-Shot)
64
+ type: cais/mmlu
65
+ config: all
66
+ split: test
67
+ args:
68
+ num_few_shot: 5
69
+ metrics:
70
+ - type: acc
71
+ value: 48.18
72
+ name: accuracy
73
+ source:
74
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wang7776/Llama-2-7b-chat-hf-10-sparsity
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: TruthfulQA (0-shot)
81
+ type: truthful_qa
82
+ config: multiple_choice
83
+ split: validation
84
+ args:
85
+ num_few_shot: 0
86
+ metrics:
87
+ - type: mc2
88
+ value: 45.29
89
+ source:
90
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wang7776/Llama-2-7b-chat-hf-10-sparsity
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: Winogrande (5-shot)
97
+ type: winogrande
98
+ config: winogrande_xl
99
+ split: validation
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 71.59
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wang7776/Llama-2-7b-chat-hf-10-sparsity
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: GSM8k (5-shot)
114
+ type: gsm8k
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 18.42
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wang7776/Llama-2-7b-chat-hf-10-sparsity
125
+ name: Open LLM Leaderboard
126
  ---
127
  # Overview
128
  This model has been pruned to 10% sparsity using the [Wanda pruning method](https://arxiv.org/abs/2306.11695). This method requires no retraining or weight updates and still achieves competitive performance. A link to the base model can be found [here](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf).
 
240
  |---|---|---|---|---|
241
  |7B| [Link](https://huggingface.co/llamaste/Llama-2-7b) | [Link](https://huggingface.co/llamaste/Llama-2-7b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-7b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-7b-chat-hf)|
242
  |13B| [Link](https://huggingface.co/llamaste/Llama-2-13b) | [Link](https://huggingface.co/llamaste/Llama-2-13b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-13b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-13b-hf)|
243
+ |70B| [Link](https://huggingface.co/llamaste/Llama-2-70b) | [Link](https://huggingface.co/llamaste/Llama-2-70b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-70b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-70b-hf)|
244
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
245
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_wang7776__Llama-2-7b-chat-hf-10-sparsity)
246
+
247
+ | Metric |Value|
248
+ |---------------------------------|----:|
249
+ |Avg. |52.48|
250
+ |AI2 Reasoning Challenge (25-Shot)|53.16|
251
+ |HellaSwag (10-Shot) |78.26|
252
+ |MMLU (5-Shot) |48.18|
253
+ |TruthfulQA (0-shot) |45.29|
254
+ |Winogrande (5-shot) |71.59|
255
+ |GSM8k (5-shot) |18.42|
256
+