aashish1904 commited on
Commit
6898ff1
1 Parent(s): bb2d003

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +149 -0
README.md ADDED
@@ -0,0 +1,149 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ language:
5
+ - en
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - esper
9
+ - esper-2
10
+ - valiant
11
+ - valiant-labs
12
+ - llama
13
+ - llama-3.2
14
+ - llama-3.2-instruct
15
+ - llama-3.2-instruct-3b
16
+ - llama-3
17
+ - llama-3-instruct
18
+ - llama-3-instruct-3b
19
+ - 3b
20
+ - code
21
+ - code-instruct
22
+ - python
23
+ - dev-ops
24
+ - terraform
25
+ - azure
26
+ - aws
27
+ - gcp
28
+ - architect
29
+ - engineer
30
+ - developer
31
+ - conversational
32
+ - chat
33
+ - instruct
34
+ base_model: meta-llama/Llama-3.2-3B-Instruct
35
+ datasets:
36
+ - sequelbox/Titanium
37
+ - sequelbox/Tachibana
38
+ - sequelbox/Supernova
39
+ model-index:
40
+ - name: ValiantLabs/Llama3.2-3B-Esper2
41
+ results:
42
+ - task:
43
+ type: text-generation
44
+ name: Text Generation
45
+ dataset:
46
+ name: Winogrande (5-Shot)
47
+ type: Winogrande
48
+ args:
49
+ num_few_shot: 5
50
+ metrics:
51
+ - type: acc
52
+ value: 65.27
53
+ name: acc
54
+ - task:
55
+ type: text-generation
56
+ name: Text Generation
57
+ dataset:
58
+ name: ARC Challenge (25-Shot)
59
+ type: arc-challenge
60
+ args:
61
+ num_few_shot: 25
62
+ metrics:
63
+ - type: acc_norm
64
+ value: 43.17
65
+ name: normalized accuracy
66
+ model_type: llama
67
+ license: llama3.2
68
+
69
+ ---
70
+
71
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
72
+
73
+
74
+ # QuantFactory/Llama3.2-3B-Esper2-GGUF
75
+ This is quantized version of [ValiantLabs/Llama3.2-3B-Esper2](https://huggingface.co/ValiantLabs/Llama3.2-3B-Esper2) created using llama.cpp
76
+
77
+ # Original Model Card
78
+
79
+
80
+
81
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/64f267a8a4f79a118e0fcc89/4I6oK8DG0so4VD8GroFsd.jpeg)
82
+
83
+
84
+ Esper 2 is a DevOps and cloud architecture code specialist built on Llama 3.2 3b.
85
+ - Expertise-driven, an AI assistant focused on AWS, Azure, GCP, Terraform, Dockerfiles, pipelines, shell scripts and more!
86
+ - Real world problem solving and high quality code instruct performance within the Llama 3.2 Instruct chat format
87
+ - Finetuned on synthetic [DevOps-instruct](https://huggingface.co/datasets/sequelbox/Titanium) and [code-instruct](https://huggingface.co/datasets/sequelbox/Tachibana) data generated with Llama 3.1 405b.
88
+ - Overall chat performance supplemented with [generalist chat data.](https://huggingface.co/datasets/sequelbox/Supernova)
89
+
90
+ Try our code-instruct AI assistant [Enigma!](https://huggingface.co/ValiantLabs/Llama3.1-8B-Enigma)
91
+
92
+
93
+ ## Version
94
+
95
+ This is the **2024-10-03** release of Esper 2 for Llama 3.2 3b.
96
+
97
+ Esper 2 is also available for [Llama 3.1 8b!](https://huggingface.co/ValiantLabs/Llama3.1-8B-Esper2)
98
+
99
+ Esper 2 will be coming to more model sizes soon :)
100
+
101
+
102
+ ## Prompting Guide
103
+ Esper 2 uses the [Llama 3.2 Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) prompt format. The example script below can be used as a starting point for general chat:
104
+
105
+ ```python
106
+ import transformers
107
+ import torch
108
+
109
+ model_id = "ValiantLabs/Llama3.2-3B-Esper2"
110
+
111
+ pipeline = transformers.pipeline(
112
+ "text-generation",
113
+ model=model_id,
114
+ model_kwargs={"torch_dtype": torch.bfloat16},
115
+ device_map="auto",
116
+ )
117
+
118
+ messages = [
119
+ {"role": "system", "content": "You are an AI assistant."},
120
+ {"role": "user", "content": "Hi, how do I optimize the size of a Docker image?"}
121
+ ]
122
+
123
+ outputs = pipeline(
124
+ messages,
125
+ max_new_tokens=2048,
126
+ )
127
+
128
+ print(outputs[0]["generated_text"][-1])
129
+ ```
130
+
131
+ ## The Model
132
+ Esper 2 is built on top of Llama 3.2 3b Instruct, improving performance through high quality DevOps, code, and chat data in Llama 3.2 Instruct prompt style.
133
+
134
+ Our current version of Esper 2 is trained on DevOps data from [sequelbox/Titanium](https://huggingface.co/datasets/sequelbox/Titanium), supplemented by code-instruct data from [sequelbox/Tachibana](https://huggingface.co/datasets/sequelbox/Tachibana) and general chat data from [sequelbox/Supernova.](https://huggingface.co/datasets/sequelbox/Supernova)
135
+
136
+
137
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63444f2687964b331809eb55/VCJ8Fmefd8cdVhXSSxJiD.jpeg)
138
+
139
+
140
+ Esper 2 is created by [Valiant Labs.](http://valiantlabs.ca/)
141
+
142
+ [Check out our HuggingFace page for Shining Valiant 2, Enigma, and our other Build Tools models for creators!](https://huggingface.co/ValiantLabs)
143
+
144
+ [Follow us on X for updates on our models!](https://twitter.com/valiant_labs)
145
+
146
+ We care about open source.
147
+ For everyone to use.
148
+
149
+ We encourage others to finetune further from our models.