mav23 commited on
Commit
ef410ab
1 Parent(s): d1662f3

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +137 -0
  3. llama3.2-3b-esper2.Q4_0.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ llama3.2-3b-esper2.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - esper
7
+ - esper-2
8
+ - valiant
9
+ - valiant-labs
10
+ - llama
11
+ - llama-3.2
12
+ - llama-3.2-instruct
13
+ - llama-3.2-instruct-3b
14
+ - llama-3
15
+ - llama-3-instruct
16
+ - llama-3-instruct-3b
17
+ - 3b
18
+ - code
19
+ - code-instruct
20
+ - python
21
+ - dev-ops
22
+ - terraform
23
+ - azure
24
+ - aws
25
+ - gcp
26
+ - architect
27
+ - engineer
28
+ - developer
29
+ - conversational
30
+ - chat
31
+ - instruct
32
+ base_model: meta-llama/Llama-3.2-3B-Instruct
33
+ datasets:
34
+ - sequelbox/Titanium
35
+ - sequelbox/Tachibana
36
+ - sequelbox/Supernova
37
+ model-index:
38
+ - name: ValiantLabs/Llama3.2-3B-Esper2
39
+ results:
40
+ - task:
41
+ type: text-generation
42
+ name: Text Generation
43
+ dataset:
44
+ name: Winogrande (5-Shot)
45
+ type: Winogrande
46
+ args:
47
+ num_few_shot: 5
48
+ metrics:
49
+ - type: acc
50
+ value: 65.27
51
+ name: acc
52
+ - task:
53
+ type: text-generation
54
+ name: Text Generation
55
+ dataset:
56
+ name: ARC Challenge (25-Shot)
57
+ type: arc-challenge
58
+ args:
59
+ num_few_shot: 25
60
+ metrics:
61
+ - type: acc_norm
62
+ value: 43.17
63
+ name: normalized accuracy
64
+ model_type: llama
65
+ license: llama3.2
66
+ ---
67
+
68
+
69
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/64f267a8a4f79a118e0fcc89/4I6oK8DG0so4VD8GroFsd.jpeg)
70
+
71
+
72
+ Esper 2 is a DevOps and cloud architecture code specialist built on Llama 3.2 3b.
73
+ - Expertise-driven, an AI assistant focused on AWS, Azure, GCP, Terraform, Dockerfiles, pipelines, shell scripts and more!
74
+ - Real world problem solving and high quality code instruct performance within the Llama 3.2 Instruct chat format
75
+ - Finetuned on synthetic [DevOps-instruct](https://huggingface.co/datasets/sequelbox/Titanium) and [code-instruct](https://huggingface.co/datasets/sequelbox/Tachibana) data generated with Llama 3.1 405b.
76
+ - Overall chat performance supplemented with [generalist chat data.](https://huggingface.co/datasets/sequelbox/Supernova)
77
+
78
+ Try our code-instruct AI assistant [Enigma!](https://huggingface.co/ValiantLabs/Llama3.1-8B-Enigma)
79
+
80
+
81
+ ## Version
82
+
83
+ This is the **2024-10-03** release of Esper 2 for Llama 3.2 3b.
84
+
85
+ Esper 2 is also available for [Llama 3.1 8b!](https://huggingface.co/ValiantLabs/Llama3.1-8B-Esper2)
86
+
87
+ Esper 2 will be coming to more model sizes soon :)
88
+
89
+
90
+ ## Prompting Guide
91
+ Esper 2 uses the [Llama 3.2 Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) prompt format. The example script below can be used as a starting point for general chat:
92
+
93
+ ```python
94
+ import transformers
95
+ import torch
96
+
97
+ model_id = "ValiantLabs/Llama3.2-3B-Esper2"
98
+
99
+ pipeline = transformers.pipeline(
100
+ "text-generation",
101
+ model=model_id,
102
+ model_kwargs={"torch_dtype": torch.bfloat16},
103
+ device_map="auto",
104
+ )
105
+
106
+ messages = [
107
+ {"role": "system", "content": "You are an AI assistant."},
108
+ {"role": "user", "content": "Hi, how do I optimize the size of a Docker image?"}
109
+ ]
110
+
111
+ outputs = pipeline(
112
+ messages,
113
+ max_new_tokens=2048,
114
+ )
115
+
116
+ print(outputs[0]["generated_text"][-1])
117
+ ```
118
+
119
+ ## The Model
120
+ Esper 2 is built on top of Llama 3.2 3b Instruct, improving performance through high quality DevOps, code, and chat data in Llama 3.2 Instruct prompt style.
121
+
122
+ Our current version of Esper 2 is trained on DevOps data from [sequelbox/Titanium](https://huggingface.co/datasets/sequelbox/Titanium), supplemented by code-instruct data from [sequelbox/Tachibana](https://huggingface.co/datasets/sequelbox/Tachibana) and general chat data from [sequelbox/Supernova.](https://huggingface.co/datasets/sequelbox/Supernova)
123
+
124
+
125
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63444f2687964b331809eb55/VCJ8Fmefd8cdVhXSSxJiD.jpeg)
126
+
127
+
128
+ Esper 2 is created by [Valiant Labs.](http://valiantlabs.ca/)
129
+
130
+ [Check out our HuggingFace page for Shining Valiant 2, Enigma, and our other Build Tools models for creators!](https://huggingface.co/ValiantLabs)
131
+
132
+ [Follow us on X for updates on our models!](https://twitter.com/valiant_labs)
133
+
134
+ We care about open source.
135
+ For everyone to use.
136
+
137
+ We encourage others to finetune further from our models.
llama3.2-3b-esper2.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1abb522cd58e8f281a7f5bd375bf81f0add39d8d55e0a4c8f9e79ff85730b7e
3
+ size 1917191200