aashish1904 commited on
Commit
aa7975a
1 Parent(s): 2c759bd

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +137 -0
README.md ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ language:
5
+ - en
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - esper
9
+ - esper-2
10
+ - valiant
11
+ - valiant-labs
12
+ - llama
13
+ - llama-3.1
14
+ - llama-3.1-instruct
15
+ - llama-3.1-instruct-8b
16
+ - llama-3
17
+ - llama-3-instruct
18
+ - llama-3-instruct-8b
19
+ - 8b
20
+ - code
21
+ - code-instruct
22
+ - python
23
+ - dev-ops
24
+ - terraform
25
+ - azure
26
+ - aws
27
+ - gcp
28
+ - architect
29
+ - engineer
30
+ - developer
31
+ - conversational
32
+ - chat
33
+ - instruct
34
+ base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
35
+ datasets:
36
+ - sequelbox/Titanium
37
+ - sequelbox/Tachibana
38
+ - sequelbox/Supernova
39
+ model_type: llama
40
+ model-index:
41
+ - name: ValiantLabs/Llama3.1-8B-Esper2
42
+ results:
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: Winogrande (5-Shot)
48
+ type: Winogrande
49
+ args:
50
+ num_few_shot: 5
51
+ metrics:
52
+ - type: acc
53
+ value: 75.85
54
+ name: acc
55
+ license: llama3.1
56
+
57
+ ---
58
+
59
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
60
+
61
+
62
+ # QuantFactory/Llama3.1-8B-Esper2-GGUF
63
+ This is quantized version of [ValiantLabs/Llama3.1-8B-Esper2](https://huggingface.co/ValiantLabs/Llama3.1-8B-Esper2) created using llama.cpp
64
+
65
+ # Original Model Card
66
+
67
+
68
+
69
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/64f267a8a4f79a118e0fcc89/4I6oK8DG0so4VD8GroFsd.jpeg)
70
+
71
+
72
+ Esper 2 is a DevOps and cloud architecture code specialist built on Llama 3.1 8b.
73
+ - Expertise-driven, an AI assistant focused on AWS, Azure, GCP, Terraform, Dockerfiles, pipelines, shell scripts and more!
74
+ - Real world problem solving and high quality code instruct performance within the Llama 3.1 Instruct chat format
75
+ - Finetuned on synthetic [DevOps-instruct](https://huggingface.co/datasets/sequelbox/Titanium) and [code-instruct](https://huggingface.co/datasets/sequelbox/Tachibana) data generated with Llama 3.1 405b.
76
+ - Overall chat performance supplemented with [generalist chat data.](https://huggingface.co/datasets/sequelbox/Supernova)
77
+
78
+ Try our code-instruct AI assistant [Enigma!](https://huggingface.co/ValiantLabs/Llama3.1-8B-Enigma)
79
+
80
+
81
+ ## Version
82
+
83
+ This is the **2024-10-02** release of Esper 2 for Llama 3.1 8b.
84
+
85
+ Esper 2 is now available for [Llama 3.2 3b!](https://huggingface.co/ValiantLabs/Llama3.2-3B-Esper2)
86
+
87
+ Esper 2 will be coming to more model sizes soon :)
88
+
89
+
90
+ ## Prompting Guide
91
+ Esper 2 uses the [Llama 3.1 Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) prompt format. The example script below can be used as a starting point for general chat:
92
+
93
+ ```python
94
+ import transformers
95
+ import torch
96
+
97
+ model_id = "ValiantLabs/Llama3.1-8B-Esper2"
98
+
99
+ pipeline = transformers.pipeline(
100
+ "text-generation",
101
+ model=model_id,
102
+ model_kwargs={"torch_dtype": torch.bfloat16},
103
+ device_map="auto",
104
+ )
105
+
106
+ messages = [
107
+ {"role": "system", "content": "You are an AI assistant."},
108
+ {"role": "user", "content": "Hi, how do I optimize the size of a Docker image?"}
109
+ ]
110
+
111
+ outputs = pipeline(
112
+ messages,
113
+ max_new_tokens=2048,
114
+ )
115
+
116
+ print(outputs[0]["generated_text"][-1])
117
+ ```
118
+
119
+ ## The Model
120
+ Esper 2 is built on top of Llama 3.1 8b Instruct, improving performance through high quality DevOps, code, and chat data in Llama 3.1 Instruct prompt style.
121
+
122
+ Our current version of Esper 2 is trained on DevOps data from [sequelbox/Titanium](https://huggingface.co/datasets/sequelbox/Titanium), supplemented by code-instruct data from [sequelbox/Tachibana](https://huggingface.co/datasets/sequelbox/Tachibana) and general chat data from [sequelbox/Supernova.](https://huggingface.co/datasets/sequelbox/Supernova)
123
+
124
+
125
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63444f2687964b331809eb55/VCJ8Fmefd8cdVhXSSxJiD.jpeg)
126
+
127
+
128
+ Esper 2 is created by [Valiant Labs.](http://valiantlabs.ca/)
129
+
130
+ [Check out our HuggingFace page for Shining Valiant 2 Enigma, and our other Build Tools models for creators!](https://huggingface.co/ValiantLabs)
131
+
132
+ [Follow us on X for updates on our models!](https://twitter.com/valiant_labs)
133
+
134
+ We care about open source.
135
+ For everyone to use.
136
+
137
+ We encourage others to finetune further from our models.