TheBloke commited on
Commit
b10d717
1 Parent(s): 06bea33

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +164 -0
README.md ADDED
@@ -0,0 +1,164 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: qblocks/zephyr_7b_norobots
3
+ datasets:
4
+ - HuggingFaceH4/no_robots
5
+ inference: false
6
+ library_name: peft
7
+ license: apache-2.0
8
+ model_creator: MonsterAPI
9
+ model_name: Zephyr 7B Norobots
10
+ model_type: mistral
11
+ prompt_template: '<|im_start|>system
12
+
13
+ {system_message}<|im_end|>
14
+
15
+ <|im_start|>user
16
+
17
+ {prompt}<|im_end|>
18
+
19
+ <|im_start|>assistant
20
+
21
+ '
22
+ quantized_by: TheBloke
23
+ tags:
24
+ - code
25
+ - instruct
26
+ - zephyr
27
+ ---
28
+
29
+ <!-- header start -->
30
+ <!-- 200823 -->
31
+ <div style="width: auto; margin-left: auto; margin-right: auto">
32
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
33
+ </div>
34
+ <div style="display: flex; justify-content: space-between; width: 100%;">
35
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
36
+ <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p>
37
+ </div>
38
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
39
+ <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
40
+ </div>
41
+ </div>
42
+ <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div>
43
+ <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
44
+ <!-- header end -->
45
+
46
+ # Zephyr 7B Norobots - FP16
47
+ - Model creator: [MonsterAPI](https://huggingface.co/qblocks)
48
+ - Original model: [Zephyr 7B Norobots](qblocks/zephyr_7b_norobots)
49
+
50
+ <!-- description start -->
51
+ ## Description
52
+
53
+ This repo contains pytorch format fp16 model files for [MonsterAPI's Zephyr 7B Norobots](qblocks/zephyr_7b_norobots).
54
+
55
+ It is the result of either merging a LoRA, or converting the source repository to float16.
56
+
57
+ These files were quantised using hardware kindly provided by [Massed Compute](https://massedcompute.com/).
58
+
59
+ <!-- description end -->
60
+ <!-- repositories-available start -->
61
+ ## Repositories available
62
+
63
+ * [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/zephyr_7b_norobots-AWQ)
64
+ * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/zephyr_7b_norobots-GPTQ)
65
+ * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/zephyr_7b_norobots-GGUF)
66
+ * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/zephyr_7b_norobots-fp16)
67
+ * [MonsterAPI's original LoRA adapter, which can be merged on to the base model.](https://huggingface.co/qblocks/zephyr_7b_norobots)
68
+
69
+ <!-- repositories-available start -->
70
+
71
+ <!-- prompt-template start -->
72
+ ## Prompt template: ChatML
73
+
74
+ ```
75
+ <|im_start|>system
76
+ {system_message}<|im_end|>
77
+ <|im_start|>user
78
+ {prompt}<|im_end|>
79
+ <|im_start|>assistant
80
+
81
+ ```
82
+
83
+ <!-- prompt-template end -->
84
+
85
+
86
+
87
+
88
+ <!-- footer start -->
89
+ <!-- 200823 -->
90
+ ## Discord
91
+
92
+ For further support, and discussions on these models and AI in general, join us at:
93
+
94
+ [TheBloke AI's Discord server](https://discord.gg/theblokeai)
95
+
96
+ ## Thanks, and how to contribute
97
+
98
+ Thanks to the [chirper.ai](https://chirper.ai) team!
99
+
100
+ Thanks to Clay from [gpus.llm-utils.org](llm-utils)!
101
+
102
+ I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
103
+
104
+ If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
105
+
106
+ Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
107
+
108
+ * Patreon: https://patreon.com/TheBlokeAI
109
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
110
+
111
+ **Special thanks to**: Aemon Algiz.
112
+
113
+ **Patreon special mentions**: Brandon Frisco, LangChain4j, Spiking Neurons AB, transmissions 11, Joseph William Delisle, Nitin Borwankar, Willem Michiel, Michael Dempsey, vamX, Jeffrey Morgan, zynix, jjj, Omer Bin Jawed, Sean Connelly, jinyuan sun, Jeromy Smith, Shadi, Pawan Osman, Chadd, Elijah Stavena, Illia Dulskyi, Sebastain Graf, Stephen Murray, terasurfer, Edmond Seymore, Celu Ramasamy, Mandus, Alex, biorpg, Ajan Kanaga, Clay Pascal, Raven Klaugh, 阿明, K, ya boyyy, usrbinkat, Alicia Loh, John Villwock, ReadyPlayerEmma, Chris Smitley, Cap'n Zoog, fincy, GodLy, S_X, sidney chen, Cory Kujawski, OG, Mano Prime, AzureBlack, Pieter, Kalila, Spencer Kim, Tom X Nguyen, Stanislav Ovsiannikov, Michael Levine, Andrey, Trailburnt, Vadim, Enrico Ros, Talal Aujan, Brandon Phillips, Jack West, Eugene Pentland, Michael Davis, Will Dee, webtim, Jonathan Leane, Alps Aficionado, Rooh Singh, Tiffany J. Kim, theTransient, Luke @flexchar, Elle, Caitlyn Gatomon, Ari Malik, subjectnull, Johann-Peter Hartmann, Trenton Dambrowitz, Imad Khwaja, Asp the Wyvern, Emad Mostaque, Rainer Wilmers, Alexandros Triantafyllidis, Nicholas, Pedro Madruga, SuperWojo, Harry Royden McLaughlin, James Bentley, Olakabola, David Ziegler, Ai Maven, Jeff Scroggin, Nikolai Manek, Deo Leter, Matthew Berman, Fen Risland, Ken Nordquist, Manuel Alberto Morcote, Luke Pendergrass, TL, Fred von Graf, Randy H, Dan Guido, NimbleBox.ai, Vitor Caleffi, Gabriel Tamborski, knownsqashed, Lone Striker, Erik Bjäreholt, John Detwiler, Leonard Tan, Iucharbius
114
+
115
+
116
+ Thank you to all my generous patrons and donaters!
117
+
118
+ And thank you again to a16z for their generous grant.
119
+
120
+ <!-- footer end -->
121
+
122
+ # Original model card: MonsterAPI's Zephyr 7B Norobots
123
+
124
+
125
+ ### Finetuning Overview:
126
+
127
+ **Model Used:** HuggingFaceH4/zephyr-7b-alpha
128
+
129
+ **Dataset:** HuggingFaceH4/no_robots
130
+
131
+ #### Dataset Insights:
132
+
133
+ [No Robots](https://huggingface.co/datasets/HuggingFaceH4/no_robots) is a high-quality dataset of 10,000 instructions and demonstrations created by skilled human annotators. This data can be used for supervised fine-tuning (SFT) to make language models follow instructions better.
134
+
135
+ #### Finetuning Details:
136
+
137
+ With the utilization of [MonsterAPI](https://monsterapi.ai)'s [LLM finetuner](https://docs.monsterapi.ai/fine-tune-a-large-language-model-llm), this finetuning:
138
+
139
+ - Was achieved with great cost-effectiveness.
140
+ - Completed in a total duration of 36mins 47secs for 1 epoch using an A6000 48GB GPU.
141
+ - Costed `$1.212` for the entire epoch.
142
+
143
+ #### Hyperparameters & Additional Details:
144
+
145
+ - **Epochs:** 1
146
+ - **Cost Per Epoch:** $1.212
147
+ - **Total Finetuning Cost:** $1.212
148
+ - **Model Path:** HuggingFaceH4/zephyr-7b-alpha
149
+ - **Learning Rate:** 0.0002
150
+ - **Data Split:** 100% train
151
+ - **Gradient Accumulation Steps:** 4
152
+ - **lora r:** 32
153
+ - **lora alpha:** 64
154
+
155
+ ---
156
+ #### Prompt Structure
157
+ ```
158
+ <|system|> <|endoftext|> <|user|> [USER PROMPT]<|endoftext|> <|assistant|> [ASSISTANT ANSWER] <|endoftext|>
159
+ ```
160
+ #### Train loss :
161
+
162
+ ![training loss](https://cdn-uploads.huggingface.co/production/uploads/63ba46aa0a9866b28cb19a14/EYUTlcPFz-2nXzNj5_gsW.png)
163
+
164
+ license: apache-2.0