LoneStriker
commited on
Commit
•
394ae61
1
Parent(s):
cece11d
Upload folder using huggingface_hub
Browse files- .gitattributes +5 -35
- Brezn-7b-Q3_K_L.gguf +3 -0
- Brezn-7b-Q4_K_M.gguf +3 -0
- Brezn-7b-Q5_K_M.gguf +3 -0
- Brezn-7b-Q6_K.gguf +3 -0
- Brezn-7b-Q8_0.gguf +3 -0
- README.md +114 -0
- brezn.png +0 -0
- pretzel.png +0 -0
.gitattributes
CHANGED
@@ -1,35 +1,5 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
-
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
-
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
-
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
-
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
-
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
-
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
-
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
-
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
-
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
-
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
-
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
-
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
-
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
-
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
-
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
-
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
-
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
-
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
-
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
-
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
-
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
-
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
-
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
-
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
-
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
-
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
-
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
-
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
-
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
1 |
+
Brezn-7b-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
2 |
+
Brezn-7b-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
3 |
+
Brezn-7b-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
4 |
+
Brezn-7b-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
5 |
+
Brezn-7b-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Brezn-7b-Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af7365bfbd6d90966a19c02b2d4ff9269f02df276a74df0f882d673e36b7b56f
|
3 |
+
size 3822024352
|
Brezn-7b-Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5022460ba346f75613e16d1ec1bf0e44ba79231fd685b3723f171502228caa42
|
3 |
+
size 4368438944
|
Brezn-7b-Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d495193516bbe40bf619da7b7751aa693d825ff4c8a5dbcee9e7ee67d3c74431
|
3 |
+
size 5131409056
|
Brezn-7b-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:645e5bddacca78eb1603f4e1f6c5e1e7b660f7a98d9ac8afa69ca343815b7798
|
3 |
+
size 5942064800
|
Brezn-7b-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10cb80ca872c05f68f19182ad2b4210a81e28df9666bb0b44f44fb63dfecd22c
|
3 |
+
size 7695857312
|
README.md
ADDED
@@ -0,0 +1,114 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- merge
|
4 |
+
- mergekit
|
5 |
+
- lazymergekit
|
6 |
+
- FelixChao/WestSeverus-7B-DPO-v2
|
7 |
+
- mayflowergmbh/Wiedervereinigung-7b-dpo-laser
|
8 |
+
- cognitivecomputations/openchat-3.5-0106-laser
|
9 |
+
- 🥨
|
10 |
+
- 🍻
|
11 |
+
base_model:
|
12 |
+
- FelixChao/WestSeverus-7B-DPO-v2
|
13 |
+
- mayflowergmbh/Wiedervereinigung-7b-dpo-laser
|
14 |
+
- cognitivecomputations/openchat-3.5-0106-laser
|
15 |
+
license: apache-2.0
|
16 |
+
language:
|
17 |
+
- de
|
18 |
+
---
|
19 |
+
|
20 |
+
# 🥨 Brezn-7B
|
21 |
+
|
22 |
+
This is right now our best performing german speaking 7B model with an apache license, with an average of 7.49 on mt-bench-de.
|
23 |
+
You can test this model here: [mayflowergmbh/Brezn-7B-GGUF-Chat](https://huggingface.co/spaces/mayflowergmbh/Brezn-7B-GGUF-Chat).
|
24 |
+
|
25 |
+
|
26 |
+
Brezn-7B is a dpo aligned merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
|
27 |
+
* [FelixChao/WestSeverus-7B-DPO-v2](https://huggingface.co/FelixChao/WestSeverus-7B-DPO-v2)
|
28 |
+
* [mayflowergmbh/Wiedervereinigung-7b-dpo-laser](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b-dpo-laser)
|
29 |
+
* [cognitivecomputations/openchat-3.5-0106-laser](https://huggingface.co/cognitivecomputations/openchat-3.5-0106-laser)
|
30 |
+
|
31 |
+
![image/png](https://huggingface.co/mayflowergmbh/Brezn-7b/resolve/main/pretzel.png)
|
32 |
+
## 💻 Usage
|
33 |
+
|
34 |
+
In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
|
35 |
+
|
36 |
+
E.g.
|
37 |
+
```
|
38 |
+
text = "<s>[INST] What is your favourite condiment? [/INST]"
|
39 |
+
"Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> "
|
40 |
+
"[INST] Do you have mayonnaise recipes? [/INST]"
|
41 |
+
```
|
42 |
+
|
43 |
+
This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
|
44 |
+
|
45 |
+
```python
|
46 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
47 |
+
|
48 |
+
device = "cuda" # the device to load the model onto
|
49 |
+
|
50 |
+
model = AutoModelForCausalLM.from_pretrained("mayflowergmbh/Brezn-7b")
|
51 |
+
tokenizer = AutoTokenizer.from_pretrained("mayflowergmbh/Brezn-7b")
|
52 |
+
|
53 |
+
messages = [
|
54 |
+
{"role": "user", "content": "Was ist dein Lieblingsgewürz??"},
|
55 |
+
{"role": "assistant", "content": "Nun, ich mag besonders gerne einen guten Spritzer frischen Zitronensaft. Er fügt genau die richtige Menge an würzigem Geschmack hinzu, egal was ich gerade in der Küche zubereite!"},
|
56 |
+
{"role": "user", "content": "Hast du Mayonnaise-Rezepte?"}
|
57 |
+
]
|
58 |
+
|
59 |
+
encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
|
60 |
+
|
61 |
+
model_inputs = encodeds.to(device)
|
62 |
+
model.to(device)
|
63 |
+
|
64 |
+
generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
|
65 |
+
decoded = tokenizer.batch_decode(generated_ids)
|
66 |
+
print(decoded[0])
|
67 |
+
```
|
68 |
+
|
69 |
+
|
70 |
+
## mt-bench-de
|
71 |
+
```yaml
|
72 |
+
{
|
73 |
+
"first_turn": 7.6625,
|
74 |
+
"second_turn": 7.31875,
|
75 |
+
"categories": {
|
76 |
+
"writing": 8.75,
|
77 |
+
"roleplay": 8.5,
|
78 |
+
"reasoning": 6.1,
|
79 |
+
"math": 5.05,
|
80 |
+
"coding": 5.4,
|
81 |
+
"extraction": 7.975,
|
82 |
+
"stem": 9,
|
83 |
+
"humanities": 9.15
|
84 |
+
},
|
85 |
+
"average": 7.490625
|
86 |
+
}
|
87 |
+
```
|
88 |
+
|
89 |
+
## 🧩 Configuration
|
90 |
+
|
91 |
+
```yaml
|
92 |
+
models:
|
93 |
+
- model: mistralai/Mistral-7B-v0.1
|
94 |
+
# no parameters necessary for base model
|
95 |
+
- model: FelixChao/WestSeverus-7B-DPO-v2
|
96 |
+
parameters:
|
97 |
+
density: 0.60
|
98 |
+
weight: 0.30
|
99 |
+
- model: mayflowergmbh/Wiedervereinigung-7b-dpo-laser
|
100 |
+
parameters:
|
101 |
+
density: 0.65
|
102 |
+
weight: 0.40
|
103 |
+
- model: cognitivecomputations/openchat-3.5-0106-laser
|
104 |
+
parameters:
|
105 |
+
density: 0.6
|
106 |
+
weight: 0.3
|
107 |
+
merge_method: dare_ties
|
108 |
+
base_model: mistralai/Mistral-7B-v0.1
|
109 |
+
parameters:
|
110 |
+
int8_mask: true
|
111 |
+
dtype: bfloat16
|
112 |
+
random_seed: 0
|
113 |
+
```
|
114 |
+
|
brezn.png
ADDED
pretzel.png
ADDED