LoneStriker commited on
Commit
394ae61
1 Parent(s): cece11d

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -1,35 +1,5 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ Brezn-7b-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
2
+ Brezn-7b-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
3
+ Brezn-7b-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
4
+ Brezn-7b-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
5
+ Brezn-7b-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Brezn-7b-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af7365bfbd6d90966a19c02b2d4ff9269f02df276a74df0f882d673e36b7b56f
3
+ size 3822024352
Brezn-7b-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5022460ba346f75613e16d1ec1bf0e44ba79231fd685b3723f171502228caa42
3
+ size 4368438944
Brezn-7b-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d495193516bbe40bf619da7b7751aa693d825ff4c8a5dbcee9e7ee67d3c74431
3
+ size 5131409056
Brezn-7b-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:645e5bddacca78eb1603f4e1f6c5e1e7b660f7a98d9ac8afa69ca343815b7798
3
+ size 5942064800
Brezn-7b-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10cb80ca872c05f68f19182ad2b4210a81e28df9666bb0b44f44fb63dfecd22c
3
+ size 7695857312
README.md ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - merge
4
+ - mergekit
5
+ - lazymergekit
6
+ - FelixChao/WestSeverus-7B-DPO-v2
7
+ - mayflowergmbh/Wiedervereinigung-7b-dpo-laser
8
+ - cognitivecomputations/openchat-3.5-0106-laser
9
+ - 🥨
10
+ - 🍻
11
+ base_model:
12
+ - FelixChao/WestSeverus-7B-DPO-v2
13
+ - mayflowergmbh/Wiedervereinigung-7b-dpo-laser
14
+ - cognitivecomputations/openchat-3.5-0106-laser
15
+ license: apache-2.0
16
+ language:
17
+ - de
18
+ ---
19
+
20
+ # 🥨 Brezn-7B
21
+
22
+ This is right now our best performing german speaking 7B model with an apache license, with an average of 7.49 on mt-bench-de.
23
+ You can test this model here: [mayflowergmbh/Brezn-7B-GGUF-Chat](https://huggingface.co/spaces/mayflowergmbh/Brezn-7B-GGUF-Chat).
24
+
25
+
26
+ Brezn-7B is a dpo aligned merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
27
+ * [FelixChao/WestSeverus-7B-DPO-v2](https://huggingface.co/FelixChao/WestSeverus-7B-DPO-v2)
28
+ * [mayflowergmbh/Wiedervereinigung-7b-dpo-laser](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b-dpo-laser)
29
+ * [cognitivecomputations/openchat-3.5-0106-laser](https://huggingface.co/cognitivecomputations/openchat-3.5-0106-laser)
30
+
31
+ ![image/png](https://huggingface.co/mayflowergmbh/Brezn-7b/resolve/main/pretzel.png)
32
+ ## 💻 Usage
33
+
34
+ In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
35
+
36
+ E.g.
37
+ ```
38
+ text = "<s>[INST] What is your favourite condiment? [/INST]"
39
+ "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> "
40
+ "[INST] Do you have mayonnaise recipes? [/INST]"
41
+ ```
42
+
43
+ This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
44
+
45
+ ```python
46
+ from transformers import AutoModelForCausalLM, AutoTokenizer
47
+
48
+ device = "cuda" # the device to load the model onto
49
+
50
+ model = AutoModelForCausalLM.from_pretrained("mayflowergmbh/Brezn-7b")
51
+ tokenizer = AutoTokenizer.from_pretrained("mayflowergmbh/Brezn-7b")
52
+
53
+ messages = [
54
+ {"role": "user", "content": "Was ist dein Lieblingsgewürz??"},
55
+ {"role": "assistant", "content": "Nun, ich mag besonders gerne einen guten Spritzer frischen Zitronensaft. Er fügt genau die richtige Menge an würzigem Geschmack hinzu, egal was ich gerade in der Küche zubereite!"},
56
+ {"role": "user", "content": "Hast du Mayonnaise-Rezepte?"}
57
+ ]
58
+
59
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
60
+
61
+ model_inputs = encodeds.to(device)
62
+ model.to(device)
63
+
64
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
65
+ decoded = tokenizer.batch_decode(generated_ids)
66
+ print(decoded[0])
67
+ ```
68
+
69
+
70
+ ## mt-bench-de
71
+ ```yaml
72
+ {
73
+ "first_turn": 7.6625,
74
+ "second_turn": 7.31875,
75
+ "categories": {
76
+ "writing": 8.75,
77
+ "roleplay": 8.5,
78
+ "reasoning": 6.1,
79
+ "math": 5.05,
80
+ "coding": 5.4,
81
+ "extraction": 7.975,
82
+ "stem": 9,
83
+ "humanities": 9.15
84
+ },
85
+ "average": 7.490625
86
+ }
87
+ ```
88
+
89
+ ## 🧩 Configuration
90
+
91
+ ```yaml
92
+ models:
93
+ - model: mistralai/Mistral-7B-v0.1
94
+ # no parameters necessary for base model
95
+ - model: FelixChao/WestSeverus-7B-DPO-v2
96
+ parameters:
97
+ density: 0.60
98
+ weight: 0.30
99
+ - model: mayflowergmbh/Wiedervereinigung-7b-dpo-laser
100
+ parameters:
101
+ density: 0.65
102
+ weight: 0.40
103
+ - model: cognitivecomputations/openchat-3.5-0106-laser
104
+ parameters:
105
+ density: 0.6
106
+ weight: 0.3
107
+ merge_method: dare_ties
108
+ base_model: mistralai/Mistral-7B-v0.1
109
+ parameters:
110
+ int8_mask: true
111
+ dtype: bfloat16
112
+ random_seed: 0
113
+ ```
114
+
brezn.png ADDED
pretzel.png ADDED