automerger commited on
Commit
38bf47d
1 Parent(s): bc58485

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -18,19 +18,19 @@ Experiment26T3q-7B is an automated merge created by [Maxime Labonne](https://hug
18
 
19
  ```yaml
20
  models:
21
- - model: yam-peleg/Experiment26-7B
22
  # No parameters necessary for base model
23
  - model: chihoonlee10/T3Q-Mistral-Orca-Math-DPO
24
  parameters:
25
  density: 0.53
26
  weight: 0.6
27
  merge_method: dare_ties
28
- base_model: yam-peleg/Experiment26-7B
29
  parameters:
30
  int8_mask: true
31
  dtype: bfloat16
32
  random_seed: 0
33
- ```
34
 
35
  ## 💻 Usage
36
 
 
18
 
19
  ```yaml
20
  models:
21
+ - model: rwitz/experiment26-truthy-iter-0
22
  # No parameters necessary for base model
23
  - model: chihoonlee10/T3Q-Mistral-Orca-Math-DPO
24
  parameters:
25
  density: 0.53
26
  weight: 0.6
27
  merge_method: dare_ties
28
+ base_model: rwitz/experiment26-truthy-iter-0
29
  parameters:
30
  int8_mask: true
31
  dtype: bfloat16
32
  random_seed: 0
33
+ ```
34
 
35
  ## 💻 Usage
36
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "yam-peleg/Experiment26-7B",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -20,7 +20,7 @@
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
- "transformers_version": "4.38.2",
24
  "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
1
  {
2
+ "_name_or_path": "rwitz/experiment26-truthy-iter-0",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.39.1",
24
  "use_cache": true,
25
  "vocab_size": 32000
26
  }
mergekit_config.yml CHANGED
@@ -1,15 +1,14 @@
1
 
2
  models:
3
- - model: yam-peleg/Experiment26-7B
4
  # No parameters necessary for base model
5
  - model: chihoonlee10/T3Q-Mistral-Orca-Math-DPO
6
  parameters:
7
  density: 0.53
8
  weight: 0.6
9
  merge_method: dare_ties
10
- base_model: yam-peleg/Experiment26-7B
11
  parameters:
12
  int8_mask: true
13
  dtype: bfloat16
14
  random_seed: 0
15
-
 
1
 
2
  models:
3
+ - model: rwitz/experiment26-truthy-iter-0
4
  # No parameters necessary for base model
5
  - model: chihoonlee10/T3Q-Mistral-Orca-Math-DPO
6
  parameters:
7
  density: 0.53
8
  weight: 0.6
9
  merge_method: dare_ties
10
+ base_model: rwitz/experiment26-truthy-iter-0
11
  parameters:
12
  int8_mask: true
13
  dtype: bfloat16
14
  random_seed: 0
 
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25bf360183bc037ce5a8d0f093778cb0c05706d1e9429b2e65294f3a651a00d0
3
  size 9942981696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0517ccecb91279c4f90b8420aceaed628d3d927a60f308fd342b9f1055564716
3
  size 9942981696
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9478bb0f8c1e2836a322970e8cd497d8914259e6a06b030e7b0860b8e2fd2c44
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4c379799eb07a13d2ab9c0843191b07313c5f1fe1e4e2dc5cfc27b3bf503719
3
  size 4540516344