|
base_model: ./evolve-test-2/input_models/Meta-Llama-3-8B_3118938735 |
|
dtype: bfloat16 |
|
merge_method: dare_ties |
|
parameters: |
|
int8_mask: 1.0 |
|
normalize: 0.0 |
|
slices: |
|
- sources: |
|
- layer_range: [0, 8] |
|
model: ./evolve-test-2/input_models/Meta-Llama-3-8B-Instruct_531182157 |
|
parameters: |
|
density: 0.7215873011769947 |
|
weight: 0.5672371220537422 |
|
- layer_range: [0, 8] |
|
model: ./evolve-test-2/input_models/Llama-3-Open-Ko-8B_4015212464 |
|
parameters: |
|
density: 0.6535475316726772 |
|
weight: 0.4297404766011409 |
|
- layer_range: [0, 8] |
|
model: ./evolve-test-2/input_models/Meta-Llama-3-8B_3118938735 |
|
- sources: |
|
- layer_range: [8, 16] |
|
model: ./evolve-test-2/input_models/Meta-Llama-3-8B-Instruct_531182157 |
|
parameters: |
|
density: 1.0 |
|
weight: 1.258044681241929 |
|
- layer_range: [8, 16] |
|
model: ./evolve-test-2/input_models/Llama-3-Open-Ko-8B_4015212464 |
|
parameters: |
|
density: 0.6160642519526411 |
|
weight: 0.4418461544593263 |
|
- layer_range: [8, 16] |
|
model: ./evolve-test-2/input_models/Meta-Llama-3-8B_3118938735 |
|
- sources: |
|
- layer_range: [16, 24] |
|
model: ./evolve-test-2/input_models/Meta-Llama-3-8B-Instruct_531182157 |
|
parameters: |
|
density: 0.5618652598821763 |
|
weight: 0.42351796523690527 |
|
- layer_range: [16, 24] |
|
model: ./evolve-test-2/input_models/Llama-3-Open-Ko-8B_4015212464 |
|
parameters: |
|
density: 0.932056672842011 |
|
weight: 0.3737246575450384 |
|
- layer_range: [16, 24] |
|
model: ./evolve-test-2/input_models/Meta-Llama-3-8B_3118938735 |
|
- sources: |
|
- layer_range: [24, 32] |
|
model: ./evolve-test-2/input_models/Meta-Llama-3-8B-Instruct_531182157 |
|
parameters: |
|
density: 0.3934598670872163 |
|
weight: 0.003967952578762479 |
|
- layer_range: [24, 32] |
|
model: ./evolve-test-2/input_models/Llama-3-Open-Ko-8B_4015212464 |
|
parameters: |
|
density: 1.0 |
|
weight: 0.4946465398634956 |
|
- layer_range: [24, 32] |
|
model: ./evolve-test-2/input_models/Meta-Llama-3-8B_3118938735 |