--- base_model: - Delta-Vector/HollandV2R1 - anthracite-org/magnum-v2-4b - Delta-Vector/Holland-4B-V1 - IntervitensInc/intv-mimi-sus library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [Delta-Vector/Holland-4B-V1](https://huggingface.co/Delta-Vector/Holland-4B-V1) as a base. ### Models Merged The following models were included in the merge: * [Delta-Vector/HollandV2R1](https://huggingface.co/Delta-Vector/HollandV2R1) * [anthracite-org/magnum-v2-4b](https://huggingface.co/anthracite-org/magnum-v2-4b) * [IntervitensInc/intv-mimi-sus](https://huggingface.co/IntervitensInc/intv-mimi-sus) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: Delta-Vector/Holland-4B-V1 dtype: bfloat16 merge_method: task_arithmetic parameters: normalize: true slices: - sources: - layer_range: [0, 32] model: Delta-Vector/Holland-4B-V1 - layer_range: [0, 32] model: Delta-Vector/HollandV2R1 parameters: weight: 0.30 - layer_range: [0, 32] model: IntervitensInc/intv-mimi-sus parameters: weight: 0.30 - layer_range: [0, 32] model: anthracite-org/magnum-v2-4b parameters: weight: 0.40 ```