πΆπ»π¦ CRIA
Collection
Set of models derived from Meta Llama 2 7b model architecture
β’
8 items
β’
Updated
This is a frankenmerge of davzoku/cria-llama2-7b-v1.3.
The configuration is the same as Undi95/Mistral-11B-v0.1, mlabonne/FrankenBeagle14-11B and the DUS technique used in upstage/SOLAR-10.7B-v1.0.
Please be aware that this model is highly experimental, and no further training has been conducted following the merge. Therefore, the model performance may not meet expectations, as described in the SOLAR paper
FrankenCRIA v1.3 comes with several variants.
This model was merged using the passthrough merge method.
The following models were included in the merge:
The following YAML configuration was used to produce this model.
# https://huggingface.co/Undi95/Mistral-11B-v0.1
slices:
- sources:
- model: davzoku/cria-llama2-7b-v1.3
layer_range: [0, 24]
- sources:
- model: davzoku/cria-llama2-7b-v1.3
layer_range: [8, 32]
merge_method: passthrough
dtype: bfloat16
Base model
davzoku/cria-llama2-7b-v1.3