My merges
Collection
A collection of language models I've merged, in chronological order
•
11 items
•
Updated
This is a a slightly experimental frankenmerge of Microsoft's Orca-2-13b and Undi and IkariDev's Noromaid-v0.2-13b.
The model recipe is as follows:
- [0, 12] from Noromaid (12 layers)
- [12, 22] from Orca (10 layers)
- [14, 24] from Orca (10 layers)
- [18, 28] from Orca (10 layers)
- [28, 40] from Noromaid (12 layers)
In my testing so far, the model performs exceptionally well. Your experience may vary.
The prompt format is Alpaca. You can use the standard format as shown, but for best results, you should customize the system prompt to your specific needs.
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{YOUR MESSAGE HERE}
### Response:
{BOT MESSAGE HERE}
<s>
</s>
32768
via YaRN (original context length was 4096
)