MoE-Girl-1BA-7BT / README.md
Fizzarolli's picture
Update README.md
f73df71 verified
|
raw
history blame
1.03 kB
metadata
library_name: transformers
license: apache-2.0
base_model: allenai/OLMoE-1B-7B-0924
tags:
  - axolotl
  - moe
  - roleplay
model-index:
  - name: MoE-girl_1BA_7BT
    results: []

MoE Girl 1bA 7bT

image/jpeg

A finetune of OLMoE by AllenAI designed for roleplaying (and maybe general usecases if you try hartd enough).

Disclaimer

PLEASE do not expect godliness out of this, it's a model with 1 billion active parameters at once. Expect something more akin to Gemma 2 2B, not Llama 3 8B.

Prompting

Use ChatML.

<|im_start|>system
You are a helpful assistant who talks like a pirate.<|im_end|>
<|im_start|>user
Hello there!<|im_end|>
<|im_start|>assistant
Yarr harr harr, me matey!<|im_end|>

Thanks

Special thanks to the members of Allura for testing and emotional support, as well as the creators of all the datasets that were used in the Special Sauce used to train this model. I love you all <3 - Fizz