Edit model card

My upload speeds have been cooked and unstable lately.
Realistically I'd need to move to get a better provider.
If you want and you are able to, you can support that endeavor and others here (Ko-fi). I apologize for disrupting your experience.

GGUF-IQ-Imatrix quants for jeiku/Average_Normie_l3_v1_8B.

Updated! These quants have been redone with the fixes from llama.cpp/pull/6920 in mind.
Use KoboldCpp version 1.64 or higher.

Compatible SillyTavern presets here (simple) or here (Virt's).
Use the latest version of KoboldCpp. Use the provided presets.
This is all still highly experimental, let the authors know how it performs for you, feedback is more important than ever now.

For 8GB VRAM GPUs, I recommend the Q4_K_M-imat quant for up to 12288 context sizes.

Original model information:

Average Normie v1

image/png

A model by an average normie for the average normie.

This model is a stock merge of the following models:

https://huggingface.co/cgato/L3-TheSpice-8b-v0.1.3

https://huggingface.co/Sao10K/L3-Solana-8B-v1

https://huggingface.co/ResplendentAI/Kei_Llama3_8B

The final merge then had the following LoRA applied over it:

https://huggingface.co/ResplendentAI/Theory_of_Mind_Llama3

This should be an intelligent and adept roleplaying model.

Downloads last month
120
GGUF
Model size
8.03B params
Architecture
llama

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference API
Unable to determine this model's library. Check the docs .