|
--- |
|
license: llama3 |
|
library_name: transformers |
|
tags: |
|
- nsfw |
|
- not-for-all-audiences |
|
- llama-3 |
|
- text-generation-inference |
|
- mergekit |
|
- merge |
|
base_model: HiroseKoichi/L3-8B-Lunar-Stheno |
|
pipeline_tag: text-generation |
|
--- |
|
|
|
# L3-8B-Lunar-Stheno-GGUF |
|
This is quantrized version of [HiroseKoichi/L3-8B-Lunar-Stheno](https://huggingface.co/HiroseKoichi/L3-8B-Lunar-Stheno?not-for-all-audiences=true) created using llama.cpp |
|
|
|
# Model Description |
|
L3-8B-Lunaris-v1 is definitely a significant improvement over L3-8B-Stheno-v3.2 in terms of situational awareness and prose, but it's not without issues: the response length can sometimes be very long, causing it to go on a rant; it tends to not take direct action, saying that it will do something but never actually doing it; and its performance outside of roleplay took a hit. |
|
|
|
This merge fixes all of those issues, and I'm genuinely impressed with the results. While I did use a SLERP merge to create this model, there was no blending of the models; all I did was replace L3-8B-Stheno-v3.2's weights with L3-8B-Lunaris-v1's. |
|
|
|
# Details |
|
- **License**: [llama3](https://llama.meta.com/llama3/license/) |
|
- **Instruct Format**: [llama-3](https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3/) |
|
- **Context Size**: 8K |
|
|
|
## Models Used |
|
- [L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2) |
|
- [L3-8B-Lunaris-v1](https://huggingface.co/Sao10K/L3-8B-Lunaris-v1) |
|
|
|
## Merge Config |
|
```yaml |
|
models: |
|
- model: Sao10K/L3-8B-Stheno-v3.2 |
|
- model: Sao10K/L3-8B-Lunaris-v1 |
|
merge_method: slerp |
|
base_model: Sao10K/L3-8B-Stheno-v3.2 |
|
parameters: |
|
t: |
|
- filter: self_attn |
|
value: 0 |
|
- filter: mlp |
|
value: 1 |
|
- value: 0 |
|
dtype: bfloat16 |
|
``` |