Edit model card

Gen Settings & Prompting

https://rentry.org/tsukasamodel

Training

axolotl was used for training on a 4x nvidia a100 gpu cluster.

the a100 GPU cluster has been graciously provided by lloorree.

rank 16 qlora (all modules) tune

base model mistralai/Mixtral-8x7B-v0.1 tuned on koishi commit 6e675d1 for one epoch

then tuned on pippa 6412b0c for one epoch (metharme completion)

then tuned on limarp Version 2023-10-19 for 2 epochs in metharme completion format with limit_data_length set to 32768 in dataprepare-templates.py

Downloads last month
8
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train ludis/tsukasa-8x7b-qlora-gptq

Collection including ludis/tsukasa-8x7b-qlora-gptq