--- base_model: [] library_name: transformers tags: - mergekit - merge ---
MidnightMiqu
# Midnight-Miqu-70B-v1.0 - EXL2 5.0bpw This is a 5.0bpw EXL2 quant of [sophosympatheia/Midnight-Miqu-70B-v1.0](https://huggingface.co/sophosympatheia/Midnight-Miqu-70B-v1.0) Details about the model and the merge info can be found at the above mode page. I have not extensively tested this quant/model other than ensuring I could load it and chat with it. ## Model Loading Below is what I used to run this model on a dual 3090 Linux server. ![image/jpg](Midnight-Miqu-70B-exl2-5-textgen.jpg) I have not tested inference above a couple K tokens. If the model blows out past 8k, consider a 8192 context without cache_8bit set. ## Quant Details This is the script used for quantization. ```bash #!/bin/bash # Activate the conda environment source ~/miniconda3/etc/profile.d/conda.sh conda activate exllamav2 # Define variables MODEL_DIR="models/sophosympatheia_Midnight-Miqu-70B-v1.0" OUTPUT_DIR="exl2_midnight70b" MEASUREMENT_FILE="measurements/midnight70b.json" BIT_PRECISION=5.0 CONVERTED_FOLDER="models/Midnight-Miqu-70B_exl2_5.0bpw" # Create directories mkdir $OUTPUT_DIR mkdir $CONVERTED_FOLDER # Run conversion commands python convert.py -i $MODEL_DIR -o $OUTPUT_DIR -nr -om $MEASUREMENT_FILE python convert.py -i $MODEL_DIR -o $OUTPUT_DIR -nr -m $MEASUREMENT_FILE -b $BIT_PRECISION -cf $CONVERTED_FOLDER ```