mpasila's picture
Clarified quantization type
b7bf76b verified
metadata
license: llama2
language:
  - en

This is an ExLlamaV2 quantized model in 3bpw of jebcarter/Psyfighter-13B using PIPPA as the calibration dataset. (I might reupload this or add a new branch with a different version that uses the default calibration dataset instead)

Alpaca and ChatML prompt templates seem to work fine with this model.

Original Model card

merge_method: task_arithmaetic
base_model: TheBloke/Llama-2-13B-fp16
models:
  - model: TheBloke/Llama-2-13B-fp16
  - model: KoboldAI/LLaMA2-13B-Tiefighter
    parameters:
      weight: 1.0
  - model: chaoyi-wu/MedLLaMA_13B
    parameters:
      weight: 0.01
  - model: Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged
    parameters:
      weight: 0.02
dtype: float16

This model was made possible thanks to the Compute provided by the KoboldAI community.