Edit model card

4-bit Llammas in gguf

This is a 4-bit quantized version of TartuNLP/Llammas Llama2 model in gguf file format.

Downloads last month
8
GGUF
Model size
6.74B params
Architecture
llama

4-bit

Inference API
Unable to determine this model's library. Check the docs .