Edit model card

This model is a GPTQ-quantized version of the Codestral-22B-v0.1 model with the FIM fix, designed to provide improved performance and efficiency. The original Codestral-22B-v0.1 model had a tokenization error for the FIM use case, which was addressed by the legraphista/Codestral-22B-v0.1-hf-FIM-fix model. This GPTQ-quantized model builds upon the fixed model, using default AutoGPTQ arguments and 4 bits per weight.

Compatible with vllm

Alternative to dan-kwiat/Codestral-22B-v0.1-hf-FIM-fix-awq

Tips for Code Completion

When using this model or any other Codestral-22B-v0.1 model for code completion in VSCode on Windows, make sure to use LF (Line Feed) instead of CRLF (Carriage Return Line Feed) for better results. This will prevent unrelated '\' characters from appearing in completions.

Downloads last month
189
Safetensors
Model size
3.33B params
Tensor type
I32
·
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.