This model is a GPTQ-quantized version of the Codestral-22B-v0.1 model with the FIM fix, designed to provide improved performance and efficiency. The original Codestral-22B-v0.1 model had a tokenization error for the FIM use case, which was addressed by the legraphista/Codestral-22B-v0.1-hf-FIM-fix model. This GPTQ-quantized model builds upon the fixed model, using default AutoGPTQ arguments and 4 bits per weight.
Compatible with vllm
Alternative to dan-kwiat/Codestral-22B-v0.1-hf-FIM-fix-awq
Tips for Code Completion
When using this model or any other Codestral-22B-v0.1 model for code completion in VSCode on Windows, make sure to use LF (Line Feed) instead of CRLF (Carriage Return Line Feed) for better results. This will prevent unrelated '\' characters from appearing in completions.
- Downloads last month
- 189