This model is a GPTQ-quantized version of the Codestral-22B-v0.1 model with the FIM fix, designed to provide improved performance and efficiency. The original Codestral-22B-v0.1 model had a tokenization error for the FIM use case, which was addressed by the legraphista/Codestral-22B-v0.1-hf-FIM-fix model. This GPTQ-quantized model builds upon the fixed model, using default AutoGPTQ arguments and 4 bits per weight.

Compatible with vllm

Alternative to dan-kwiat/Codestral-22B-v0.1-hf-FIM-fix-awq

Tips for Code Completion

When using this model or any other Codestral-22B-v0.1 model for code completion in VSCode on Windows, make sure to use LF (Line Feed) instead of CRLF (Carriage Return Line Feed) for better results. This will prevent unrelated '\' characters from appearing in completions.

Downloads last month
48
Safetensors
Model size
3.33B params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.