Base model: https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2

This is just a custom 4bit imatrix quant made to run optiomally on a macbook with 8gb of ram.

For use with llama.cpp https://github.com/ggerganov/llama.cpp

Downloads last month
17
GGUF
Model size
7.24B params
Architecture
llama

16-bit

Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for nisten/mistral-instruct0.2-imatrix4bit.gguf

Quantized
(88)
this model