Model Card for Mistral-7B-v0.2

The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) was fine-tuned on top of Mistral-7B-v0.2.

Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1

  • 32k context window (vs 8k context in v0.1)
  • Rope-theta = 1e6
  • No Sliding-Window Attention

For full details of this model please read our paper and release blog post.

Downloads last month
11
Safetensors
Model size
7.24B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MaziyarPanahi/Mistral-7B-v0.2

Quantizations
2 models