YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

GGUF quants for : https://huggingface.co/alchemonaut/QuartetAnemoi-70B-t0.0001

Available : Q3_K_M, IQ3_XXS. Otw : IQ2_XS

I recommand you folks to try this model, because it's quite an efficient merge of Miqu, WinterGoddess, AuroraNights, and XWin.

The Theta Rope 1,000,000 of Miqu, hence the 32k context, is functional up to 16k accordingly to my tests, and probably above (I need a smaller quant to test, which is otw).

To use it with a quantized KV cache for a higher context, here's KoboldCPP Frankenstein version with several different KV cache quantization lebels to chose from : https://github.com/Nexesenex/kobold.cpp/releases

Downloads last month
123
GGUF
Model size
69B params
Architecture
llama

1-bit

2-bit

3-bit

4-bit

Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Collection including Nexesenex/alchemonaut_QuartetAnemoi-70B-iMat.GGUF