mobiuslabsgmbh/Meta-Llama-3-8B-Instruct_4bitgs64_hqq_hf Text Generation • Updated 5 days ago • 16 • 2
DeepSeek-R1-ReDistill Collection Re-distilled DeepSeek R1 models • 4 items • Updated 13 days ago • 12
view post Post 2078 Releasing HQQ Llama-3.1-70b 4-bit quantized version! Check it out at mobiuslabsgmbh/Llama-3.1-70b-instruct_4bitgs64_hqq. Achieves 99% of the base model performance across various benchmarks! Details in the model card. 🔥 8 8 + Reply
view post Post 1792 Excited to announce the release of our high-quality Llama-3.1 8B 4-bit HQQ/calibrated quantized model! Achieving an impressive 99.3% relative performance to FP16, it also delivers the fastest inference speed for transformers. mobiuslabsgmbh/Llama-3.1-8b-instruct_4bitgs64_hqq_calib 1 reply · 🔥 9 9 + Reply
mobiuslabsgmbh/Llama-3.1-8b-instruct_4bitgs64_hqq_calib Text Generation • Updated 6 days ago • 27 • 55