Inquiry Regarding Fine-Tuned LLaMA 3.3 70B and Potential 4-Bit AWQ Quantized Model Release

#1
by caffeineWoo - opened

I hope this message finds you well.

I was excited to learn about the release of the fine-tuned LLaMA 3.3 70B model and its impressive capabilities. The advancements in fine-tuning are inspiring, and I truly appreciate the efforts your team has made in bringing this model to the community.

I wanted to inquire if there are any plans to release a 4-bit AWQ quantized version of this model. Such a release would undoubtedly enable broader accessibility and practical application, especially for those working with hardware-constrained environments.

Your insights would be greatly appreciated, and I look forward to hearing about any updates or plans for the model’s development.

Thank you for your time and consideration, and please let me know if there’s anything further I can assist with or provide feedback on.

한국분이신가요 ?
네 quantization 해서 ollama에 공개 하여 드리겠습니다.
=Yes, I will make it available on Ollama with quantization.
=はい、量子化してollamaで公開させていただきます。

한국분이셨군요. 라마 70B 모델의 일본어 학습 모델의 활용 기대가 큽니다. 감사합니다!!

Sign up or log in to comment