This is a Llama 2 architecture model series trained on the FineWeb dataset. This is ~500 Million model uses lamma tokenizer. trained using code from Karpathy lamma2
- Downloads last month
- 2
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the HF Inference API does not support Transformers models with pipeline type text-generation