This is a Llama 2 architecture model series trained on the FineWeb dataset. This is ~500 Million model uses lamma tokenizer. trained using code from Karpathy lamma2

Downloads last month
2
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the HF Inference API does not support Transformers models with pipeline type text-generation

Dataset used to train sabareesh88/fw14k