model avg arc truthfulqa hellaswag mmlu
crumb/d1536-250MT-full 30.61 23.29 47.1 26.60 25.45787898435144

trained for 250MT/16GT, compared to GPT-2 suite: between gpt2-124m and gpt2-345m (the params are between as well). 254,883,841 non-embedding parameters, this is a roughly 1Tok/Param initial saturation of weights to be further pretrained with ReLora.

Downloads last month
29
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API does not yet support model repos that contain custom code.