lm1-misc-pile / 221m60b60b /evaluation /rankeval /lm1-221m-60b-results_lm-eval_global_step115203_2023-01-23-18-55-40_1shots.csv
Muennighoff's picture
Add
c305798
task,metric,value,err,version
anli_r1,acc,0.313,0.014671272822977886,0
anli_r2,acc,0.317,0.014721675438880212,0
anli_r3,acc,0.3358333333333333,0.013639261190932887,0
arc_challenge,acc,0.19368600682593856,0.01154842540997854,0
arc_challenge,acc_norm,0.23208191126279865,0.012336718284948854,0
arc_easy,acc,0.45580808080808083,0.01021963176343785,0
arc_easy,acc_norm,0.42508417508417506,0.010143966195717845,0
boolq,acc,0.5088685015290519,0.008743679265456024,1
cb,acc,0.42857142857142855,0.06672848092813058,1
cb,f1,0.265993265993266,,1
copa,acc,0.63,0.04852365870939099,0
hellaswag,acc,0.2898824935271858,0.00452780401625378,0
hellaswag,acc_norm,0.31736705835490936,0.004645003662067885,0
piqa,acc,0.6322089227421109,0.011250616646678795,0
piqa,acc_norm,0.6294885745375408,0.011267826475447665,0
rte,acc,0.4729241877256318,0.030052303463143706,0
sciq,acc,0.804,0.012559527926707387,0
sciq,acc_norm,0.776,0.013190830072364466,0
storycloze_2016,acc,0.5863174772848744,0.011388832439704382,0
winogrande,acc,0.5098658247829518,0.014049749833367585,0