Updated Krutrim 1 IF_eval
Browse files
README.md
CHANGED
@@ -86,7 +86,7 @@ We use the LM Evaluation Harness to evaluate our model on the En benchmarks task
|
|
86 |
| ARC_Challenge (0-shot) - Accuracy | 0.48 | 0.59 | 0.60 | 0.93 (25-shot) | - | 0.50 |
|
87 |
| ARC_Easy (0-shot) - Accuracy | 0.73 | 0.80 | 0.82 | - | - | - |
|
88 |
| HumanEval - Pass@10 | 0.00 | 0.23 | 0.80 | 0.88 | 0.74 (0-shot) | 0.90 |
|
89 |
-
| IF_Eval (0-shot) - Accuracy | 0.
|
90 |
|
91 |
### Indic Benchmarks
|
92 |
|
|
|
86 |
| ARC_Challenge (0-shot) - Accuracy | 0.48 | 0.59 | 0.60 | 0.93 (25-shot) | - | 0.50 |
|
87 |
| ARC_Easy (0-shot) - Accuracy | 0.73 | 0.80 | 0.82 | - | - | - |
|
88 |
| HumanEval - Pass@10 | 0.00 | 0.23 | 0.80 | 0.88 | 0.74 (0-shot) | 0.90 |
|
89 |
+
| IF_Eval (0-shot) - Accuracy | 0.27 | 0.46 | 0.73 | 0.92 | - | 0.84 |
|
90 |
|
91 |
### Indic Benchmarks
|
92 |
|