updated description
Browse files
README.md
CHANGED
@@ -30,6 +30,8 @@ Krutrim-2 is a 12B parameter language model developed by the OLA Krutrim team. I
|
|
30 |
|
31 |
After fine-tuning, the model underwent Direct Preference Optimization (DPO) to enhance alignment across multiple aspects. DPO was applied to improve response helpfulness, safety, and reasoning.
|
32 |
|
|
|
|
|
33 |
## Key Features
|
34 |
- 12B parameter dense transformer model leading to better generalization compared to Krutrim-1 7B;
|
35 |
- Supports context up to 128K tokens making it suitable for long multi-turn conversations, long-form generations, document translations and others;
|
|
|
30 |
|
31 |
After fine-tuning, the model underwent Direct Preference Optimization (DPO) to enhance alignment across multiple aspects. DPO was applied to improve response helpfulness, safety, and reasoning.
|
32 |
|
33 |
+
The model delivers best-in-class performance across Indic tasks and a promising performance on English benchmarks equivalent to models 5-10x the size. We present details of the model architecture, pre-training, post-training and evaluation results. We also publicly release the post-trained versions of the model. We are continuously improving the model through post-training techniques such as RLHF.
|
34 |
+
|
35 |
## Key Features
|
36 |
- 12B parameter dense transformer model leading to better generalization compared to Krutrim-1 7B;
|
37 |
- Supports context up to 128K tokens making it suitable for long multi-turn conversations, long-form generations, document translations and others;
|