PyTorch
mistral
Krutrim
language-model
krutrim-admin commited on
Commit
0cc38b7
·
verified ·
1 Parent(s): 5f656f6

updated description

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -30,6 +30,8 @@ Krutrim-2 is a 12B parameter language model developed by the OLA Krutrim team. I
30
 
31
  After fine-tuning, the model underwent Direct Preference Optimization (DPO) to enhance alignment across multiple aspects. DPO was applied to improve response helpfulness, safety, and reasoning.
32
 
 
 
33
  ## Key Features
34
  - 12B parameter dense transformer model leading to better generalization compared to Krutrim-1 7B;
35
  - Supports context up to 128K tokens making it suitable for long multi-turn conversations, long-form generations, document translations and others;
 
30
 
31
  After fine-tuning, the model underwent Direct Preference Optimization (DPO) to enhance alignment across multiple aspects. DPO was applied to improve response helpfulness, safety, and reasoning.
32
 
33
+ The model delivers best-in-class performance across Indic tasks and a promising performance on English benchmarks equivalent to models 5-10x the size. We present details of the model architecture, pre-training, post-training and evaluation results. We also publicly release the post-trained versions of the model. We are continuously improving the model through post-training techniques such as RLHF.
34
+
35
  ## Key Features
36
  - 12B parameter dense transformer model leading to better generalization compared to Krutrim-1 7B;
37
  - Supports context up to 128K tokens making it suitable for long multi-turn conversations, long-form generations, document translations and others;