Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,17 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- Intel/orca_dpo_pairs
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
- es
|
8 |
---
|
9 |
+
Barcenas Tiny 1.1b DPO
|
10 |
+
|
11 |
+
It is a model based on the famous TinyLlama/TinyLlama-1.1B-Chat-v1.0 and trained with DPO using the Intel/orca_dpo_pairs dataset.
|
12 |
+
|
13 |
+
With its reinforcement based training we hope to improve the Tiny model in a huge way and have a better model with better responses with a small size and accessible to most people.
|
14 |
+
|
15 |
+
Many thanks to Maxime Labonne (mlabonne) for his tutorial on how to train a LLM model using DPO, without his tutorial this model would not have been possible.
|
16 |
+
|
17 |
+
Made with ❤️ in Guadalupe, Nuevo Leon, Mexico 🇲🇽
|