Update README.md
Browse filesAdd technical report link
README.md
CHANGED
@@ -19,8 +19,8 @@ tokens of text and code data in 8K sequence length. Fox-1 uses Grouped Query Att
|
|
19 |
Fox-1-Instruct-v0.1 is an instruction-tuned (SFT) version of Fox-1-1.6B that has an 8K native context length. The model
|
20 |
was finetuned with 5B tokens of instruction following and multi-turn conversation data.
|
21 |
|
22 |
-
For the full details of this model please read
|
23 |
-
|
24 |
|
25 |
## Getting-Started
|
26 |
|
|
|
19 |
Fox-1-Instruct-v0.1 is an instruction-tuned (SFT) version of Fox-1-1.6B that has an 8K native context length. The model
|
20 |
was finetuned with 5B tokens of instruction following and multi-turn conversation data.
|
21 |
|
22 |
+
For the full details of this model please read [Fox-1 technical report](https://arxiv.org/abs/2411.05281)
|
23 |
+
and [release blog post](https://blog.tensoropera.ai/tensoropera-unveils-fox-foundation-model-a-pioneering-open-source-slm-leading-the-way-against-tech-giants).
|
24 |
|
25 |
## Getting-Started
|
26 |
|