Update README.md
Browse files
README.md
CHANGED
@@ -6,7 +6,19 @@ base_model:
|
|
6 |
- deepseek-ai/DeepSeek-R1
|
7 |
pipeline_tag: text-generation
|
8 |
tags:
|
9 |
-
-
|
10 |
- gguf-connector
|
11 |
---
|
12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
- deepseek-ai/DeepSeek-R1
|
7 |
pipeline_tag: text-generation
|
8 |
tags:
|
9 |
+
- deepseek-r1
|
10 |
- gguf-connector
|
11 |
---
|
12 |
|
13 |
+
# GGUF quantized version of **DeepSeek-R1**
|
14 |
+
|
15 |
+
### review
|
16 |
+
- fixed the error loading message: "unknown pre-tokenizer type: deepseek-r1-qwen"
|
17 |
+
- working with llama architecture
|
18 |
+
|
19 |
+
### run the model
|
20 |
+
use any gguf connector to interact with gguf file(s), i.e., [connector](https://pypi.org/project/gguf-connector/)
|
21 |
+
|
22 |
+
### reference
|
23 |
+
- base model: deepseek-ai/[DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1)
|
24 |
+
- tool used for quantization: [cutter](https://pypi.org/project/gguf-cutter)
|