Update README.md
Browse files
README.md
CHANGED
@@ -20,6 +20,9 @@ license: apache-2.0
|
|
20 |
This model was converted to GGUF format from [`crestf411/Q2.5-32B-Slush`](https://huggingface.co/crestf411/Q2.5-32B-Slush) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
21 |
Refer to the [original model card](https://huggingface.co/crestf411/Q2.5-32B-Slush) for more details on the model.
|
22 |
|
|
|
|
|
|
|
23 |
Slush is a two-stage model trained with high LoRA dropout, where stage 1 is a pretraining continuation on the base model, aimed at boosting the model's creativity and writing capabilities. This is then merged into the instruction tune model, and stage 2 is a fine tuning step on top of this to further enhance its roleplaying capabilities and/or to repair any damage caused in the stage 1 merge.
|
24 |
|
25 |
This is still early stage. As always, feedback is welcome, and begone if you demand perfection.
|
@@ -84,6 +87,7 @@ parameters:
|
|
84 |
tokenizer_source: Qwen/Qwen2.5-32B-Instruct
|
85 |
dtype: bfloat16
|
86 |
|
|
|
87 |
## Use with llama.cpp
|
88 |
Install llama.cpp through brew (works on Mac and Linux)
|
89 |
|
|
|
20 |
This model was converted to GGUF format from [`crestf411/Q2.5-32B-Slush`](https://huggingface.co/crestf411/Q2.5-32B-Slush) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
21 |
Refer to the [original model card](https://huggingface.co/crestf411/Q2.5-32B-Slush) for more details on the model.
|
22 |
|
23 |
+
---
|
24 |
+
Model details:
|
25 |
+
-
|
26 |
Slush is a two-stage model trained with high LoRA dropout, where stage 1 is a pretraining continuation on the base model, aimed at boosting the model's creativity and writing capabilities. This is then merged into the instruction tune model, and stage 2 is a fine tuning step on top of this to further enhance its roleplaying capabilities and/or to repair any damage caused in the stage 1 merge.
|
27 |
|
28 |
This is still early stage. As always, feedback is welcome, and begone if you demand perfection.
|
|
|
87 |
tokenizer_source: Qwen/Qwen2.5-32B-Instruct
|
88 |
dtype: bfloat16
|
89 |
|
90 |
+
---
|
91 |
## Use with llama.cpp
|
92 |
Install llama.cpp through brew (works on Mac and Linux)
|
93 |
|