Triangle104
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -13,6 +13,50 @@ license: apache-2.0
|
|
13 |
This model was converted to GGUF format from [`SubtleOne/Qwen2.5-32b-Erudite-Writer`](https://huggingface.co/SubtleOne/Qwen2.5-32b-Erudite-Writer) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
14 |
Refer to the [original model card](https://huggingface.co/SubtleOne/Qwen2.5-32b-Erudite-Writer) for more details on the model.
|
15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
## Use with llama.cpp
|
17 |
Install llama.cpp through brew (works on Mac and Linux)
|
18 |
|
|
|
13 |
This model was converted to GGUF format from [`SubtleOne/Qwen2.5-32b-Erudite-Writer`](https://huggingface.co/SubtleOne/Qwen2.5-32b-Erudite-Writer) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
14 |
Refer to the [original model card](https://huggingface.co/SubtleOne/Qwen2.5-32b-Erudite-Writer) for more details on the model.
|
15 |
|
16 |
+
---
|
17 |
+
This model is a merge using Rombos's top-ranked 32b model, based on Qwen 2.5, and merging three creative writing finetunes. The creative content is a serious upgrade over the base it started with and has a much more literary style than the previous Writer model. I won't call it better or worse, merely a very distinct flavor and style. I quite like it, and enjoin you to try it as well. Enjoy!
|
18 |
+
|
19 |
+
Merge Method
|
20 |
+
|
21 |
+
This model was merged using the DELLA merge method using rombodawg/Rombos-LLM-V2.5-Qwen-32b as a base.
|
22 |
+
|
23 |
+
Models Merged
|
24 |
+
|
25 |
+
The following models were included in the merge:
|
26 |
+
|
27 |
+
nbeerbower/Qwen2.5-Gutenberg-Doppel-32B
|
28 |
+
ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3
|
29 |
+
EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2
|
30 |
+
|
31 |
+
Configuration
|
32 |
+
|
33 |
+
The following YAML configuration was used to produce this model:
|
34 |
+
|
35 |
+
base_model: rombodawg/Rombos-LLM-V2.5-Qwen-32b
|
36 |
+
parameters:
|
37 |
+
int8_mask: true
|
38 |
+
rescale: false
|
39 |
+
normalize: true
|
40 |
+
lambda: 1.04
|
41 |
+
epsilon: 0.05
|
42 |
+
dtype: bfloat16
|
43 |
+
tokenizer_source: union
|
44 |
+
merge_method: della
|
45 |
+
models:
|
46 |
+
- model: EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2
|
47 |
+
parameters:
|
48 |
+
weight: [0.40]
|
49 |
+
density: [0.53]
|
50 |
+
- model: nbeerbower/Qwen2.5-Gutenberg-Doppel-32B
|
51 |
+
parameters:
|
52 |
+
weight: [0.30]
|
53 |
+
density: [0.53]
|
54 |
+
- model: ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3
|
55 |
+
parameters:
|
56 |
+
weight: [0.40]
|
57 |
+
density: [0.53]
|
58 |
+
|
59 |
+
---
|
60 |
## Use with llama.cpp
|
61 |
Install llama.cpp through brew (works on Mac and Linux)
|
62 |
|