LoneStriker commited on
Commit
712f880
Β·
verified Β·
1 Parent(s): 5afba0b

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -1,35 +1,5 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ AlphaMonarch-7B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
2
+ AlphaMonarch-7B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
3
+ AlphaMonarch-7B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
4
+ AlphaMonarch-7B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
5
+ AlphaMonarch-7B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
AlphaMonarch-7B-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91e92331dc34ad697e34801d57b9d9f504bf1c289e91d7a711cb9ea11ce3cd12
3
+ size 3822024672
AlphaMonarch-7B-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3a38583d7737f3f1ccf43ba378dd3fdfb626bac9519f7ad0754afc8cc1ca824
3
+ size 4368439264
AlphaMonarch-7B-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38bbbaa87e309bf3b4d35704c82a5c701cdebe28ec8ec16aee8458222f22829d
3
+ size 5131409376
AlphaMonarch-7B-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9551a2251b94244006dac073c6d472529f711881ea19d63953d977ddf7ed884a
3
+ size 5942065120
AlphaMonarch-7B-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69d1cf29382408c973a68bead59167883af5b4bcfa6f0b1815c7dd44f5009b0b
3
+ size 7695857632
README.md ADDED
@@ -0,0 +1,136 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ tags:
4
+ - merge
5
+ - lazymergekit
6
+ - dpo
7
+ - rlhf
8
+ dataset:
9
+ - mlabonne/truthy-dpo-v0.1
10
+ - mlabonne/distilabel-intel-orca-dpo-pairs
11
+ - mlabonne/chatml-OpenHermes2.5-dpo-binarized-alpha
12
+ base_model:
13
+ - mlabonne/NeuralMonarch-7B
14
+ language:
15
+ - en
16
+ ---
17
+
18
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/TI7C8F2gk43gmI9U2L0uk.jpeg)
19
+
20
+ # πŸ‘‘ AlphaMonarch-7B
21
+
22
+ **tl;dr: AlphaMonarch-7B is a new DPO merge that retains all the reasoning abilities of the very best merges and significantly improves its conversational abilities. Kind of the best of both worlds in a 7B model. πŸŽ‰**
23
+
24
+ AlphaMonarch-7B is a DPO fine-tuned of [mlabonne/NeuralMonarch-7B](https://huggingface.co/mlabonne/NeuralMonarch-7B/) using the [argilla/OpenHermes2.5-dpo-binarized-alpha](https://huggingface.co/datasets/argilla/OpenHermes2.5-dpo-binarized-alpha) preference dataset.
25
+
26
+ It is based on a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
27
+ * [mlabonne/OmniTruthyBeagle-7B-v0](https://huggingface.co/mlabonne/OmniTruthyBeagle-7B-v0)
28
+ * [mlabonne/NeuBeagle-7B](https://huggingface.co/mlabonne/NeuBeagle-7B)
29
+ * [mlabonne/NeuralOmniBeagle-7B](https://huggingface.co/mlabonne/NeuralOmniBeagle-7B)
30
+
31
+ Special thanks to [Jon Durbin](https://huggingface.co/jondurbin), [Intel](https://huggingface.co/Intel), [Argilla](https://huggingface.co/argilla), and [Teknium](https://huggingface.co/teknium) for the preference datasets.
32
+
33
+ **Try the demo**: https://huggingface.co/spaces/mlabonne/AlphaMonarch-7B-GGUF-Chat
34
+
35
+ ## πŸ” Applications
36
+
37
+ This model uses a context window of 8k. I recommend using it with the Mistral Instruct chat template (works perfectly with LM Studio).
38
+
39
+ It is one of the very best 7B models in terms of instructing following and reasoning abilities and can be used for conversations, RP, and storytelling. Note that it tends to have a quite formal and sophisticated style, but it can be changed by modifying the prompt.
40
+
41
+ ## ⚑ Quantized models
42
+
43
+ * **GGUF**: https://huggingface.co/mlabonne/AlphaMonarch-7B-GGUF
44
+
45
+ ## πŸ† Evaluation
46
+
47
+ ### Nous
48
+
49
+ AlphaMonarch-7B is the best-performing 7B model on Nous' benchmark suite (evaluation performed using [LLM AutoEval](https://github.com/mlabonne/llm-autoeval)). See the entire leaderboard [here](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard).
50
+
51
+ | Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
52
+ |---|---:|---:|---:|---:|---:|
53
+ | [**AlphaMonarch-7B**](https://huggingface.co/mlabonne/AlphaMonarch-7B) [πŸ“„](https://gist.github.com/mlabonne/1d33c86824b3a11d2308e36db1ba41c1) | **62.74** | **45.37** | **77.01** | **78.39** | **50.2** |
54
+ | [NeuralMonarch-7B](https://huggingface.co/mlabonne/NeuralMonarch-7B) [πŸ“„](https://gist.github.com/mlabonne/64050c96c6aa261a8f5b403190c8dee4) | 62.73 | 45.31 | 76.99 | 78.35 | 50.28 |
55
+ | [Monarch-7B](https://huggingface.co/mlabonne/Monarch-7B) [πŸ“„](https://gist.github.com/mlabonne/0b8d057c5ece41e0290580a108c7a093) | 62.68 | 45.48 | 77.07 | 78.04 | 50.14 |
56
+ | [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) [πŸ“„](https://gist.github.com/mlabonne/88b21dd9698ffed75d6163ebdc2f6cc8) | 52.42 | 42.75 | 72.99 | 52.99 | 40.94 |
57
+ | [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B) [πŸ“„](https://gist.github.com/mlabonne/14687f1eb3425b166db511f31f8e66f6) | 53.51 | 43.67 | 73.24 | 55.37 | 41.76 |
58
+ | [mlabonne/NeuralBeagle14-7B](https://huggingface.co/mlabonne/NeuralBeagle14-7B) [πŸ“„](https://gist.github.com/mlabonne/ad0c665bbe581c8420136c3b52b3c15c) | 60.25 | 46.06 | 76.77 | 70.32 | 47.86 |
59
+ | [mlabonne/NeuralOmniBeagle-7B](https://huggingface.co/mlabonne/NeuralOmniBeagle-7B) [πŸ“„](https://gist.github.com/mlabonne/0e49d591787185fa5ae92ca5d9d4a1fd) | 62.3 | 45.85 | 77.26 | 76.06 | 50.03 |
60
+ | [eren23/dpo-binarized-NeuralTrix-7B](https://huggingface.co/eren23/dpo-binarized-NeuralTrix-7B) [πŸ“„](https://gist.github.com/CultriX-Github/dbdde67ead233df0c7c56f1b091f728c) | 62.5 | 44.57 | 76.34 | 79.81 | 49.27 |
61
+ | [CultriX/NeuralTrix-7B-dpo](https://huggingface.co/CultriX/NeuralTrix-7B-dpo) [πŸ“„](https://gist.github.com/CultriX-Github/df0502599867d4043b45d9dafb5976e8) | 62.5 | 44.61 | 76.33 | 79.8 | 49.24 |
62
+
63
+ ### EQ-bench
64
+
65
+ AlphaMonarch-7B is also outperforming 70B and 120B parameter models on [EQ-bench](https://eqbench.com/) by [Samuel J. Paech](https://twitter.com/sam_paech), who kindly ran the evaluations.
66
+
67
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/dnCFxieqLiAC3Ll6CfdZW.png)
68
+
69
+ ### MT-Bench
70
+
71
+ ```
72
+ ########## First turn ##########
73
+ score
74
+ model turn
75
+ gpt-4 1 8.95625
76
+ OmniBeagle-7B 1 8.31250
77
+ AlphaMonarch-7B 1 8.23750
78
+ claude-v1 1 8.15000
79
+ NeuralMonarch-7B 1 8.09375
80
+ gpt-3.5-turbo 1 8.07500
81
+ claude-instant-v1 1 7.80000
82
+
83
+ ########## Second turn ##########
84
+ score
85
+ model turn
86
+ gpt-4 2 9.025000
87
+ claude-instant-v1 2 8.012658
88
+ OmniBeagle-7B 2 7.837500
89
+ gpt-3.5-turbo 2 7.812500
90
+ claude-v1 2 7.650000
91
+ AlphaMonarch-7B 2 7.618750
92
+ NeuralMonarch-7B 2 7.375000
93
+
94
+ ########## Average ##########
95
+ score
96
+ model
97
+ gpt-4 8.990625
98
+ OmniBeagle-7B 8.075000
99
+ gpt-3.5-turbo 7.943750
100
+ AlphaMonarch-7B 7.928125
101
+ claude-instant-v1 7.905660
102
+ claude-v1 7.900000
103
+ NeuralMonarch-7B 7.734375
104
+ NeuralBeagle14-7B 7.628125
105
+ ```
106
+
107
+ ### Open LLM Leaderboard
108
+
109
+ AlphaMonarch-7B is one of the best-performing non-merge 7B models on the Open LLM Leaderboard:
110
+
111
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/njHxX_ERQaBssHqp17fMy.png)
112
+
113
+ ## πŸ’» Usage
114
+
115
+ ```python
116
+ !pip install -qU transformers accelerate
117
+
118
+ from transformers import AutoTokenizer
119
+ import transformers
120
+ import torch
121
+
122
+ model = "mlabonne/AlphaMonarch-7B"
123
+ messages = [{"role": "user", "content": "What is a large language model?"}]
124
+
125
+ tokenizer = AutoTokenizer.from_pretrained(model)
126
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
127
+ pipeline = transformers.pipeline(
128
+ "text-generation",
129
+ model=model,
130
+ torch_dtype=torch.float16,
131
+ device_map="auto",
132
+ )
133
+
134
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
135
+ print(outputs[0]["generated_text"])
136
+ ```