jphme commited on
Commit
679eef5
·
verified ·
1 Parent(s): e428ecc

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "ellamind/qwen2-72b-instruct-awq-de",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
@@ -9,7 +9,7 @@
9
  "hidden_act": "silu",
10
  "hidden_size": 8192,
11
  "initializer_range": 0.02,
12
- "intermediate_size": 29568,
13
  "max_position_embeddings": 32768,
14
  "max_window_layers": 80,
15
  "model_type": "qwen2",
@@ -17,13 +17,8 @@
17
  "num_hidden_layers": 80,
18
  "num_key_value_heads": 8,
19
  "quantization_config": {
20
- "backend": "autoawq",
21
  "bits": 4,
22
- "do_fuse": false,
23
- "exllama_config": null,
24
- "fuse_max_seq_len": null,
25
  "group_size": 128,
26
- "modules_to_fuse": null,
27
  "modules_to_not_convert": null,
28
  "quant_method": "awq",
29
  "version": "gemm",
@@ -31,11 +26,11 @@
31
  },
32
  "rms_norm_eps": 1e-06,
33
  "rope_theta": 1000000.0,
34
- "sliding_window": 131072,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "float16",
37
- "transformers_version": "4.42.3",
38
- "use_cache": false,
39
  "use_sliding_window": false,
40
  "vocab_size": 152064
41
  }
 
1
  {
2
+ "_name_or_path": "Qwen2-72b-Instruct-fixed",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
 
9
  "hidden_act": "silu",
10
  "hidden_size": 8192,
11
  "initializer_range": 0.02,
12
+ "intermediate_size": 29696,
13
  "max_position_embeddings": 32768,
14
  "max_window_layers": 80,
15
  "model_type": "qwen2",
 
17
  "num_hidden_layers": 80,
18
  "num_key_value_heads": 8,
19
  "quantization_config": {
 
20
  "bits": 4,
 
 
 
21
  "group_size": 128,
 
22
  "modules_to_not_convert": null,
23
  "quant_method": "awq",
24
  "version": "gemm",
 
26
  },
27
  "rms_norm_eps": 1e-06,
28
  "rope_theta": 1000000.0,
29
+ "sliding_window": null,
30
  "tie_word_embeddings": false,
31
  "torch_dtype": "float16",
32
+ "transformers_version": "4.43.0.dev0",
33
+ "use_cache": true,
34
  "use_sliding_window": false,
35
  "vocab_size": 152064
36
  }
generation_config.json CHANGED
@@ -10,5 +10,5 @@
10
  "temperature": 0.7,
11
  "top_k": 20,
12
  "top_p": 0.8,
13
- "transformers_version": "4.42.3"
14
  }
 
10
  "temperature": 0.7,
11
  "top_k": 20,
12
  "top_p": 0.8,
13
+ "transformers_version": "4.43.0.dev0"
14
  }
model-00001-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da844731a49a2e22de43eb30cd7a6acd7e9ed5308bcb766e2cd6c737dada8d46
3
- size 4975862192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54879f413b906c601f62022d805fa55091a474f8a68297f0c9e6e3909b3a538f
3
+ size 4984578480
model-00002-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9768e173cfdef4a6c02d3beb6d2b4c8dd5dcf4487e0d656889cb18fdc7c67ffd
3
- size 4890455440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf938ec98046696bff0ce196a3a7a4ce95e3f5bf56213010ac6f280c81ef0a10
3
+ size 4907888016
model-00003-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9c2aa56e1cf64503a7d3b6ce07cdaf25559f0e8ac00b67d5c672c315610c536
3
- size 4981431808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b1132919c11503a1ecd1b414739587ffbf7ab0e44fd944baa9e9d959fea3d32
3
+ size 4999409152
model-00004-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1288b33fd35de53df3cb199180b5e067bf4067f9411d643de52ebabf70650215
3
- size 4972694016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d551ddd7d9f968ba3f27c508a25a8a561b63c5762be888fe16f872627c7d289
3
+ size 4990671360
model-00005-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4376868ec29c7250ddff11798e270d5a0ec8fd174d0e1791a456fcd668768cbf
3
- size 4890422552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67f0c24270edd8b38d5cd4cd0d99367177d398996c47b3f134b90e08c4cfdd2b
3
+ size 4907855128
model-00006-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e27eebb39f5b2eebe964d8d4f3a5ee6864fa391888df4d5aa7e5ebd37938f86a
3
- size 4890455560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0de4b77cc8340386b638898e54deda41022590b0752668893b5b0dd2a3b488a8
3
+ size 4907888136
model-00007-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee8800c8ba73fc80f2ea78b7e6295059aeec7e842c0a75f4fea06cd1e086a9ee
3
- size 4890455560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8882a6ce2543de2a6f84b116570008c5695e5e889ccabc821d08cc13e845eb4b
3
+ size 4907888136
model-00008-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c61493d1ab181e788db3d399722b74dc027974588bb9b7ce32077d5ae32f5693
3
- size 4481817968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00db327344b81ac084d7e74e413ef09f29535289c980dd1eb05eb8318e54ccd9
3
+ size 4498161008
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff