Caleb Ellington
commited on
Commit
·
572e7a9
1
Parent(s):
31d1df4
initial upload
Browse files- config.json +32 -0
- generation_config.json +5 -0
- pytorch_model-00002-of-00013.bin +3 -0
- pytorch_model-00003-of-00013.bin +3 -0
- pytorch_model-00004-of-00013.bin +3 -0
- pytorch_model-00005-of-00013.bin +3 -0
- pytorch_model-00007-of-00013.bin +3 -0
- pytorch_model-00008-of-00013.bin +3 -0
- pytorch_model-00009-of-00013.bin +3 -0
- pytorch_model-00010-of-00013.bin +3 -0
- pytorch_model.bin.index.json +0 -0
config.json
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_linear_bias": true,
|
3 |
+
"architectures": [
|
4 |
+
"FM4BioForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.0,
|
7 |
+
"experts_per_token": 2,
|
8 |
+
"hidden_act": "swiglu",
|
9 |
+
"hidden_dropout_prob": 0.0,
|
10 |
+
"hidden_size": 2304,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 7680,
|
13 |
+
"layer_norm_eps": 1e-05,
|
14 |
+
"max_position_embeddings": 2048,
|
15 |
+
"model_type": "fm4bio",
|
16 |
+
"moe": true,
|
17 |
+
"normalization_type": "RMSNorm",
|
18 |
+
"num_attention_heads": 36,
|
19 |
+
"num_experts": 8,
|
20 |
+
"num_hidden_layers": 36,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"position_embedding_type": "rope",
|
23 |
+
"rotary_percent": 1.0,
|
24 |
+
"seq_len_interpolation_factor": null,
|
25 |
+
"tie_word_embeddings": false,
|
26 |
+
"torch_dtype": "float32",
|
27 |
+
"transformers_version": "4.38.0.dev0",
|
28 |
+
"type_vocab_size": 2,
|
29 |
+
"use_cache": true,
|
30 |
+
"use_lm_head": false,
|
31 |
+
"vocab_size": 128
|
32 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"pad_token_id": 0,
|
4 |
+
"transformers_version": "4.38.0.dev0"
|
5 |
+
}
|
pytorch_model-00002-of-00013.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e9a99e2f63551e07dbe0a34e33ad2bd7b96c8a5ba914d1a1b9cfca4c960dceb
|
3 |
+
size 4928193593
|
pytorch_model-00003-of-00013.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8449e8c6b8026b4338c328f2d032a275924621a9e0361be74d37ca45e8cc1d25
|
3 |
+
size 4928193521
|
pytorch_model-00004-of-00013.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f84de12149bfb3d52471842b7ebb972186f67b19851bbcee6cd1297dba6496ce
|
3 |
+
size 1161822208
|
pytorch_model-00005-of-00013.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2d8de2960b47bf93ba68fe83e93ca7977ba48d4faf232e1d6f874c7ace20164
|
3 |
+
size 4984746078
|
pytorch_model-00007-of-00013.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15638d62de5299486387575df3dc48625e7d224a82c131b06b5f8973a8972131
|
3 |
+
size 4928193673
|
pytorch_model-00008-of-00013.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e28a13b7d9245d0c075f256e02ce0a6cdd492430e2459582fd59d3a9e788200
|
3 |
+
size 4928193649
|
pytorch_model-00009-of-00013.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5b1fb269f9a5cbae9b69dfd77062ec87a041b29ce95a7beaa534ca493a2616bb
|
3 |
+
size 4984746078
|
pytorch_model-00010-of-00013.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6d3bb01148ae5343a9edbf2383447feb49f311e1700a3ec5ba78a6b1f5225ab
|
3 |
+
size 4998982555
|
pytorch_model.bin.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|