duyphu commited on
Commit
5e910f6
·
verified ·
1 Parent(s): dc08e17

Training in progress, step 13

Browse files
adapter_config.json CHANGED
@@ -22,11 +22,11 @@
22
  "target_modules": [
23
  "v_proj",
24
  "down_proj",
25
- "q_proj",
26
  "k_proj",
 
 
27
  "o_proj",
28
- "up_proj",
29
- "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
22
  "target_modules": [
23
  "v_proj",
24
  "down_proj",
 
25
  "k_proj",
26
+ "q_proj",
27
+ "gate_proj",
28
  "o_proj",
29
+ "up_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:26511612ee7722c9f6777610c9b4d633630fa9340341886d4fe30e3979a1a5a6
3
  size 50899792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:387437f25c3f442fa0739b337e027bfa8f5dcfec3c7c574d82e670e3c4cb863e
3
  size 50899792
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:801e82d92b7a0fcb5504aa4cdd50beb9d418ff099379be3aeb254bf20048e61c
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c902d61e4dc7c442277de529acbaa11a0ae0719ae1a5ba205c36831150aa27ee
3
  size 6776