laquythang commited on
Commit
d543f62
·
verified ·
1 Parent(s): 06167fa

End of training

Browse files
README.md CHANGED
@@ -95,7 +95,7 @@ xformers_attention: true
95
 
96
  This model is a fine-tuned version of [jhflow/mistral7b-lora-multi-turn-v2](https://huggingface.co/jhflow/mistral7b-lora-multi-turn-v2) on the None dataset.
97
  It achieves the following results on the evaluation set:
98
- - Loss: 1.8998
99
 
100
  ## Model description
101
 
@@ -127,7 +127,7 @@ The following hyperparameters were used during training:
127
 
128
  | Training Loss | Epoch | Step | Validation Loss |
129
  |:-------------:|:------:|:----:|:---------------:|
130
- | 1.6939 | 0.0008 | 50 | 1.8998 |
131
 
132
 
133
  ### Framework versions
 
95
 
96
  This model is a fine-tuned version of [jhflow/mistral7b-lora-multi-turn-v2](https://huggingface.co/jhflow/mistral7b-lora-multi-turn-v2) on the None dataset.
97
  It achieves the following results on the evaluation set:
98
+ - Loss: 1.9077
99
 
100
  ## Model description
101
 
 
127
 
128
  | Training Loss | Epoch | Step | Validation Loss |
129
  |:-------------:|:------:|:----:|:---------------:|
130
+ | 1.7459 | 0.0008 | 50 | 1.9077 |
131
 
132
 
133
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
- "gate_proj",
25
- "v_proj",
26
  "up_proj",
27
- "k_proj",
28
  "down_proj",
29
- "q_proj"
 
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "up_proj",
 
24
  "down_proj",
25
+ "q_proj",
26
+ "gate_proj",
27
+ "v_proj",
28
+ "o_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c12eb922bad910df0073e52dda16e4e95361307ce4538b058601f3cfc067484c
3
  size 167934026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2e40d39b5c2e78a2fb2235a676050ee3a170b0171242d15bb760b5e5964e5bf
3
  size 167934026
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4db5b10da495cd3da2aa9f0f75345820c9c373de6c063a83ba78aedb561da428
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c2962af7a1109d6a39532a90feedcc00b1a831b5317ffc2d5dbc69541a9257e
3
  size 167832240
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e6d3575d9d7395536b922b4c7f22e6bc0a972ec39128afa719686c80fe08735
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be7949611575e93d2b5e50db3dfc1ab90e2b762df23a744fa28b764f5bd6127e
3
  size 6776