FatCat87 commited on
Commit
b0b2af5
·
verified ·
1 Parent(s): ce16e61

End of training

Browse files
Files changed (2) hide show
  1. README.md +17 -19
  2. adapter_model.bin +1 -1
README.md CHANGED
@@ -3,9 +3,9 @@ library_name: peft
3
  tags:
4
  - axolotl
5
  - generated_from_trainer
6
- base_model: oopsung/llama2-7b-n-ox-test-v1
7
  model-index:
8
- - name: 695b6d76-a4f0-4d1e-b337-f6d87e10453d
9
  results: []
10
  ---
11
 
@@ -18,19 +18,19 @@ should probably proofread and complete it, then remove this comment. -->
18
  axolotl version: `0.4.1`
19
  ```yaml
20
  adapter: lora
21
- base_model: oopsung/llama2-7b-n-ox-test-v1
22
  bf16: auto
23
  datasets:
24
  - data_files:
25
- - 35ca5ef06350813d_train_data.json
26
  ds_type: json
27
  format: custom
28
- path: 35ca5ef06350813d_train_data.json
29
  type:
30
  field: null
31
  field_input: null
32
- field_instruction: instruction
33
- field_output: output
34
  field_system: null
35
  format: null
36
  no_input_format: null
@@ -50,7 +50,7 @@ fsdp_config: null
50
  gradient_accumulation_steps: 4
51
  gradient_checkpointing: true
52
  group_by_length: false
53
- hub_model_id: FatCat87/695b6d76-a4f0-4d1e-b337-f6d87e10453d
54
  learning_rate: 0.0002
55
  load_in_4bit: false
56
  load_in_8bit: true
@@ -81,9 +81,9 @@ val_set_size: 0.1
81
  wandb_entity: fatcat87-taopanda
82
  wandb_log_model: null
83
  wandb_mode: online
84
- wandb_name: 695b6d76-a4f0-4d1e-b337-f6d87e10453d
85
  wandb_project: subnet56
86
- wandb_runid: 695b6d76-a4f0-4d1e-b337-f6d87e10453d
87
  wandb_watch: null
88
  warmup_ratio: 0.05
89
  weight_decay: 0.0
@@ -93,12 +93,12 @@ xformers_attention: null
93
 
94
  </details><br>
95
 
96
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/fatcat87-taopanda/subnet56/runs/euufo96d)
97
- # 695b6d76-a4f0-4d1e-b337-f6d87e10453d
98
 
99
- This model is a fine-tuned version of [oopsung/llama2-7b-n-ox-test-v1](https://huggingface.co/oopsung/llama2-7b-n-ox-test-v1) on the None dataset.
100
  It achieves the following results on the evaluation set:
101
- - Loss: 2.0870
102
 
103
  ## Model description
104
 
@@ -128,17 +128,15 @@ The following hyperparameters were used during training:
128
  - total_eval_batch_size: 4
129
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
130
  - lr_scheduler_type: cosine
131
- - lr_scheduler_warmup_steps: 2
132
  - num_epochs: 1
133
 
134
  ### Training results
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
- | 2.8009 | 0.0331 | 1 | 2.9392 |
139
- | 2.309 | 0.2645 | 8 | 2.3503 |
140
- | 2.2402 | 0.5289 | 16 | 2.1435 |
141
- | 2.1873 | 0.7934 | 24 | 2.0870 |
142
 
143
 
144
  ### Framework versions
 
3
  tags:
4
  - axolotl
5
  - generated_from_trainer
6
+ base_model: oopsung/llama2-7b-koNqa-test-v1
7
  model-index:
8
+ - name: 7cf75911-52d7-4b9f-b49f-422caa1e3165
9
  results: []
10
  ---
11
 
 
18
  axolotl version: `0.4.1`
19
  ```yaml
20
  adapter: lora
21
+ base_model: oopsung/llama2-7b-koNqa-test-v1
22
  bf16: auto
23
  datasets:
24
  - data_files:
25
+ - b1db72b837c3bf33_train_data.json
26
  ds_type: json
27
  format: custom
28
+ path: b1db72b837c3bf33_train_data.json
29
  type:
30
  field: null
31
  field_input: null
32
+ field_instruction: question
33
+ field_output: query
34
  field_system: null
35
  format: null
36
  no_input_format: null
 
50
  gradient_accumulation_steps: 4
51
  gradient_checkpointing: true
52
  group_by_length: false
53
+ hub_model_id: FatCat87/7cf75911-52d7-4b9f-b49f-422caa1e3165
54
  learning_rate: 0.0002
55
  load_in_4bit: false
56
  load_in_8bit: true
 
81
  wandb_entity: fatcat87-taopanda
82
  wandb_log_model: null
83
  wandb_mode: online
84
+ wandb_name: 7cf75911-52d7-4b9f-b49f-422caa1e3165
85
  wandb_project: subnet56
86
+ wandb_runid: 7cf75911-52d7-4b9f-b49f-422caa1e3165
87
  wandb_watch: null
88
  warmup_ratio: 0.05
89
  weight_decay: 0.0
 
93
 
94
  </details><br>
95
 
96
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/fatcat87-taopanda/subnet56/runs/ni7922kd)
97
+ # 7cf75911-52d7-4b9f-b49f-422caa1e3165
98
 
99
+ This model is a fine-tuned version of [oopsung/llama2-7b-koNqa-test-v1](https://huggingface.co/oopsung/llama2-7b-koNqa-test-v1) on the None dataset.
100
  It achieves the following results on the evaluation set:
101
+ - Loss: 2.3653
102
 
103
  ## Model description
104
 
 
128
  - total_eval_batch_size: 4
129
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
130
  - lr_scheduler_type: cosine
 
131
  - num_epochs: 1
132
 
133
  ### Training results
134
 
135
  | Training Loss | Epoch | Step | Validation Loss |
136
  |:-------------:|:------:|:----:|:---------------:|
137
+ | 2.6121 | 0.1739 | 1 | 3.0362 |
138
+ | 2.6016 | 0.3478 | 2 | 2.7754 |
139
+ | 2.088 | 0.6957 | 4 | 2.3653 |
 
140
 
141
 
142
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab4f52457a073d6ac822788e123aea2637035d6f1e06680bd9c795f3e5da993d
3
  size 319977674
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9de5d782948a6513fedc80c0d95e8fe0a085689c42d8b95dbcbb6db0a57e9c55
3
  size 319977674