VERSIL91 commited on
Commit
c691026
·
verified ·
1 Parent(s): f7c38e0

End of training

Browse files
Files changed (2) hide show
  1. README.md +6 -5
  2. adapter_model.bin +1 -1
README.md CHANGED
@@ -75,7 +75,7 @@ lora_target_modules:
75
  lr_scheduler: cosine
76
  max_memory:
77
  0: 70GiB
78
- max_steps: 5
79
  micro_batch_size: 2
80
  mlflow_experiment_name: /tmp/1c2d230e0db4aaf5_train_data.json
81
  model_type: AutoModelForCausalLM
@@ -118,7 +118,7 @@ xformers_attention: null
118
 
119
  This model is a fine-tuned version of [Korabbit/llama-2-ko-7b](https://huggingface.co/Korabbit/llama-2-ko-7b) on the None dataset.
120
  It achieves the following results on the evaluation set:
121
- - Loss: 0.8985
122
 
123
  ## Model description
124
 
@@ -146,15 +146,16 @@ The following hyperparameters were used during training:
146
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
147
  - lr_scheduler_type: cosine
148
  - lr_scheduler_warmup_steps: 10
149
- - training_steps: 5
150
 
151
  ### Training results
152
 
153
  | Training Loss | Epoch | Step | Validation Loss |
154
  |:-------------:|:------:|:----:|:---------------:|
155
  | 0.9265 | 0.0024 | 1 | 0.9870 |
156
- | 1.0298 | 0.0048 | 2 | 0.9388 |
157
- | 0.9506 | 0.0095 | 4 | 0.8985 |
 
158
 
159
 
160
  ### Framework versions
 
75
  lr_scheduler: cosine
76
  max_memory:
77
  0: 70GiB
78
+ max_steps: 50
79
  micro_batch_size: 2
80
  mlflow_experiment_name: /tmp/1c2d230e0db4aaf5_train_data.json
81
  model_type: AutoModelForCausalLM
 
118
 
119
  This model is a fine-tuned version of [Korabbit/llama-2-ko-7b](https://huggingface.co/Korabbit/llama-2-ko-7b) on the None dataset.
120
  It achieves the following results on the evaluation set:
121
+ - Loss: 0.0540
122
 
123
  ## Model description
124
 
 
146
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
147
  - lr_scheduler_type: cosine
148
  - lr_scheduler_warmup_steps: 10
149
+ - training_steps: 50
150
 
151
  ### Training results
152
 
153
  | Training Loss | Epoch | Step | Validation Loss |
154
  |:-------------:|:------:|:----:|:---------------:|
155
  | 0.9265 | 0.0024 | 1 | 0.9870 |
156
+ | 0.3231 | 0.0310 | 13 | 0.3032 |
157
+ | 0.1219 | 0.0620 | 26 | 0.0904 |
158
+ | 0.0602 | 0.0930 | 39 | 0.0540 |
159
 
160
 
161
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1508bbce101c2fb481eb76fc83c844edf9e2118f937bbb6efa6c399b6095ed49
3
  size 80115210
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5114fa34ed382549acc5ff56201c6cc1710b082b8968d1e0b63ff44e58ca4d0
3
  size 80115210