mrferr3t commited on
Commit
f8cacb5
·
verified ·
1 Parent(s): dae6de3

End of training

Browse files
Files changed (2) hide show
  1. README.md +6 -5
  2. adapter_model.bin +1 -1
README.md CHANGED
@@ -18,6 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
  axolotl version: `0.4.1`
19
  ```yaml
20
  adapter: lora
 
21
  base_model: katuni4ka/tiny-random-dbrx
22
  bf16: auto
23
  chat_template: llama3
@@ -65,7 +66,7 @@ lora_r: 8
65
  lora_target_linear: true
66
  lr_scheduler: cosine
67
  max_steps: 99
68
- micro_batch_size: 16
69
  mlflow_experiment_name: /tmp/f1fb178455e95e80_train_data.json
70
  model_type: AutoModelForCausalLM
71
  num_epochs: 1
@@ -122,8 +123,8 @@ More information needed
122
 
123
  The following hyperparameters were used during training:
124
  - learning_rate: 0.0005
125
- - train_batch_size: 16
126
- - eval_batch_size: 16
127
  - seed: 42
128
  - optimizer: Use adamw_bnb_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
129
  - lr_scheduler_type: cosine
@@ -134,8 +135,8 @@ The following hyperparameters were used during training:
134
 
135
  | Training Loss | Epoch | Step | Validation Loss |
136
  |:-------------:|:------:|:----:|:---------------:|
137
- | No log | 0.0031 | 1 | 11.5 |
138
- | 11.5 | 0.1553 | 50 | 11.5 |
139
 
140
 
141
  ### Framework versions
 
18
  axolotl version: `0.4.1`
19
  ```yaml
20
  adapter: lora
21
+ auto_find_batch_size: true
22
  base_model: katuni4ka/tiny-random-dbrx
23
  bf16: auto
24
  chat_template: llama3
 
66
  lora_target_linear: true
67
  lr_scheduler: cosine
68
  max_steps: 99
69
+ micro_batch_size: 8
70
  mlflow_experiment_name: /tmp/f1fb178455e95e80_train_data.json
71
  model_type: AutoModelForCausalLM
72
  num_epochs: 1
 
123
 
124
  The following hyperparameters were used during training:
125
  - learning_rate: 0.0005
126
+ - train_batch_size: 8
127
+ - eval_batch_size: 8
128
  - seed: 42
129
  - optimizer: Use adamw_bnb_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
130
  - lr_scheduler_type: cosine
 
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
+ | No log | 0.0016 | 1 | 11.5 |
139
+ | 11.5 | 0.0778 | 50 | 11.5 |
140
 
141
 
142
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9fb695301b465ad407f443dcc4d464d6d6d49197785c634716886725f9f1efae
3
  size 9170
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e3a66310ef2cc438504a3240fb8d6729c1e6b8af05c667a166b8fd1ebb2dd6f
3
  size 9170