ali619 commited on
Commit
5e57e53
·
verified ·
1 Parent(s): a7087da

Training in progress, step 100, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -199,4 +199,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
- - PEFT 0.13.0
 
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
+ - PEFT 0.14.0
last-checkpoint/adapter_config.json CHANGED
@@ -3,6 +3,8 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "unsloth/Llama-3.2-1B-Instruct-bnb-4bit",
5
  "bias": "none",
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
@@ -11,6 +13,7 @@
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 16,
 
14
  "lora_dropout": 0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
@@ -20,13 +23,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
- "up_proj",
25
  "v_proj",
26
- "gate_proj",
 
27
  "q_proj",
28
  "k_proj",
29
- "down_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "unsloth/Llama-3.2-1B-Instruct-bnb-4bit",
5
  "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
  "fan_in_fan_out": false,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
 
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
  "lora_alpha": 16,
16
+ "lora_bias": false,
17
  "lora_dropout": 0,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
26
  "v_proj",
27
+ "up_proj",
28
+ "down_proj",
29
  "q_proj",
30
  "k_proj",
31
+ "gate_proj",
32
+ "o_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e360210c7fe6bc9c91e079ae013f535381a6a13f5485dae05bd22c5d7e5f393
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a92839096cdccaad56ee91ea5f7994c6be6ba9d33a99134e5ed273129ef18ef0
3
  size 45118424
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2e1b5838cd931aa02079ffe7f652864e12610c4e917f3e31958ce7debf53515
3
  size 23159290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9402c026f3d7e964601394b7516d2092189b94d8b08e5b09f4062426fb8e7bb
3
  size 23159290
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0bc5c637ed138ad6c1b0f857643b0f621a7e8c72bf01723b5254644a04aa9b48
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e0bfbbe2be6a1399a4c175bf33e1a680e875704102055b2bbfcb74670a4c677
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.00303342838075593,
5
  "eval_steps": 500,
6
  "global_step": 100,
7
  "is_hyper_param_search": false,
@@ -9,15 +9,15 @@
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.00303342838075593,
13
- "grad_norm": 3.9159162044525146,
14
- "learning_rate": 0.002,
15
- "loss": 3.4799,
16
  "step": 100
17
  }
18
  ],
19
  "logging_steps": 100,
20
- "max_steps": 32966,
21
  "num_input_tokens_seen": 0,
22
  "num_train_epochs": 1,
23
  "save_steps": 100,
@@ -33,7 +33,7 @@
33
  "attributes": {}
34
  }
35
  },
36
- "total_flos": 7833024320274432.0,
37
  "train_batch_size": 8,
38
  "trial_name": null,
39
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.07417022065640645,
5
  "eval_steps": 500,
6
  "global_step": 100,
7
  "is_hyper_param_search": false,
 
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.07417022065640645,
13
+ "grad_norm": 1.151120901107788,
14
+ "learning_rate": 0.0002,
15
+ "loss": 3.5121,
16
  "step": 100
17
  }
18
  ],
19
  "logging_steps": 100,
20
+ "max_steps": 1348,
21
  "num_input_tokens_seen": 0,
22
  "num_train_epochs": 1,
23
  "save_steps": 100,
 
33
  "attributes": {}
34
  }
35
  },
36
+ "total_flos": 3.87089027825664e+16,
37
  "train_batch_size": 8,
38
  "trial_name": null,
39
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6fb6b6e3adc261d0a7885521544cee5190193ba91f162fdd596a9773bd61697
3
  size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e21f2a882f703e54677b6d29c5166af27a38da370fe6b9685b41d457c46a3d9d
3
  size 5688