mingxilei commited on
Commit
be9f880
·
verified ·
1 Parent(s): 0db728b

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -1,8 +1,13 @@
1
  {
2
  "alpha_pattern": {},
3
- "auto_mapping": null,
 
 
 
4
  "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
5
  "bias": "none",
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
@@ -10,20 +15,21 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 32,
14
- "lora_dropout": 0.05,
 
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 16,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
- "task_type": "CAUSAL_LM",
27
  "use_dora": false,
28
  "use_rslora": false
29
  }
 
1
  {
2
  "alpha_pattern": {},
3
+ "auto_mapping": {
4
+ "base_model_class": "LlamaForCausalLM",
5
+ "parent_library": "transformers.models.llama.modeling_llama"
6
+ },
7
  "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
8
  "bias": "none",
9
+ "eva_config": null,
10
+ "exclude_modules": null,
11
  "fan_in_fan_out": false,
12
  "inference_mode": true,
13
  "init_lora_weights": true,
 
15
  "layers_pattern": null,
16
  "layers_to_transform": null,
17
  "loftq_config": {},
18
+ "lora_alpha": 16,
19
+ "lora_bias": false,
20
+ "lora_dropout": 0.0,
21
  "megatron_config": null,
22
  "megatron_core": "megatron.core",
23
  "modules_to_save": null,
24
  "peft_type": "LORA",
25
+ "r": 8,
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "q_proj",
30
+ "v_proj"
31
  ],
32
+ "task_type": null,
33
  "use_dora": false,
34
  "use_rslora": false
35
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bc979d8f45d3a68b1fb596defd32fa57a9350e5cc1c3cd5638c1b0a347b6a54
3
- size 33571624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a6a7db52feb0622cd0c0879c56a47d3b77955e9f74b5c261452aa39e1665f22
3
+ size 16794200
runs/Feb07_11-24-11_autodl-container-ae2d4d944b-64da26b0/events.out.tfevents.1738898662.autodl-container-ae2d4d944b-64da26b0.1665.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09ba8ce926ea590888567b341fa611027b129cd0603f043e84d56ff6e2f05e78
3
+ size 5161
runs/Feb07_11-25-04_autodl-container-ae2d4d944b-64da26b0/events.out.tfevents.1738898716.autodl-container-ae2d4d944b-64da26b0.1833.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97ab01abe2c0af56ac4a52760304eefb303524c30580b8330afb3b8ca7bcb4eb
3
+ size 286549
runs/Feb07_11-31-54_autodl-container-ae2d4d944b-64da26b0/events.out.tfevents.1738899123.autodl-container-ae2d4d944b-64da26b0.2083.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb27deab953c0942970635d74cea290e43e2c92f22f974a4eadf2ed5edb0ee34
3
+ size 14476
runs/Feb07_11-32-52_autodl-container-ae2d4d944b-64da26b0/events.out.tfevents.1738899183.autodl-container-ae2d4d944b-64da26b0.2219.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce452010389b0069f332b74f6df9d5eaaf9d6f3b73722f2af9a1f0f2e1b129c8
3
+ size 9508
runs/Feb07_11-37-42_autodl-container-ae2d4d944b-64da26b0/events.out.tfevents.1738899472.autodl-container-ae2d4d944b-64da26b0.2403.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc9bba4d21289b482f671d5527c937ffad82e947d5bc67ca607c07a73f632aec
3
+ size 73439
runs/Feb07_11-46-37_autodl-container-ae2d4d944b-64da26b0/events.out.tfevents.1738900006.autodl-container-ae2d4d944b-64da26b0.2668.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf0b06eed5895eb1a9478ba2a7838652d337b23d7311d7e15e9922bb514a7c8a
3
+ size 136145
runs/Feb07_12-02-54_autodl-container-ae2d4d944b-64da26b0/events.out.tfevents.1738900984.autodl-container-ae2d4d944b-64da26b0.3043.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f5ec6fba81aac983c95bccabd9e9e36c09c87e4677730bbc4b5222740d87346
3
+ size 110192
special_tokens_map.json CHANGED
@@ -13,7 +13,6 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "</s>",
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
16
  "unk_token": {
17
  "content": "<unk>",
18
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -34,7 +34,7 @@
34
  "extra_special_tokens": {},
35
  "legacy": false,
36
  "model_max_length": 1000000000000000019884624838656,
37
- "pad_token": "</s>",
38
  "padding_side": "right",
39
  "sp_model_kwargs": {},
40
  "tokenizer_class": "LlamaTokenizer",
 
34
  "extra_special_tokens": {},
35
  "legacy": false,
36
  "model_max_length": 1000000000000000019884624838656,
37
+ "pad_token": null,
38
  "padding_side": "right",
39
  "sp_model_kwargs": {},
40
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c475dd4b8c9f09f7d2e2020d9060e7404ece41105bbfcad9d9739bd67c7b37a4
3
- size 5816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f781d828578bf6b12bdd9b91a328bf105f029202a0ff7eaf0c08d29afede3b7e
3
+ size 5496