aariz120 commited on
Commit
1f2585a
·
verified ·
1 Parent(s): 4ca789d

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e99b93475b2bcc22e6bec389a1f84e25874c0a596ef91804dc915ee65e30bde3
3
  size 9022864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddbd17f7b937a7917fca4cb4b24ddaec8265a8d0746222c758747c6bd56dd0d2
3
  size 9022864
runs/May28_05-08-01_11a1ae0cd486/events.out.tfevents.1716872882.11a1ae0cd486.743.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e29c662e46b7c1f46d4eab6ad453f43ad39adf26660fb230163bbe172453eca
3
+ size 6128
tokenizer_config.json CHANGED
@@ -31,7 +31,6 @@
31
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
- "legacy": false,
35
  "model_max_length": 2048,
36
  "pad_token": "</s>",
37
  "padding_side": "right",
 
31
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
 
34
  "model_max_length": 2048,
35
  "pad_token": "</s>",
36
  "padding_side": "right",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9236ec1bbec1c8fdeb20881ee86c546ce5c866f857bfeb231ab6b1262fd4012
3
- size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56f1978054c125d3fcc8974e3456c2d6f0a030758a4f5d2e58ac97fc82dda680
3
+ size 5112