AugustGislerudRolfsen commited on
Commit
3cca85e
·
verified ·
1 Parent(s): f52abef

Model save

Browse files
Files changed (1) hide show
  1. README.md +4 -8
README.md CHANGED
@@ -14,7 +14,7 @@ model-index:
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/august-gislerud-rolfsen-relu/Dummy/runs/yp8679pw)
18
  # fine-tuned-visionllama
19
 
20
  This model is a fine-tuned version of [meta-llama/Llama-3.2-11B-Vision-Instruct](https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct) on an unknown dataset.
@@ -37,19 +37,15 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0002
40
- - train_batch_size: 1
41
  - eval_batch_size: 8
42
  - seed: 42
43
  - gradient_accumulation_steps: 2
44
- - total_train_batch_size: 2
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: constant
47
  - lr_scheduler_warmup_ratio: 0.03
48
- - num_epochs: 0.01
49
-
50
- ### Training results
51
-
52
-
53
 
54
  ### Framework versions
55
 
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/august-gislerud-rolfsen-relu/Dummy/runs/pbz15hgu)
18
  # fine-tuned-visionllama
19
 
20
  This model is a fine-tuned version of [meta-llama/Llama-3.2-11B-Vision-Instruct](https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct) on an unknown dataset.
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0002
40
+ - train_batch_size: 2
41
  - eval_batch_size: 8
42
  - seed: 42
43
  - gradient_accumulation_steps: 2
44
+ - total_train_batch_size: 4
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: constant
47
  - lr_scheduler_warmup_ratio: 0.03
48
+ - num_epochs: 0.001
 
 
 
 
49
 
50
  ### Framework versions
51