File size: 1,977 Bytes
2863d67 3d04364 2863d67 ba8ef5f 2863d67 b4d546c 2863d67 ba8ef5f 2863d67 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 |
---
language:
- sv
license: llama3.1
library_name: transformers
tags:
- unsloth
datasets:
- neph1/bellman-7b-finetune
- neph1/codefeedback-swedish
---
# Model Card for Bellman
This version of bellman is finetuned from llama-3.1-instruct-8b.
It's finetuned for prompt question answering, based on a dataset created from
Swedish wikipedia, with a lot of Sweden-centric questions.
New from previous versions is questions from a translated code-feedback dataset, as well as a number of stories. It's not great at generating stories,
but better than previosly.
Try out the Q8 version here: https://huggingface.co/spaces/neph1/bellman (cpu)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/653cd3049107029eb004f968/IDGX3d9lGe6yx-yHjsrav.png)
## Model Details
Training run on 240724:
Step Training Loss Validation Loss<br>
25 1.352200 1.034565<br>
50 1.033600 1.009348<br>
75 1.022400 0.996665<br>
100 1.002900 0.988050<br>
125 1.014600 0.981633<br>
150 1.006300 0.975584<br>
175 0.988800 0.970966<br>
200 0.985300 0.967037<br>
225 0.992400 0.964120<br>
250 0.950000 0.962472<br>
275 0.931000 0.960848<br>
300 0.932000 0.958946 <-- picked checkpoint <br>
### Training Parameters
per_device_train_batch_size = 4,<br>
gradient_accumulation_steps = 16,<br>
num_train_epochs=3,<br>
warmup_steps = 5,<br>
learning_rate = 3e-5,<br>
logging_steps = 25,<br>
optim = "adamw_8bit",<br>
weight_decay = 0.01,<br>
lr_scheduler_type = "linear",<br>
seed = 3407,<br>
per_device_eval_batch_size = 2,<br>
eval_strategy="steps",<br>
eval_accumulation_steps = 32,<br>
eval_steps = 25,<br>
eval_delay = 0,<br>
save_strategy="steps",<br>
save_steps=50,<br>
### Model Description
- **Developed by:** Me
- **Funded by:** Me
- **Model type:** Instruct
- **Language(s) (NLP):** Swedish
- **License:** llama-3.1
- **Finetuned from model:** Llama3.1 Instruct 8b
## Model Card Contact
[email protected] |