|
--- |
|
base_model: [meta-llama/Meta-Llama-3.1-405B] |
|
library_name: transformers |
|
tags: |
|
- mergekit |
|
- merge |
|
|
|
--- |
|
|
|
|
|
|
|
# lobotollama-368b prune [Meta-Llama-3.1-405B-Base](https://huggingface.co/meta-llama/Meta-Llama-3.1-405B). |
|
This is a negative-merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). |
|
|
|
# Just so you meow, this did not turn out all that great in the perplexity benchmarks. Needs healing, you'll probably need 32xh100 to do a full finetune. |
|
# Model was designed to fin in a M2 mac-studio 192gb in 4bit. |
|
|
|
```verilog |
|
perplexity: 167.37 seconds per pass - ETA 33.47 minutes - meta-405b-base - q8_0 - newest base was identical in bf16 and q8_0 |
|
[1]1.3927,[2]1.6952,[3]1.5905,[4]1.4674,[5]1.3652,[6]1.3054,[7]1.2885,[8]1.2673,[9]1.2397,[10]1.2179,[11]1.2149,[12]1.2162, |
|
Final estimate: PPL = 1.2162 +/- 0.02128 |
|
|
|
perplexity: 2197.87 seconds per pass - ETA 1 hours 49.88 minutes -- llama 405b - instruct - old BF16 -8head |
|
[1]2.1037,[2]2.4201,[3]2.0992,[4]1.8446,[5]1.6823,[6]1.5948,[7]1.5575,[8]1.5121,[9]1.4750,[10]1.4570,[11]1.4567,[12]1.4666, |
|
Final estimate: PPL = 1.4666 +/- 0.03184 |
|
|
|
./llama-perplexity -m /scratch-10/lobotollama-q8_0.gguf -f wiki.test.raw -t 96 --chunks 12 -b 1024 |
|
perplexity: 331.47 seconds per pass - ETA 33.13 minutes |
|
[1]2.6744,[2]3.4041,[3]2.9683,[4]2.8669,[5]2.7924,[6]2.7590,[7]2.8274,[8]2.8306,[9]2.7943,[10]2.7910,[11]2.8164,[12]2.9396, |
|
Final estimate: PPL = 2.9396 +/- 0.09497 |
|
``` |
|
|
|
|
|
|
|
## Merge Details |
|
### Merge Method |
|
|
|
This model was merged using the passthrough merge method. |
|
|
|
### Models Merged |
|
|
|
The following models were included in the merge: |
|
* /Meta-Llama-3.1-405B |
|
|
|
### Configuration |
|
|
|
The following YAML configuration was used to produce this model: |
|
|
|
```yaml |
|
dtype: bfloat16 |
|
merge_method: passthrough |
|
slices: |
|
- sources: |
|
- layer_range: [0, 29] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [30, 35] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [36, 40] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [41, 45] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [46, 49] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [50, 54] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [55, 59] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [60, 64] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [65, 69] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [70, 74] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [75, 79] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [80, 84] |
|
model: /Meta-Llama-3.1-405B |
|
- sources: |
|
- layer_range: [85, 126] |
|
model: /Meta-Llama-3.1-405B |
|
``` |
|
|