|
|
|
A set of 50% weight-sparse Llama3.1-8B pruned by [Wanda](https://github.com/locuslab/wanda). |
|
Model links are in the table below. Models can be loaded as is with Huggingface Transformers. |
|
|
|
### Perplexity |
|
![Perplexity over Sparsity](llama3.1-8B_Wanda_sparsity.png) |
|
### MMLU (5-shot) |
|
| MMLU (5-shot) | Accuracy (%) | Relative to Dense (%) | Model Link| |
|
|----------------|--------------|-----------------------|-----------| |
|
| Dense | 65.1 | baseline | [Meta-Llama-3.1-8B-wanda-unstructured-0.0](https://huggingface.co/vuiseng9/Meta-Llama-3.1-8B-wanda-unstructured-0.0) | |
|
| Unstructured | 50.0 | -15.1 | [Meta-Llama-3.1-8B-wanda-unstructured-0.5](https://huggingface.co/vuiseng9/Meta-Llama-3.1-8B-wanda-unstructured-0.5) | |
|
| 4:8 | 39.3 | -25.8 | [Meta-Llama-3.1-8B-wanda-4of8](https://huggingface.co/vuiseng9/Meta-Llama-3.1-8B-wanda-4of8) | |
|
| 2:4 | 28.7 | -36.4 | [Meta-Llama-3.1-8B-wanda-2of4](https://huggingface.co/vuiseng9/Meta-Llama-3.1-8B-wanda-2of4) | |
|
|
|
|