typos
Browse files
README.md
CHANGED
@@ -9,12 +9,13 @@ tags:
|
|
9 |
- sharded
|
10 |
- 8-bit
|
11 |
- quantized
|
12 |
-
|
|
|
13 |
---
|
14 |
|
15 |
# stablelm-tuned-alpha-7b-sharded-8bit
|
16 |
|
17 |
-
This is a sharded checkpoint (with ~4GB shards) of the `stabilityai/stablelm-
|
18 |
|
19 |
Refer to the [original model](https://huggingface.co/stabilityai/stablelm-tuned-alpha-7b) for all details w.r.t. to the model. For more info on loading 8bit models, refer to the [example repo](https://huggingface.co/ybelkada/bloom-1b7-8bit) and/or the `4.28.0` [release info](https://github.com/huggingface/transformers/releases/tag/v4.28.0).
|
20 |
|
@@ -44,8 +45,8 @@ Load the model. As it is serialized in 8bit you don't need to do anything specia
|
|
44 |
```python
|
45 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
46 |
|
47 |
-
model_name = "ethzanalytics/stablelm-
|
48 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
49 |
|
50 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
51 |
-
```
|
|
|
9 |
- sharded
|
10 |
- 8-bit
|
11 |
- quantized
|
12 |
+
- tuned
|
13 |
+
inference: false
|
14 |
---
|
15 |
|
16 |
# stablelm-tuned-alpha-7b-sharded-8bit
|
17 |
|
18 |
+
This is a sharded checkpoint (with ~4GB shards) of the `stabilityai/stablelm-tuned-alpha-7b` model **in `8bit` precision** using `bitsandbytes`.
|
19 |
|
20 |
Refer to the [original model](https://huggingface.co/stabilityai/stablelm-tuned-alpha-7b) for all details w.r.t. to the model. For more info on loading 8bit models, refer to the [example repo](https://huggingface.co/ybelkada/bloom-1b7-8bit) and/or the `4.28.0` [release info](https://github.com/huggingface/transformers/releases/tag/v4.28.0).
|
21 |
|
|
|
45 |
```python
|
46 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
47 |
|
48 |
+
model_name = "ethzanalytics/stablelm-tuned-alpha-7b-sharded-8bit"
|
49 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
50 |
|
51 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
52 |
+
```
|