Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,6 @@ license: apache-2.0
|
|
5 |
[Optimum Habana](https://github.com/huggingface/optimum-habana) is the interface between the Transformers library and Habana's Gaudi processor (HPU). It provides a set of tools enabling easy and fast model loading and fine-tuning on single- and multi-HPU settings for different downstream tasks.
|
6 |
Learn more about how to take advantage of the power of Habana HPUs to train Transformers models at [hf.co/Habana](https://huggingface.co/Habana).
|
7 |
|
8 |
-
|
9 |
## RoBERTa Base model HPU configuration
|
10 |
|
11 |
This model contains just the `GaudiConfig` file for running the [roberta-base](https://huggingface.co/roberta-base) model on Habana's Gaudi processors (HPU).
|
@@ -21,7 +20,6 @@ This enables to specify:
|
|
21 |
- `use_fused_adam`: whether to use Habana's custom AdamW implementation
|
22 |
- `use_fused_clip_norm`: whether to use Habana's fused gradient norm clipping operator
|
23 |
|
24 |
-
|
25 |
## Usage
|
26 |
|
27 |
The model is instantiated the same way as in the Transformers library.
|
@@ -31,7 +29,6 @@ The only difference is that the Gaudi configuration has to be loaded and provide
|
|
31 |
from optimum.habana import GaudiConfig, GaudiTrainer, GaudiTrainingArguments
|
32 |
from transformers import RobertaModel, RobertaTokenizer
|
33 |
|
34 |
-
|
35 |
tokenizer = RobertaTokenizer.from_pretrained("roberta-base")
|
36 |
model = RobertaModel.from_pretrained("roberta-base")
|
37 |
gaudi_config = GaudiConfig.from_pretrained("Habana/roberta-base")
|
|
|
5 |
[Optimum Habana](https://github.com/huggingface/optimum-habana) is the interface between the Transformers library and Habana's Gaudi processor (HPU). It provides a set of tools enabling easy and fast model loading and fine-tuning on single- and multi-HPU settings for different downstream tasks.
|
6 |
Learn more about how to take advantage of the power of Habana HPUs to train Transformers models at [hf.co/Habana](https://huggingface.co/Habana).
|
7 |
|
|
|
8 |
## RoBERTa Base model HPU configuration
|
9 |
|
10 |
This model contains just the `GaudiConfig` file for running the [roberta-base](https://huggingface.co/roberta-base) model on Habana's Gaudi processors (HPU).
|
|
|
20 |
- `use_fused_adam`: whether to use Habana's custom AdamW implementation
|
21 |
- `use_fused_clip_norm`: whether to use Habana's fused gradient norm clipping operator
|
22 |
|
|
|
23 |
## Usage
|
24 |
|
25 |
The model is instantiated the same way as in the Transformers library.
|
|
|
29 |
from optimum.habana import GaudiConfig, GaudiTrainer, GaudiTrainingArguments
|
30 |
from transformers import RobertaModel, RobertaTokenizer
|
31 |
|
|
|
32 |
tokenizer = RobertaTokenizer.from_pretrained("roberta-base")
|
33 |
model = RobertaModel.from_pretrained("roberta-base")
|
34 |
gaudi_config = GaudiConfig.from_pretrained("Habana/roberta-base")
|