regisss HF staff commited on
Commit
8cf5c76
·
1 Parent(s): 49793c7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -3
README.md CHANGED
@@ -5,7 +5,6 @@ license: apache-2.0
5
  [Optimum Habana](https://github.com/huggingface/optimum-habana) is the interface between the Transformers library and Habana's Gaudi processor (HPU). It provides a set of tools enabling easy and fast model loading and fine-tuning on single- and multi-HPU settings for different downstream tasks.
6
  Learn more about how to take advantage of the power of Habana HPUs to train Transformers models at [hf.co/Habana](https://huggingface.co/Habana).
7
 
8
-
9
  ## RoBERTa Base model HPU configuration
10
 
11
  This model contains just the `GaudiConfig` file for running the [roberta-base](https://huggingface.co/roberta-base) model on Habana's Gaudi processors (HPU).
@@ -21,7 +20,6 @@ This enables to specify:
21
  - `use_fused_adam`: whether to use Habana's custom AdamW implementation
22
  - `use_fused_clip_norm`: whether to use Habana's fused gradient norm clipping operator
23
 
24
-
25
  ## Usage
26
 
27
  The model is instantiated the same way as in the Transformers library.
@@ -31,7 +29,6 @@ The only difference is that the Gaudi configuration has to be loaded and provide
31
  from optimum.habana import GaudiConfig, GaudiTrainer, GaudiTrainingArguments
32
  from transformers import RobertaModel, RobertaTokenizer
33
 
34
-
35
  tokenizer = RobertaTokenizer.from_pretrained("roberta-base")
36
  model = RobertaModel.from_pretrained("roberta-base")
37
  gaudi_config = GaudiConfig.from_pretrained("Habana/roberta-base")
 
5
  [Optimum Habana](https://github.com/huggingface/optimum-habana) is the interface between the Transformers library and Habana's Gaudi processor (HPU). It provides a set of tools enabling easy and fast model loading and fine-tuning on single- and multi-HPU settings for different downstream tasks.
6
  Learn more about how to take advantage of the power of Habana HPUs to train Transformers models at [hf.co/Habana](https://huggingface.co/Habana).
7
 
 
8
  ## RoBERTa Base model HPU configuration
9
 
10
  This model contains just the `GaudiConfig` file for running the [roberta-base](https://huggingface.co/roberta-base) model on Habana's Gaudi processors (HPU).
 
20
  - `use_fused_adam`: whether to use Habana's custom AdamW implementation
21
  - `use_fused_clip_norm`: whether to use Habana's fused gradient norm clipping operator
22
 
 
23
  ## Usage
24
 
25
  The model is instantiated the same way as in the Transformers library.
 
29
  from optimum.habana import GaudiConfig, GaudiTrainer, GaudiTrainingArguments
30
  from transformers import RobertaModel, RobertaTokenizer
31
 
 
32
  tokenizer = RobertaTokenizer.from_pretrained("roberta-base")
33
  model = RobertaModel.from_pretrained("roberta-base")
34
  gaudi_config = GaudiConfig.from_pretrained("Habana/roberta-base")