OPEA
/

Safetensors
llama
4-bit precision
intel/auto-round
cicdatopea commited on
Commit
5894db9
·
verified ·
1 Parent(s): 43d645d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -24,7 +24,7 @@ tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir)
24
 
25
  model = AutoModelForCausalLM.from_pretrained(
26
  quantized_model_dir,
27
- torch_dtype='auto',
28
  device_map="auto",
29
  ##revision="12cbcc0", ##AutoGPTQ format
30
  )
 
24
 
25
  model = AutoModelForCausalLM.from_pretrained(
26
  quantized_model_dir,
27
+ torch_dtype=torch.float16,
28
  device_map="auto",
29
  ##revision="12cbcc0", ##AutoGPTQ format
30
  )