Text Generation
Transformers
Safetensors
Japanese
qwen2
conversational
text-generation-inference
Inference Endpoints
mkshing commited on
Commit
2877f44
·
verified ·
1 Parent(s): 260e206

fix the usage

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -40,14 +40,14 @@ model.to(device)
40
 
41
  # 2. prepare inputs
42
  text = "知識蒸留について簡単に教えてください。"
43
- messages = [
44
- {"role": "system", "content": "あなたは誠実で優秀な日本人のアシスタントです。"},
45
- {"role": "user", "content": text},
46
- ]
47
- inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
48
 
49
  # 3. generate
50
- output_ids = model.generate(**inputs.to(device))
 
 
 
51
  output_ids = output_ids[:, inputs.input_ids.shape[1] :]
52
  generated_text = tokenizer.batch_decode(output_ids, skip_special_tokens=True)[0]
53
  print(generated_text)
 
40
 
41
  # 2. prepare inputs
42
  text = "知識蒸留について簡単に教えてください。"
43
+ messages = [{"role": "user", "content": text}]
44
+ inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt")
 
 
 
45
 
46
  # 3. generate
47
+ output_ids = model.generate(
48
+ inputs.to(device),
49
+ max_new_tokens=1024,
50
+ )
51
  output_ids = output_ids[:, inputs.input_ids.shape[1] :]
52
  generated_text = tokenizer.batch_decode(output_ids, skip_special_tokens=True)[0]
53
  print(generated_text)