justinj92 commited on
Commit
88d842f
·
verified ·
1 Parent(s): ad3b2c3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -1,15 +1,17 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
3
  library_name: transformers
4
- model_name: Qwen-1.5B-GRPO
5
  tags:
6
  - generated_from_trainer
7
  - trl
8
  - grpo
9
  licence: license
 
 
10
  ---
11
 
12
- # Model Card for Qwen-1.5B-GRPO
13
 
14
  This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
@@ -19,8 +21,8 @@ It has been trained using [TRL](https://github.com/huggingface/trl).
19
  ```python
20
  from transformers import pipeline
21
 
22
- question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
- generator = pipeline("text-generation", model="justinj92/Qwen-1.5B-GRPO", device="cuda")
24
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
  print(output["generated_text"])
26
  ```
 
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
3
  library_name: transformers
4
+ model_name: Qwen2.5-1.5B-Thinking
5
  tags:
6
  - generated_from_trainer
7
  - trl
8
  - grpo
9
  licence: license
10
+ datasets:
11
+ - microsoft/orca-math-word-problems-200k
12
  ---
13
 
14
+ # Model Card for Qwen2.5-1.5B-Thinking
15
 
16
  This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
 
21
  ```python
22
  from transformers import pipeline
23
 
24
+ question = "Mia can decorate 2 dozen Easter eggs per hour. Her little brother Billy can only decorate 10 eggs per hour. They need to decorate 170 eggs for the Easter egg hunt. If they work together, how long will it take them to decorate all the eggs?"
25
+ generator = pipeline("text-generation", model="justinj92/Qwen2.5-1.5B-Thinking", device="cuda")
26
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
27
  print(output["generated_text"])
28
  ```