Update README.md
Browse files
README.md
CHANGED
@@ -45,7 +45,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
|
45 |
class TextGenerationAssistant:
|
46 |
def __init__(self, model_id:str):
|
47 |
self.tokenizer = AutoTokenizer.from_pretrained(model_id)
|
48 |
-
self.model = AutoModelForCausalLM.from_pretrained(model_id, device_map='auto'
|
49 |
self.pipe = pipeline("text-generation",
|
50 |
model=self.model,
|
51 |
tokenizer=self.tokenizer,
|
@@ -56,10 +56,10 @@ class TextGenerationAssistant:
|
|
56 |
)
|
57 |
|
58 |
self.sampling_params = dict(do_sample=True, temperature=0.5, top_k=50, top_p=0.9)
|
59 |
-
self.
|
60 |
|
61 |
def format_prompt(self, user_input):
|
62 |
-
return "[INST] " + self.
|
63 |
|
64 |
def generate_response(self, user_query):
|
65 |
prompt = self.format_prompt(user_query)
|
|
|
45 |
class TextGenerationAssistant:
|
46 |
def __init__(self, model_id:str):
|
47 |
self.tokenizer = AutoTokenizer.from_pretrained(model_id)
|
48 |
+
self.model = AutoModelForCausalLM.from_pretrained(model_id, device_map='auto')
|
49 |
self.pipe = pipeline("text-generation",
|
50 |
model=self.model,
|
51 |
tokenizer=self.tokenizer,
|
|
|
56 |
)
|
57 |
|
58 |
self.sampling_params = dict(do_sample=True, temperature=0.5, top_k=50, top_p=0.9)
|
59 |
+
self.system_prompt = "Sen yardımcı bir asistansın. Sana verilen talimat ve girdilere en uygun cevapları üreteceksin. \n\n\n"
|
60 |
|
61 |
def format_prompt(self, user_input):
|
62 |
+
return "[INST] " + self.system_prompt + user_input + " [/INST]"
|
63 |
|
64 |
def generate_response(self, user_query):
|
65 |
prompt = self.format_prompt(user_query)
|