Update README.md
Browse files
README.md
CHANGED
@@ -70,7 +70,7 @@ class TextGenerationAssistant:
|
|
70 |
assistant = TextGenerationAssistant(model_id="Commencis/Commencis-LLM")
|
71 |
|
72 |
# Enter your query here.
|
73 |
-
user_query = "Faiz
|
74 |
response = assistant.generate_response(user_query)
|
75 |
print(response)
|
76 |
|
@@ -84,7 +84,7 @@ import transformers
|
|
84 |
import torch
|
85 |
|
86 |
model = "Commencis/Commencis-LLM"
|
87 |
-
messages = [{"role": "user", "content": "Faiz
|
88 |
|
89 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
90 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
|
|
70 |
assistant = TextGenerationAssistant(model_id="Commencis/Commencis-LLM")
|
71 |
|
72 |
# Enter your query here.
|
73 |
+
user_query = "Faiz oranı yükseldiğinde kredi maliyetim nasıl etkilenir?"
|
74 |
response = assistant.generate_response(user_query)
|
75 |
print(response)
|
76 |
|
|
|
84 |
import torch
|
85 |
|
86 |
model = "Commencis/Commencis-LLM"
|
87 |
+
messages = [{"role": "user", "content": "Faiz oranı yükseldiğinde kredi maliyetim nasıl etkilenir?"}]
|
88 |
|
89 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
90 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|