Text Generation
Transformers
Safetensors
mistral
text-generation-inference
unsloth
Mistral_Star
Mistral_Quiet
Mistral
Mixtral
Question-Answer
Token-Classification
Sequence-Classification
SpydazWeb-AI
chemistry
biology
legal
code
climate
medical
LCARS_AI_StarTrek_Computer
chain-of-thought
tree-of-knowledge
forest-of-thoughts
visual-spacial-sketchpad
alpha-mind
knowledge-graph
entity-detection
encyclopedia
wikipedia
stack-exchange
Reddit
Cyber-series
MegaMind
Cybertron
SpydazWeb
Spydaz
LCARS
star-trek
mega-transformers
Mulit-Mega-Merge
Multi-Lingual
Afro-Centric
African-Model
Ancient-One
conversational
Inference Endpoints
Update README.md
Browse files
README.md
CHANGED
@@ -119,8 +119,19 @@ on top of this stack we will do finally train for conversations and dialogues .
|
|
119 |
as well as general tasks with chain of thoughts and react really sea as well as self critique . model ranking , intent detection , requirements gathering and various other agent tasks such as general q/a and instruct . and business data . .
|
120 |
the result is a genralised Intelivence.
|
121 |
|
122 |
-
now since this is a language model we should also add modalitys and other inferance heads . to also enable for the tensors to retask the embedding space with enhanced richness !
|
123 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
124 |
|
125 |
— # Leroy Dyer (1972-Present)
|
126 |
<img src="https://cdn-avatars.huggingface.co/v1/production/uploads/65d883893a52cd9bcd8ab7cf/tRsCJlHNZo1D02kBTmfy9.jpeg" width="300"/>
|
|
|
119 |
as well as general tasks with chain of thoughts and react really sea as well as self critique . model ranking , intent detection , requirements gathering and various other agent tasks such as general q/a and instruct . and business data . .
|
120 |
the result is a genralised Intelivence.
|
121 |
|
122 |
+
now since this is a language model we should also add modalitys and other inferance heads . to also enable for the tensors to retask the embedding space with enhanced richness !
|
123 |
+
|
124 |
+
# the value of a PROMPT
|
125 |
+
the true value of a prompt is the fact it tasks the model or triggers the trained tasks . so when releasing a model prompts should also be released or tasks trained may not be known . in fact tasks are similar to commands so given the prompt the model will remember the task it was trained with this prompt . as with training a model can become over fit ! so we just change the prompt . when training the prompt is associated with the expected output so the embeddings relate to this prompt . a large prompt can be considered a instruction but if it is not trained with multiple examples the expected output is unknown. so a large prompt can give the user more accidental relation to the tasks. smaller prompts are useful as they remove the expectation of the task and he genralise the response
|
126 |
+
so after fully training a task remove your prompt or simplify the prompt so that the task will still execute despite the prompt not matching the original task prompt .
|
127 |
+
for specific processes large sets of examples must be trained give as much examples as possible . this can drastically change the expected output shape . so when training you should use data which you have already trained ! just reframed I to the reponse style you desire as well as the execution. strategy desired .
|
128 |
+
playing around with prompts in training is so useful as the model does learn to think . as a single question now could have many possible responses and many possible solutions. henxe now temperature and topK will actually have an effect on your model !
|
129 |
+
|
130 |
+
# grounding a model
|
131 |
+
grounding a model by simply training a general alpaca with no prompt . this can make your trained model become a great base model as all past training is contained within sub layers the knowledge is still contained and the model has no objective , IE the last fine-tuning !
|
132 |
+
this enables for your universal user to guide the model to Thier desires !
|
133 |
+
for this i generally use a dialogue data set such as (Samantha ) or a duacussive dataset such as medical council counciling . enabling for the top layer to be a chat level !
|
134 |
+
this will enable for the model to suprise the user ! sometimes reasoning deed and sometimes anßering directly ..
|
135 |
|
136 |
— # Leroy Dyer (1972-Present)
|
137 |
<img src="https://cdn-avatars.huggingface.co/v1/production/uploads/65d883893a52cd9bcd8ab7cf/tRsCJlHNZo1D02kBTmfy9.jpeg" width="300"/>
|