Text Generation
Transformers
Safetensors
mistral
mergekit
Merge
Mistral_Star
Mistral_Quiet
Mistral
Mixtral
Question-Answer
Token-Classification
Sequence-Classification
SpydazWeb-AI
chemistry
biology
legal
code
climate
medical
LCARS_AI_StarTrek_Computer
text-generation-inference
chain-of-thought
tree-of-knowledge
forest-of-thoughts
visual-spacial-sketchpad
alpha-mind
knowledge-graph
entity-detection
encyclopedia
wikipedia
stack-exchange
Reddit
Cyber-series
MegaMind
Cybertron
SpydazWeb
Spydaz
LCARS
star-trek
mega-transformers
Mulit-Mega-Merge
Multi-Lingual
Afro-Centric
African-Model
Ancient-One
conversational
Inference Endpoints
Update README.md
Browse files
README.md
CHANGED
@@ -71,6 +71,9 @@ datasets:
|
|
71 |
- xz56/react-llama
|
72 |
- BeIR/hotpotqa
|
73 |
- YBXL/medical_book_train_filtered
|
|
|
|
|
|
|
74 |
tags:
|
75 |
- mergekit
|
76 |
- merge
|
@@ -250,8 +253,4 @@ Fine-tuning with a focus on attention can help prevent overfitting to specific p
|
|
250 |
# “Epochs are the key to effective training, rather than merely mass dumping examples—unless those examples are interconnected within a single or multiple conversations that teach through dialogue.”
|
251 |
|
252 |
My personal training methods are unconventional. I prioritize creating conversations that allow the model to learn new topics from diverse perspectives. This approach is essential, as many models are losing their unique personalities. Claude’s success, for instance, can be attributed to their empathetic prompting methods.
|
253 |
-
It’s important for the model to express itself, even during training, which can be challenging. Role-playing and conversational training are effective strategies to help the model learn to communicate naturally. Currently, the training has become overly focused on technical methodologies and task expectations, resulting in a loss of personality.
|
254 |
-
|
255 |
-
|
256 |
-
|
257 |
-
|
|
|
71 |
- xz56/react-llama
|
72 |
- BeIR/hotpotqa
|
73 |
- YBXL/medical_book_train_filtered
|
74 |
+
- SkunkworksAI/reasoning-0.01
|
75 |
+
- rewoo/planner_instruction_tuning_2k
|
76 |
+
- THUDM/LongWriter-6k
|
77 |
tags:
|
78 |
- mergekit
|
79 |
- merge
|
|
|
253 |
# “Epochs are the key to effective training, rather than merely mass dumping examples—unless those examples are interconnected within a single or multiple conversations that teach through dialogue.”
|
254 |
|
255 |
My personal training methods are unconventional. I prioritize creating conversations that allow the model to learn new topics from diverse perspectives. This approach is essential, as many models are losing their unique personalities. Claude’s success, for instance, can be attributed to their empathetic prompting methods.
|
256 |
+
It’s important for the model to express itself, even during training, which can be challenging. Role-playing and conversational training are effective strategies to help the model learn to communicate naturally. Currently, the training has become overly focused on technical methodologies and task expectations, resulting in a loss of personality.
|
|
|
|
|
|
|
|