Update librechat.yaml
Browse files- librechat.yaml +67 -31
librechat.yaml
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
# Configuration version (required)
|
2 |
-
version: 1.0.
|
3 |
|
4 |
# Cache settings: Set to true to enable caching
|
5 |
cache: false
|
@@ -12,44 +12,20 @@ endpoints:
|
|
12 |
- name: "Mistral"
|
13 |
apiKey: "user_provided"
|
14 |
baseURL: "https://api.mistral.ai/v1"
|
15 |
-
|
16 |
-
# Models configuration
|
17 |
models:
|
18 |
default: [
|
19 |
"mistral-tiny",
|
20 |
"mistral-small",
|
21 |
"mistral-medium"
|
22 |
]
|
23 |
-
fetch:
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
# Title Conversation setting
|
28 |
-
titleConvo: true # Set to true to enable title conversation
|
29 |
-
|
30 |
-
# Title Method: Choose between "completion" or "functions".
|
31 |
-
titleMethod: "completion" # Defaults to "completion" if omitted.
|
32 |
-
|
33 |
-
# Title Model: Specify the model to use for titles.
|
34 |
-
titleModel: "mistral-tiny" # Defaults to "gpt-3.5-turbo" if omitted.
|
35 |
-
|
36 |
-
# Summarize setting: Set to true to enable summarization.
|
37 |
summarize: false
|
38 |
-
|
39 |
-
# Summary Model: Specify the model to use if summarization is enabled.
|
40 |
-
summaryModel: "mistral-tiny" # Defaults to "gpt-3.5-turbo" if omitted.
|
41 |
-
|
42 |
-
# Force Prompt setting: If true, sends a `prompt` parameter instead of `messages`.
|
43 |
forcePrompt: false
|
44 |
-
|
45 |
-
# The label displayed for the AI model in messages.
|
46 |
-
modelDisplayLabel: "Mistral" # Default is "AI" when not set.
|
47 |
-
|
48 |
-
# Add additional parameters to the request. Default params will be overwritten.
|
49 |
-
# addParams:
|
50 |
-
# safe_mode: true # This field is specific to Mistral AI: https://docs.mistral.ai/api/
|
51 |
-
|
52 |
-
# NOTE: For Mistral, it is necessary to drop the following parameters or you will encounter a 422 Error:
|
53 |
dropParams: ["stop", "user", "frequency_penalty", "presence_penalty"]
|
54 |
|
55 |
# OpenRouter.ai
|
@@ -153,6 +129,66 @@ endpoints:
|
|
153 |
forcePrompt: false
|
154 |
modelDisplayLabel: "OpenRouter"
|
155 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
156 |
# NagaAI
|
157 |
- name: "NagaAI"
|
158 |
apiKey: "user_provided"
|
|
|
1 |
# Configuration version (required)
|
2 |
+
version: 1.0.1
|
3 |
|
4 |
# Cache settings: Set to true to enable caching
|
5 |
cache: false
|
|
|
12 |
- name: "Mistral"
|
13 |
apiKey: "user_provided"
|
14 |
baseURL: "https://api.mistral.ai/v1"
|
|
|
|
|
15 |
models:
|
16 |
default: [
|
17 |
"mistral-tiny",
|
18 |
"mistral-small",
|
19 |
"mistral-medium"
|
20 |
]
|
21 |
+
fetch: false
|
22 |
+
titleConvo: true
|
23 |
+
titleMethod: "completion"
|
24 |
+
titleModel: "mistral-tiny"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
summarize: false
|
26 |
+
summaryModel: "mistral-tiny"
|
|
|
|
|
|
|
|
|
27 |
forcePrompt: false
|
28 |
+
modelDisplayLabel: "Mistral"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
dropParams: ["stop", "user", "frequency_penalty", "presence_penalty"]
|
30 |
|
31 |
# OpenRouter.ai
|
|
|
129 |
forcePrompt: false
|
130 |
modelDisplayLabel: "OpenRouter"
|
131 |
|
132 |
+
# together.ai
|
133 |
+
- name: "together.ai"
|
134 |
+
apiKey: "user_provided"
|
135 |
+
baseURL: "https://api.together.xyz"
|
136 |
+
models:
|
137 |
+
default: [
|
138 |
+
"zero-one-ai/Yi-34B-Chat",
|
139 |
+
"Austism/chronos-hermes-13b",
|
140 |
+
"DiscoResearch/DiscoLM-mixtral-8x7b-v2",
|
141 |
+
"Gryphe/MythoMax-L2-13b",
|
142 |
+
"lmsys/vicuna-13b-v1.5",
|
143 |
+
"lmsys/vicuna-7b-v1.5",
|
144 |
+
"lmsys/vicuna-13b-v1.5-16k",
|
145 |
+
"codellama/CodeLlama-13b-Instruct-hf",
|
146 |
+
"codellama/CodeLlama-34b-Instruct-hf",
|
147 |
+
"codellama/CodeLlama-70b-Instruct-hf",
|
148 |
+
"codellama/CodeLlama-7b-Instruct-hf",
|
149 |
+
"togethercomputer/llama-2-13b-chat",
|
150 |
+
"togethercomputer/llama-2-70b-chat",
|
151 |
+
"togethercomputer/llama-2-7b-chat",
|
152 |
+
"NousResearch/Nous-Capybara-7B-V1p9",
|
153 |
+
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
|
154 |
+
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT",
|
155 |
+
"NousResearch/Nous-Hermes-Llama2-70b",
|
156 |
+
"NousResearch/Nous-Hermes-llama-2-7b",
|
157 |
+
"NousResearch/Nous-Hermes-Llama2-13b",
|
158 |
+
"NousResearch/Nous-Hermes-2-Yi-34B",
|
159 |
+
"openchat/openchat-3.5-1210",
|
160 |
+
"Open-Orca/Mistral-7B-OpenOrca",
|
161 |
+
"togethercomputer/Qwen-7B-Chat",
|
162 |
+
"snorkelai/Snorkel-Mistral-PairRM-DPO",
|
163 |
+
"togethercomputer/alpaca-7b",
|
164 |
+
"togethercomputer/falcon-40b-instruct",
|
165 |
+
"togethercomputer/falcon-7b-instruct",
|
166 |
+
"togethercomputer/GPT-NeoXT-Chat-Base-20B",
|
167 |
+
"togethercomputer/Llama-2-7B-32K-Instruct",
|
168 |
+
"togethercomputer/Pythia-Chat-Base-7B-v0.16",
|
169 |
+
"togethercomputer/RedPajama-INCITE-Chat-3B-v1",
|
170 |
+
"togethercomputer/RedPajama-INCITE-7B-Chat",
|
171 |
+
"togethercomputer/StripedHyena-Nous-7B",
|
172 |
+
"Undi95/ReMM-SLERP-L2-13B",
|
173 |
+
"Undi95/Toppy-M-7B",
|
174 |
+
"WizardLM/WizardLM-13B-V1.2",
|
175 |
+
"garage-bAInd/Platypus2-70B-instruct",
|
176 |
+
"mistralai/Mistral-7B-Instruct-v0.1",
|
177 |
+
"mistralai/Mistral-7B-Instruct-v0.2",
|
178 |
+
"mistralai/Mixtral-8x7B-Instruct-v0.1",
|
179 |
+
"teknium/OpenHermes-2-Mistral-7B",
|
180 |
+
"teknium/OpenHermes-2p5-Mistral-7B",
|
181 |
+
"upstage/SOLAR-10.7B-Instruct-v1.0"
|
182 |
+
]
|
183 |
+
fetch: false
|
184 |
+
titleConvo: true
|
185 |
+
titleModel: "openchat/openchat-3.5-1210"
|
186 |
+
summarize: false
|
187 |
+
summaryModel: "openchat/openchat-3.5-1210"
|
188 |
+
forcePrompt: false
|
189 |
+
modelDisplayLabel: "together.ai"
|
190 |
+
iconURL: "https://cdn.discordapp.com/attachments/1171849907812978800/1203453238477717514/togetherai.png"
|
191 |
+
|
192 |
# NagaAI
|
193 |
- name: "NagaAI"
|
194 |
apiKey: "user_provided"
|