# Configuration version (required) version: 1.0.1 cache: true registration: socialLogins: ["discord", "facebook", "github", "google", "openid"] endpoints: custom: # Anyscale # # Model list: https://console.anyscale.com/v2/playground - name: "Anyscale" apiKey: "${ANYSCALE_API_KEY}" baseURL: "https://api.endpoints.anyscale.com/v1" models: default: [ "meta-llama/Llama-2-7b-chat-hf", "meta-llama/Llama-2-13b-chat-hf", "meta-llama/Llama-2-70b-chat-hf", "codellama/CodeLlama-34b-Instruct-hf", "codellama/CodeLlama-70b-Instruct-hf", "mistralai/Mistral-7B-Instruct-v0.1", "mistralai/Mixtral-8x7B-Instruct-v0.1", "mlabonne/NeuralHermes-2.5-Mistral-7B", "Open-Orca/Mistral-7B-OpenOrca", "HuggingFaceH4/zephyr-7b-beta", "google/gemma-7b-it" ] fetch: false titleConvo: true titleModel: "meta-llama/Llama-2-7b-chat-hf" summarize: false summaryModel: "meta-llama/Llama-2-7b-chat-hf" forcePrompt: false modelDisplayLabel: "Anyscale" # APIpie # https://apipie.ai/dashboard/ # Script to fetch models: https://github.com/LibreChat-AI/librechat-config-yaml/blob/main/scripts/apipie.py - name: "APIpie" apiKey: "${APIPIE_API_KEY}" baseURL: "https://apipie.ai/v1/" models: default: [ "GPT-JT-Moderation-6B", "LLaMA-2-7B-32K", "Llama-2-13b-chat-hf", "Llama-2-13b-hf", "Llama-2-70b-chat-hf", "Llama-2-70b-hf", "Llama-2-7B-32K-Instruct", "Llama-2-7b-chat-hf", "Llama-2-7b-hf", "Meta-Llama-3-70B-Instruct", "Meta-Llama-3-8B", "Meta-Llama-3-8B-Instruct", "Mistral-7B-Instruct-v0.1", "Mistral-7B-Instruct-v0.2", "Mistral-7B-OpenOrca", "Mixtral-8x22B-Instruct-v0.1", "Mixtral-8x22B-v0.1", "Mixtral-8x7B-Instruct-v0.1", "Mixtral-8x7B-v0.1", "MythoMax-L2-13b", "NexusRaven-V2-13B", "Nous-Hermes-2-Mixtral-8x7B-DPO", "Nous-Hermes-2-Mixtral-8x7B-SFT", "Nous-Hermes-Llama2-13b", "Nous-Hermes-llama-2-7b", "ReMM-SLERP-L2-13B", "RedPajama-INCITE-7B-Base", "RedPajama-INCITE-7B-Chat", "RedPajama-INCITE-Chat-3B-v1", "TinyLlama-1.1B-Chat-v1.0", "Toppy-M-7B", "WizardLM-2-7B", "WizardLM-2-8x22B", "Yi-34B-Chat", "airoboros-70b", "airoboros-l2-70b", "alpaca-7b", "babbage-002", "chat-bison", "chatgpt-4o-latest", "chronos-hermes-13b", "chronos-hermes-13b-v2", "claude-1", "claude-1.2", "claude-2", "claude-2.0", "claude-2.1", "claude-3-5-sonnet", "claude-3-5-sonnet-20240620-v1", "claude-3-haiku", "claude-3-haiku-20240307-v1", "claude-3-opus", "claude-3-sonnet", "claude-3-sonnet-20240229-v1", "claude-3.5-sonnet", "claude-instant-1", "claude-instant-1.0", "claude-instant-1.1", "claude-instant-1.2", "claude-instant-v1", "claude-v2", "codellama-34b-instruct", "codellama-70b-instruct", "codestral-mamba", "command", "command-light", "command-light-nightly", "command-light-text-v14", "command-nightly", "command-r", "command-r-plus", "command-r-plus-v1", "command-r-v1", "command-text-v14", "davinci-002", "dbrx-instruct", "deepseek-chat", "deepseek-coder", "dolphin-2.5-mixtral-8x7b", "dolphin-2.6-mixtral-8x7b", "dolphin-llama-3-70b", "dolphin-mixtral-8x22b", "dolphin-mixtral-8x7b", "eagle-7b", "fimbulvetr-11b-v2", "firellava-13b", "gemini-1.5-flash", "gemini-1.5-pro", "gemini-flash-1.5", "gemini-pro", "gemini-pro-1.5", "gemini-pro-1.5-exp", "gemini-pro-vision", "gemma-1.1-7b-it", "gemma-2-27b-it", "gemma-2-9b-it", "gemma-7b-it", "goliath-120b", "gpt-3.5-turbo", "gpt-3.5-turbo-0125", "gpt-3.5-turbo-0301", "gpt-3.5-turbo-0613", "gpt-3.5-turbo-1106", "gpt-3.5-turbo-16k", "gpt-3.5-turbo-16k-0613", "gpt-3.5-turbo-instruct", "gpt-3.5-turbo-instruct-0914", "gpt-4", "gpt-4-0125-preview", "gpt-4-0314", "gpt-4-0613", "gpt-4-1106-preview", "gpt-4-1106-vision-preview", "gpt-4-32k", "gpt-4-32k-0314", "gpt-4-turbo", "gpt-4-turbo-2024-04-09", "gpt-4-turbo-preview", "gpt-4-vision-preview", "gpt-4o", "gpt-4o-2024-05-13", "gpt-4o-2024-08-06", "gpt-4o-mini", "gpt-4o-mini-2024-07-18", "hermes-2-pro-llama-3-8b", "hermes-2-theta-llama-3-8b", "hermes-3-llama-3.1-405b", "hermes-3-llama-3.1-70b", "j2-grande-instruct", "j2-jumbo-instruct", "j2-mid", "j2-mid-v1", "j2-ultra", "j2-ultra-v1", "jamba-1-5-large", "jamba-1-5-mini", "jamba-instruct", "jamba-instruct-v1", "l3-euryale-70b", "l3-lunaris-8b", "l3-stheno-8b", "large-latest", "llama-2-13b-chat", "llama-2-70b-chat", "llama-3-70b", "llama-3-70b-instruct", "llama-3-8b", "llama-3-8b-instruct", "llama-3-lumimaid-70b", "llama-3-lumimaid-8b", "llama-3-sonar-large-32k-chat", "llama-3-sonar-large-32k-online", "llama-3-sonar-small-32k-chat", "llama-3-sonar-small-32k-online", "llama-3.1-405b", "llama-3.1-405b-instruct", "llama-3.1-70b-instruct", "llama-3.1-8b-instruct", "llama-3.1-sonar-huge-128k-online", "llama-3.1-sonar-large-128k-chat", "llama-3.1-sonar-large-128k-online", "llama-3.1-sonar-small-128k-chat", "llama-3.1-sonar-small-128k-online", "llama-guard-2-8b", "llama2-13b-chat-v1", "llama2-70b-chat-v1", "llama3-70b-instruct-v1", "llama3-70b-instruct-v1:0", "llama3-8b-instruct-v1", "llama3-8b-instruct-v1:0", "llava-1.5-7b-hf", "lzlv-70b-fp16-hf", "lzlv_70b_fp16_hf", "magnum-72b", "medium", "midnight-rose-70b", "mistral-7b-instruct", "mistral-7b-instruct-v0", "mistral-7b-instruct-v0.1", "mistral-7b-instruct-v0.2", "mistral-7b-instruct-v0.3", "mistral-7b-openorca", "mistral-large", "mistral-large-2402-v1", "mistral-medium", "mistral-nemo", "mistral-small", "mistral-small-2402-v1", "mistral-tiny", "mixtral-8x22b", "mixtral-8x22b-instruct", "mixtral-8x7b", "mixtral-8x7b-instruct", "mixtral-8x7b-instruct-v0", "mn-celeste-12b", "mn-starcannon-12b", "mythalion-13b", "mythomax-l2-13b", "mythomist-7b", "noromaid-20b", "nous-capybara-7b", "nous-hermes-2-mistral-7b-dpo", "nous-hermes-2-mixtral-8x7b-dpo", "nous-hermes-2-mixtral-8x7b-sft", "nous-hermes-2-vision-7b", "nous-hermes-llama2-13b", "nous-hermes-yi-34b", "olmo-7b-instruct", "olympus-premier-v1", "openchat-3.5-1210", "openchat-7b", "openchat-8b", "openchat_3.5", "openhermes-2-mistral-7b", "openhermes-2.5-mistral-7b", "palm-2-chat-bison", "palm-2-chat-bison-32k", "palm-2-codechat-bison", "palm-2-codechat-bison-32k", "phi-2", "phi-3-medium-128k-instruct", "phi-3-medium-4k-instruct", "phi-3-mini-128k-instruct", "phi-3.5-mini-128k-instruct", "phind-codellama-34b", "pplx-70b-online", "pplx-7b-chat", "qwen-110b-chat", "qwen-14b-chat", "qwen-2-72b-instruct", "qwen-2-7b-instruct", "qwen-32b-chat", "qwen-4b-chat", "qwen-72b-chat", "qwen-7b-chat", "remm-slerp-l2-13b", "small", "snowflake-arctic-instruct", "soliloquy-l3", "sonar-medium-online", "sonar-small-chat", "sonar-small-online", "stripedhyena-hessian-7b", "stripedhyena-nous-7b", "text-babbage-002", "text-bison", "text-davinci-002", "tiny", "titan-text-express-v1", "titan-text-lite-v1", "titan-text-premier-v1", "titan-tg1-large", "toppy-m-7b", "vicuna-13b-v1.5", "vicuna-7b-v1.5", "weaver", "wizardlm-2-7b", "wizardlm-2-8x22b", "xwin-lm-70b", "yi-1.5-34b-chat", "yi-34b", "yi-34b-chat", "yi-6b", "yi-large", "yi-large-fc", "yi-large-turbo", "yi-vision", "zephyr-7b-beta", "zephyr-orpo-141b-A35b-v0.1" ] fetch: false titleConvo: true titleModel: "claude-3-haiku" summarize: false summaryModel: "claude-3-haiku" dropParams: ["stream"] modelDisplayLabel: "APIpie" iconURL: "https://raw.githubusercontent.com/fuegovic/lc-config-yaml/main/icons/APIpie.png" # cohere # Model list: https://dashboard.cohere.com/playground/chat - name: "cohere" apiKey: "${COHERE_API_KEY}" baseURL: "https://api.cohere.ai/v1" models: default: [ "c4ai-aya-23-35b", "c4ai-aya-23-8b", "command", "command-light", "command-light-nightly", "command-nightly", "command-r", "command-r-plus", ] fetch: false modelDisplayLabel: "cohere" titleModel: "command" dropParams: ["stop", "user", "frequency_penalty", "presence_penalty", "temperature", "top_p"] # DEEPNIGHT # https://github.com/brahmai-research/aiforcause # Model list: https://aiforcause.deepnight.tech/models - name: "DEEPNIGHT" apiKey: "sk-free1234" baseURL: "https://aiforcause.deepnight.tech/openai/" models: default: [ "gpt-35-turbo", "gpt-35-turbo-16k", "gpt-4-turbo" ] fetch: false titleConvo: true titleModel: "gpt-35-turbo" summarize: false summaryModel: "gpt-35-turbo" forcePrompt: false modelDisplayLabel: "DEEPNIGHT" addParams: stream: True iconURL: "https://raw.githubusercontent.com/fuegovic/lc-config-yaml/main/icons/DEEPNIGHT.png" # deepseek # https://platform.deepseek.com/api_keys # Model list: https://platform.deepseek.com/api-docs/pricing - name: "deepseek" apiKey: "${DEEPSEEK_API_KEY}" baseURL: "https://api.deepseek.com" models: default: [ "deepseek-chat", "deepseek-coder" ] fetch: false titleConvo: true titleModel: "deepseek-chat" summarize: false summaryModel: "deepseek-chat" forcePrompt: false modelDisplayLabel: "DeepSeek" # Fireworks.ai # Models: https://fireworks.ai/models?show=Serverless - name: "Fireworks" apiKey: "${FIREWORKS_API_KEY}" baseURL: "https://api.fireworks.ai/inference/v1" models: default: [ "accounts/fireworks/models/devashisht-test-v2", "accounts/fireworks/models/dt-fc-rc-v1", "accounts/fireworks/models/firefunction-v1", "accounts/fireworks/models/firefunction-v2", "accounts/fireworks/models/firellava-13b", "accounts/devashisht-72fdad/models/function-calling-v11", "accounts/fireworks/models/fw-function-call-34b-v0", "accounts/stability/models/japanese-stablelm-instruct-beta-70b", "accounts/stability/models/japanese-stablelm-instruct-gamma-7b", "accounts/fireworks/models/japanese-stable-vlm", "accounts/fireworks/models/gemma2-9b-it", "accounts/fireworks/models/llama-v3p1-405b-instruct", "accounts/fireworks/models/llama-v3p1-70b-instruct", "accounts/fireworks/models/llama-v3p1-8b-instruct", "accounts/fireworks/models/llama-v3-70b-instruct", "accounts/fireworks/models/llama-v3-70b-instruct-hf", "accounts/fireworks/models/llama-v3-8b-hf", "accounts/fireworks/models/llama-v3-8b-instruct", "accounts/fireworks/models/llama-v3-8b-instruct-hf", "accounts/fireworks/models/llama-v2-13b-chat", "accounts/fireworks/models/llama-v2-13b-code-instruct", "accounts/fireworks/models/llama-v2-34b-code-instruct", "accounts/fireworks/models/llama-v2-70b-chat", "accounts/fireworks/models/llama-v2-70b-code-instruct", "accounts/fireworks/models/llama-v2-7b-chat", "accounts/fireworks/models/deepseek-coder-v2-instruct", "accounts/fireworks/models/deepseek-coder-v2-lite-instruct", "accounts/fireworks/models/llava-v15-13b-fireworks", "accounts/fireworks/models/mistral-7b-instruct-4k", "accounts/dev-e24710/models/mistral-spellbound-format", "accounts/fireworks/models/mixtral-8x22b-instruct", "accounts/fireworks/models/mixtral-8x7b-instruct", "accounts/fireworks/models/mixtral-8x7b-instruct-hf", "accounts/fireworks/models/new-mixtral-chat", "accounts/fireworks/models/qwen-14b-chat", "accounts/fireworks/models/qwen-1-8b-chat", "accounts/fireworks/models/qwen-72b-chat", "accounts/stability/models/stablelm-zephyr-3b", "accounts/fireworks/models/yi-34b-200k-capybara", ] fetch: false titleConvo: true titleModel: "accounts/fireworks/models/llama-v2-7b-chat" summarize: false summaryModel: "accounts/fireworks/models/llama-v2-7b-chat" forcePrompt: false modelDisplayLabel: "Fireworks" dropParams: ["user"] # groq # Model list: https://console.groq.com/settings/limits - name: "groq" apiKey: "${GROQ_API_KEY}" baseURL: "https://api.groq.com/openai/v1/" models: default: [ "llama-3.1-405b-reasoning", "llama-3.1-70b-versatile", "llama-3.1-8b-instant", "llama3-groq-70b-8192-tool-use-preview", "llama3-groq-8b-8192-tool-use-preview", "llama3-70b-8192", "llama3-8b-8192", "mixtral-8x7b-32768", "gemma-7b-it", "gemma2-9b-it" ] fetch: false titleConvo: true titleModel: "mixtral-8x7b-32768" modelDisplayLabel: "groq" # HuggingFace # https://huggingface.co/settings/tokens - name: 'HuggingFace' apiKey: '${HUGGINGFACE_TOKEN}' baseURL: 'https://api-inference.huggingface.co/v1' models: default: [ "codellama/CodeLlama-34b-Instruct-hf", "google/gemma-1.1-2b-it", "google/gemma-1.1-7b-it", "HuggingFaceH4/starchat2-15b-v0.1", "HuggingFaceH4/zephyr-7b-beta", "meta-llama/Meta-Llama-3-8B-Instruct", "microsoft/Phi-3-mini-4k-instruct", "mistralai/Mistral-7B-Instruct-v0.1", "mistralai/Mistral-7B-Instruct-v0.2", "mistralai/Mixtral-8x7B-Instruct-v0.1", "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", ] fetch: true titleConvo: true titleModel: "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO" dropParams: ["top_p"] # Mistral AI API # Model list: https://docs.mistral.ai/getting-started/models/ - name: "Mistral" apiKey: "${MISTRAL_API_KEY}" baseURL: "https://api.mistral.ai/v1" models: default: [ "mistral-tiny", "mistral-small", "mistral-medium", "mistral-large-latest" ] fetch: false titleConvo: true titleMethod: "completion" titleModel: "mistral-tiny" summarize: false summaryModel: "mistral-tiny" forcePrompt: false modelDisplayLabel: "Mistral" dropParams: ["stop", "user", "frequency_penalty", "presence_penalty"] # OpenRouter.ai # Model list: https://openrouter.ai/models # Script to fetch models: https://github.com/LibreChat-AI/librechat-config-yaml/blob/main/scripts/openrouter.py - name: "OpenRouter" apiKey: "${OPENROUTER_KEY}" baseURL: "https://openrouter.ai/api/v1" models: default: [ "openrouter/auto", "---FREE---", "google/gemma-2-9b-it:free", "google/gemma-7b-it:free", "gryphe/mythomist-7b:free", "huggingfaceh4/zephyr-7b-beta:free", "meta-llama/llama-3-8b-instruct:free", "meta-llama/llama-3.1-8b-instruct:free", "microsoft/phi-3-medium-128k-instruct:free", "microsoft/phi-3-mini-128k-instruct:free", "mistralai/mistral-7b-instruct:free", "nousresearch/nous-capybara-7b:free", "openchat/openchat-7b:free", "qwen/qwen-2-7b-instruct:free", "undi95/toppy-m-7b:free", "---NITRO---", "google/gemma-7b-it:nitro", "gryphe/mythomax-l2-13b:nitro", "meta-llama/llama-3-70b-instruct:nitro", "meta-llama/llama-3-8b-instruct:nitro", "mistralai/mistral-7b-instruct:nitro", "mistralai/mixtral-8x7b-instruct:nitro", "undi95/toppy-m-7b:nitro", "---BETA---", "anthropic/claude-2.0:beta", "anthropic/claude-2.1:beta", "anthropic/claude-2:beta", "anthropic/claude-3-haiku:beta", "anthropic/claude-3-opus:beta", "anthropic/claude-3-sonnet:beta", "anthropic/claude-3.5-sonnet:beta", "anthropic/claude-instant-1:beta", "---EXTENDED---", "gryphe/mythomax-l2-13b:extended", "meta-llama/llama-3-8b-instruct:extended", "neversleep/llama-3-lumimaid-8b:extended", "nousresearch/hermes-3-llama-3.1-405b:extended", "openai/gpt-4o:extended", "undi95/remm-slerp-l2-13b:extended", "---01-AI---", "01-ai/yi-1.5-34b-chat", "01-ai/yi-34b", "01-ai/yi-34b-chat", "01-ai/yi-6b", "01-ai/yi-large", "01-ai/yi-large-fc", "01-ai/yi-large-turbo", "01-ai/yi-vision", "---AI21---", "ai21/jamba-1-5-large", "ai21/jamba-1-5-mini", "ai21/jamba-instruct", "---ANTHROPIC---", "anthropic/claude-1", "anthropic/claude-1.2", "anthropic/claude-2", "anthropic/claude-2.0", "anthropic/claude-2.1", "anthropic/claude-3-haiku", "anthropic/claude-3-opus", "anthropic/claude-3-sonnet", "anthropic/claude-3.5-sonnet", "anthropic/claude-instant-1", "anthropic/claude-instant-1.0", "anthropic/claude-instant-1.1", "---COGNITIVECOMPUTATIONS---", "cognitivecomputations/dolphin-llama-3-70b", "cognitivecomputations/dolphin-mixtral-8x22b", "cognitivecomputations/dolphin-mixtral-8x7b", "---COHERE---", "cohere/command", "cohere/command-r", "cohere/command-r-plus", "---GOOGLE---", "google/gemini-flash-1.5", "google/gemini-pro", "google/gemini-pro-1.5", "google/gemini-pro-1.5-exp", "google/gemini-pro-vision", "google/gemma-2-27b-it", "google/gemma-2-9b-it", "google/gemma-7b-it", "google/palm-2-chat-bison", "google/palm-2-chat-bison-32k", "google/palm-2-codechat-bison", "google/palm-2-codechat-bison-32k", "---META-LLAMA---", "meta-llama/codellama-34b-instruct", "meta-llama/codellama-70b-instruct", "meta-llama/llama-2-13b-chat", "meta-llama/llama-2-70b-chat", "meta-llama/llama-3-70b", "meta-llama/llama-3-70b-instruct", "meta-llama/llama-3-8b", "meta-llama/llama-3-8b-instruct", "meta-llama/llama-3.1-405b", "meta-llama/llama-3.1-405b-instruct", "meta-llama/llama-3.1-70b-instruct", "meta-llama/llama-3.1-8b-instruct", "meta-llama/llama-guard-2-8b", "---MICROSOFT---", "microsoft/phi-3-medium-128k-instruct", "microsoft/phi-3-medium-4k-instruct", "microsoft/phi-3-mini-128k-instruct", "microsoft/phi-3.5-mini-128k-instruct", "microsoft/wizardlm-2-7b", "microsoft/wizardlm-2-8x22b", "---MISTRALAI---", "mistralai/codestral-mamba", "mistralai/mistral-7b-instruct", "mistralai/mistral-7b-instruct-v0.1", "mistralai/mistral-7b-instruct-v0.2", "mistralai/mistral-7b-instruct-v0.3", "mistralai/mistral-large", "mistralai/mistral-medium", "mistralai/mistral-nemo", "mistralai/mistral-small", "mistralai/mistral-tiny", "mistralai/mixtral-8x22b", "mistralai/mixtral-8x22b-instruct", "mistralai/mixtral-8x7b", "mistralai/mixtral-8x7b-instruct", "---NEVERSLEEP---", "neversleep/llama-3-lumimaid-70b", "neversleep/llama-3-lumimaid-8b", "neversleep/noromaid-20b", "---NOUSRESEARCH---", "nousresearch/hermes-2-pro-llama-3-8b", "nousresearch/hermes-2-theta-llama-3-8b", "nousresearch/hermes-3-llama-3.1-405b", "nousresearch/hermes-3-llama-3.1-70b", "nousresearch/nous-capybara-7b", "nousresearch/nous-hermes-2-mistral-7b-dpo", "nousresearch/nous-hermes-2-mixtral-8x7b-dpo", "nousresearch/nous-hermes-2-mixtral-8x7b-sft", "nousresearch/nous-hermes-llama2-13b", "nousresearch/nous-hermes-yi-34b", "---OPENAI---", "openai/chatgpt-4o-latest", "openai/gpt-3.5-turbo", "openai/gpt-3.5-turbo-0125", "openai/gpt-3.5-turbo-0301", "openai/gpt-3.5-turbo-0613", "openai/gpt-3.5-turbo-1106", "openai/gpt-3.5-turbo-16k", "openai/gpt-3.5-turbo-instruct", "openai/gpt-4", "openai/gpt-4-0314", "openai/gpt-4-1106-preview", "openai/gpt-4-32k", "openai/gpt-4-32k-0314", "openai/gpt-4-turbo", "openai/gpt-4-turbo-preview", "openai/gpt-4-vision-preview", "openai/gpt-4o", "openai/gpt-4o-2024-05-13", "openai/gpt-4o-2024-08-06", "openai/gpt-4o-mini", "openai/gpt-4o-mini-2024-07-18", "---PERPLEXITY---", "perplexity/llama-3-sonar-large-32k-chat", "perplexity/llama-3-sonar-large-32k-online", "perplexity/llama-3-sonar-small-32k-chat", "perplexity/llama-3-sonar-small-32k-online", "perplexity/llama-3.1-sonar-huge-128k-online", "perplexity/llama-3.1-sonar-large-128k-chat", "perplexity/llama-3.1-sonar-large-128k-online", "perplexity/llama-3.1-sonar-small-128k-chat", "perplexity/llama-3.1-sonar-small-128k-online", "---QWEN---", "qwen/qwen-110b-chat", "qwen/qwen-14b-chat", "qwen/qwen-2-72b-instruct", "qwen/qwen-2-7b-instruct", "qwen/qwen-32b-chat", "qwen/qwen-4b-chat", "qwen/qwen-72b-chat", "qwen/qwen-7b-chat", "---SAO10K---", "sao10k/fimbulvetr-11b-v2", "sao10k/l3-euryale-70b", "sao10k/l3-lunaris-8b", "sao10k/l3-stheno-8b", "---OTHERS---", "aetherwiing/mn-starcannon-12b", "allenai/olmo-7b-instruct", "alpindale/goliath-120b", "alpindale/magnum-72b", "austism/chronos-hermes-13b", "databricks/dbrx-instruct", "deepseek/deepseek-chat", "deepseek/deepseek-coder", "gryphe/mythomax-l2-13b", "gryphe/mythomist-7b", "jondurbin/airoboros-l2-70b", "lizpreciatior/lzlv-70b-fp16-hf", "mancer/weaver", "nothingiisreal/mn-celeste-12b", "open-orca/mistral-7b-openorca", "openchat/openchat-7b", "openchat/openchat-8b", "openrouter/flavor-of-the-week", "phind/phind-codellama-34b", "pygmalionai/mythalion-13b", "recursal/eagle-7b", "recursal/rwkv-5-3b-ai-town", "rwkv/rwkv-5-world-3b", "snowflake/snowflake-arctic-instruct", "sophosympatheia/midnight-rose-70b", "teknium/openhermes-2-mistral-7b", "teknium/openhermes-2.5-mistral-7b", "togethercomputer/stripedhyena-hessian-7b", "togethercomputer/stripedhyena-nous-7b", "undi95/remm-slerp-l2-13b", "undi95/toppy-m-7b", "xwin-lm/xwin-lm-70b" ] fetch: false dropParams: ["stop"] titleConvo: true titleModel: "gpt-3.5-turbo" summarize: false summaryModel: "gpt-3.5-turbo" forcePrompt: false modelDisplayLabel: "OpenRouter" # Preplexity # Model list: https://docs.perplexity.ai/docs/model-cards - name: "Perplexity" apiKey: "${PERPLEXITY_API_KEY}" baseURL: "https://api.perplexity.ai/" models: default: [ "llama-3.1-sonar-small-128k-chat", "llama-3.1-sonar-small-128k-online", "llama-3.1-sonar-large-128k-chat", "llama-3.1-sonar-large-128k-online", "llama-3.1-sonar-huge-128k-online", "llama-3.1-8b-instruct", "llama-3.1-70b-instruct" ] fetch: false # fetching list of models is not supported titleConvo: true titleModel: "llama-3.1-sonar-small-128k-chat" summarize: false summaryModel: "llama-3.1-sonar-small-128k-chat" forcePrompt: false dropParams: ["stop", "frequency_penalty"] modelDisplayLabel: "Perplexity" # ShuttleAI API - name: "ShuttleAI" apiKey: "${SHUTTLEAI_API_KEY}" baseURL: "https://api.shuttleai.app/v1" models: default: [ "shuttle-2-turbo", "shuttle-turbo", "gpt-4o-2024-05-13", "gpt-4o", "im-also-a-good-gpt2-chatbot", "gpt-4-turbo-2024-04-09", "gpt-4-turbo", "gpt-4-0125-preview", "gpt-4-turbo-preview", "gpt-4-1106-preview", "gpt-4-1106-vision-preview", "gpt-4-vision-preview", "gpt-4-0613", "gpt-4", "gpt-4-bing", "gpt-4-turbo-bing", "gpt-4-32k-0613", "gpt-4-32k", "gpt-3.5-turbo-0125", "gpt-3.5-turbo", "gpt-3.5-turbo-1106", "claude-3-opus-20240229", "claude-3-opus", "claude-3-sonnet-20240229", "claude-3-sonnet", "claude-3-haiku-20240307", "claude-3-haiku", "claude-2.1", "claude-2.0", "claude-2", "claude-instant-1.2", "claude-instant-1.1", "claude-instant-1.0", "claude-instant", "meta-llama-3-70b-instruct", "llama-3-70b-instruct", "meta-llama-3-8b-instruct", "llama-3-8b-instruct", "llama-3-sonar-large-32k-online", "llama-3-sonar-small-32k-online", "llama-3-sonar-large-32k-chat", "llama-3-sonar-small-32k-chat", "blackbox", "blackbox-code", "wizardlm-2-8x22b", "wizardlm-2-70b", "dolphin-2.6-mixtral-8x7b", "dolphin-mixtral-8x7b", "mistral-large", "mistral-next", "mistral-medium", "mistral-small", "mistral-tiny", "mixtral-8x7b-instruct-v0.1", "mixtral-8x7b-instruct", "mixtral-8x22b-instruct-v0.1", "mixtral-8x22b-instruct", "mistral-7b-instruct-v0.2", "mistral-7b-instruct-2", "mistral-7b-instruct-v0.1", "mistral-7b-instruct", "nous-hermes-2-mixtral-8x7b", "gemini-1.5-pro-latest", "gemini-1.5-pro", "gemini-1.0-pro-latest", "gemini-1.0-pro", "gemini-pro", "gemini-1.0-pro-vision", "gemini-pro-vision", "lzlv-70b", "figgs-rp", "cinematika-7b" ] fetch: true titleConvo: true titleMethod: "completion" titleModel: "shuttle-2-turbo" summarize: false summaryModel: "shuttle-2-turbo" forcePrompt: false dropParams: ["user", "frequency_penalty", "presence_penalty", "repition_penalty"] modelDisplayLabel: "ShuttleAI" - name: "together.ai" apiKey: "${TOGETHERAI_API_KEY}" baseURL: "https://api.together.xyz" models: default: [ "Austism/chronos-hermes-13b", "Gryphe/MythoMax-L2-13b", "HuggingFaceH4/zephyr-7b-beta", "NousResearch/Hermes-2-Theta-Llama-3-70B", "NousResearch/Nous-Capybara-7B-V1p9", "NousResearch/Nous-Hermes-2-Mistral-7B-DPO", "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", "NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT", "NousResearch/Nous-Hermes-2-Yi-34B", "NousResearch/Nous-Hermes-Llama2-13b", "NousResearch/Nous-Hermes-Llama2-70b", "NousResearch/Nous-Hermes-llama-2-7b", "Open-Orca/Mistral-7B-OpenOrca", "Qwen/Qwen1.5-0.5B-Chat", "Qwen/Qwen1.5-1.8B-Chat", "Qwen/Qwen1.5-110B-Chat", "Qwen/Qwen1.5-14B-Chat", "Qwen/Qwen1.5-32B-Chat", "Qwen/Qwen1.5-4B-Chat", "Qwen/Qwen1.5-72B-Chat", "Qwen/Qwen1.5-7B-Chat", "Qwen/Qwen2-1.5B-Instruct", "Qwen/Qwen2-72B-Instruct", "Qwen/Qwen2-7B-Instruct", "Snowflake/snowflake-arctic-instruct", "Undi95/ReMM-SLERP-L2-13B", "Undi95/Toppy-M-7B", "WizardLM/WizardLM-13B-V1.2", "allenai/OLMo-7B-Instruct", "carson/ml31405bit", "carson/ml3170bit", "carson/ml318bit", "carson/ml318br", "codellama/CodeLlama-13b-Instruct-hf", "codellama/CodeLlama-34b-Instruct-hf", "codellama/CodeLlama-70b-Instruct-hf", "codellama/CodeLlama-7b-Instruct-hf", "cognitivecomputations/dolphin-2.5-mixtral-8x7b", "databricks/dbrx-instruct", "deepseek-ai/deepseek-coder-33b-instruct", "deepseek-ai/deepseek-llm-67b-chat", "garage-bAInd/Platypus2-70B-instruct", "google/gemma-2-27b-it", "google/gemma-2-9b-it", "google/gemma-2b-it", "google/gemma-7b-it", "gradientai/Llama-3-70B-Instruct-Gradient-1048k", "lmsys/vicuna-13b-v1.3", "lmsys/vicuna-13b-v1.5", "lmsys/vicuna-13b-v1.5-16k", "lmsys/vicuna-7b-v1.3", "lmsys/vicuna-7b-v1.5", "meta-llama/Llama-2-13b-chat-hf", "meta-llama/Llama-2-70b-chat-hf", "meta-llama/Llama-2-7b-chat-hf", "meta-llama/Llama-3-70b-chat-hf", "meta-llama/Llama-3-8b-chat-hf", "meta-llama/Meta-Llama-3-70B-Instruct", "meta-llama/Meta-Llama-3-70B-Instruct-Lite", "meta-llama/Meta-Llama-3-70B-Instruct-Turbo", "meta-llama/Meta-Llama-3-8B-Instruct", "meta-llama/Meta-Llama-3-8B-Instruct-Lite", "meta-llama/Meta-Llama-3-8B-Instruct-Turbo", "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", "meta-llama/Meta-Llama-3.1-70B-Instruct-Reference", "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", "meta-llama/Meta-Llama-3.1-70B-Reference", "meta-llama/Meta-Llama-3.1-8B-Instruct-Reference", "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo", "microsoft/WizardLM-2-8x22B", "mistralai/Mistral-7B-Instruct-v0.1", "mistralai/Mistral-7B-Instruct-v0.2", "mistralai/Mistral-7B-Instruct-v0.3", "mistralai/Mixtral-8x22B-Instruct-v0.1", "mistralai/Mixtral-8x7B-Instruct-v0.1", "openchat/openchat-3.5-1210", "snorkelai/Snorkel-Mistral-PairRM-DPO", "teknium/OpenHermes-2-Mistral-7B", "teknium/OpenHermes-2p5-Mistral-7B", "togethercomputer/CodeLlama-13b-Instruct", "togethercomputer/CodeLlama-34b-Instruct", "togethercomputer/CodeLlama-7b-Instruct", "togethercomputer/Koala-13B", "togethercomputer/Koala-7B", "togethercomputer/Llama-2-7B-32K-Instruct", "togethercomputer/Llama-3-8b-chat-hf-int4", "togethercomputer/Llama-3-8b-chat-hf-int8", "togethercomputer/SOLAR-10.7B-Instruct-v1.0-int4", "togethercomputer/StripedHyena-Nous-7B", "togethercomputer/alpaca-7b", "togethercomputer/guanaco-13b", "togethercomputer/guanaco-33b", "togethercomputer/guanaco-65b", "togethercomputer/guanaco-7b", "togethercomputer/llama-2-13b-chat", "togethercomputer/llama-2-70b-chat", "togethercomputer/llama-2-7b-chat", "upstage/SOLAR-10.7B-Instruct-v1.0", "zero-one-ai/Yi-34B-Chat" ] fetch: false titleConvo: true titleModel: "togethercomputer/llama-2-7b-chat" summarize: false summaryModel: "togethercomputer/llama-2-7b-chat" forcePrompt: false modelDisplayLabel: "together.ai" # Unify # Model list: https://unify.ai/chat - name: "Unify" apiKey: "${UNIFY_API_KEY}" baseURL: "https://api.unify.ai/v0/" models: default: [ "router@q:1|c:2.12e-01|t:5.00e-04|i:2.78e-04", "chatgpt-4o-latest@openai", "gpt-3.5-turbo@openai", "gpt-4-turbo@openai", "gpt-4@openai", "gpt-4o-2024-08-06@openai", "gpt-4o-mini@openai", "gpt-4o@openai", "claude-3-haiku@anthropic", "claude-3-opus@anthropic", "claude-3-sonnet@anthropic", "claude-3.5-sonnet@anthropic", "claude-3-haiku@aws-bedrock", "claude-3-opus@aws-bedrock", "claude-3-sonnet@aws-bedrock", "claude-3.5-sonnet@aws-bedrock", "command-r-plus@aws-bedrock", "llama-3-70b-chat@aws-bedrock", "llama-3-8b-chat@aws-bedrock", "llama-3.1-405b-chat@aws-bedrock", "llama-3.1-70b-chat@aws-bedrock", "llama-3.1-8b-chat@aws-bedrock", "mistral-7b-instruct-v0.2@aws-bedrock", "mistral-large@aws-bedrock", "mixtral-8x7b-instruct-v0.1@aws-bedrock", "codellama-13b-instruct@fireworks-ai", "codellama-34b-instruct@fireworks-ai", "gemma-2-9b-it@fireworks-ai", "gemma-7b-it@fireworks-ai", "llama-3-70b-chat@fireworks-ai", "llama-3-8b-chat@fireworks-ai", "llama-3.1-405b-chat@fireworks-ai", "llama-3.1-70b-chat@fireworks-ai", "llama-3.1-8b-chat@fireworks-ai", "mistral-7b-instruct-v0.1@fireworks-ai", "mistral-7b-instruct-v0.2@fireworks-ai", "mistral-7b-instruct-v0.3@fireworks-ai", "mistral-nemo@fireworks-ai", "mixtral-8x22b-instruct-v0.1@fireworks-ai", "mixtral-8x7b-instruct-v0.1@fireworks-ai", "qwen-2-72b-instruct@fireworks-ai", "codellama-13b-instruct@octoai", "codellama-34b-instruct@octoai", "codellama-7b-instruct@octoai", "llama-3-70b-chat@octoai", "llama-3-8b-chat@octoai", "llama-3.1-405b-chat@octoai", "llama-3.1-70b-chat@octoai", "llama-3.1-8b-chat@octoai", "mistral-7b-instruct-v0.2@octoai", "mistral-7b-instruct-v0.3@octoai", "mixtral-8x22b-instruct-v0.1@octoai", "mixtral-8x7b-instruct-v0.1@octoai", "qwen-2-7b-instruct@octoai", "codellama-13b-instruct@together-ai", "codellama-34b-instruct@together-ai", "codellama-70b-instruct@together-ai", "codellama-7b-instruct@together-ai", "deepseek-coder-33b-instruct@together-ai", "gemma-2b-it@together-ai", "gemma-7b-it@together-ai", "llama-3-70b-chat@together-ai", "llama-3-8b-chat@together-ai", "llama-3.1-405b-chat@together-ai", "llama-3.1-70b-chat@together-ai", "llama-3.1-8b-chat@together-ai", "mistral-7b-instruct-v0.1@together-ai", "mistral-7b-instruct-v0.2@together-ai", "mistral-7b-instruct-v0.3@together-ai", "mixtral-8x22b-instruct-v0.1@together-ai", "mixtral-8x7b-instruct-v0.1@together-ai", "phind-codellama-34b-v2@together-ai", "qwen-2-72b-instruct@together-ai", "codellama-34b-instruct@deepinfra", "gemma-2-27b-it@deepinfra", "gemma-2-9b-it@deepinfra", "gemma-7b-it@deepinfra", "llama-3-70b-chat@deepinfra", "llama-3-8b-chat@deepinfra", "llama-3.1-405b-chat@deepinfra", "llama-3.1-70b-chat@deepinfra", "llama-3.1-8b-chat@deepinfra", "mistral-7b-instruct-v0.1@deepinfra", "mistral-7b-instruct-v0.3@deepinfra", "mixtral-8x22b-instruct-v0.1@deepinfra", "mixtral-8x7b-instruct-v0.1@deepinfra", "nemotron-4-340b-instruct@deepinfra", "phi-3-medium-4k-instruct@deepinfra", "phind-codellama-34b-v2@deepinfra", "qwen-2-72b-instruct@deepinfra", "qwen-2-7b-instruct@deepinfra", "codellama-34b-instruct@perplexity-ai", "llama-3.1-70b-chat@perplexity-ai", "llama-3.1-8b-chat@perplexity-ai", "mistral-7b-instruct-v0.2@perplexity-ai", "mixtral-8x7b-instruct-v0.1@perplexity-ai", "gemini-1.5-flash@vertex-ai", "gemini-1.5-pro@vertex-ai", "gemma-2-9b-it@vertex-ai", "gemma-2-9b-it@groq", "gemma-7b-it@groq", "llama-3-70b-chat@groq", "llama-3-8b-chat@groq", "mixtral-8x7b-instruct-v0.1@groq", "gemma-7b-it@lepton-ai", "llama-3-70b-chat@lepton-ai", "llama-3-8b-chat@lepton-ai", "llama-3.1-405b-chat@lepton-ai", "llama-3.1-70b-chat@lepton-ai", "llama-3.1-8b-chat@lepton-ai", "mistral-7b-instruct-v0.3@lepton-ai", "mixtral-8x7b-instruct-v0.1@lepton-ai", "gpt-4o-mini@azure-ai", "gpt-4o@azure-ai", "llama-3.1-405b-chat@azure-ai", "llama-3.1-70b-chat@azure-ai", "llama-3.1-8b-chat@azure-ai", "llama-3-70b-chat@replicate", "llama-3-8b-chat@replicate", "llama-3.1-405b-chat@replicate", "mistral-7b-instruct-v0.2@replicate", "mixtral-8x7b-instruct-v0.1@replicate", "mistral-7b-instruct-v0.2@mistral-ai", "mistral-7b-instruct-v0.3@mistral-ai", "mistral-large@mistral-ai", "mistral-nemo@mistral-ai", "mistral-small@mistral-ai", "mixtral-8x22b-instruct-v0.1@mistral-ai", "mixtral-8x7b-instruct-v0.1@mistral-ai", ] fetch: false titleConvo: true titleModel: "router@q:1|c:2.12e-01|t:5.00e-04|i:2.78e-04" dropParams: ["stop", "user", "frequency_penalty", "presence_penalty"]