import os import together_gradio from utils import get_app demo = get_app( models=[ "meta-llama/Llama-Vision-Free", "meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo", "meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo", "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo", "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", "meta-llama/Meta-Llama-3-8B-Instruct-Turbo", "meta-llama/Meta-Llama-3-70B-Instruct-Turbo", "meta-llama/Llama-3.2-3B-Instruct-Turbo", "meta-llama/Meta-Llama-3-8B-Instruct-Lite", "meta-llama/Meta-Llama-3-70B-Instruct-Lite", "meta-llama/Llama-3-8b-chat-hf", "meta-llama/Llama-3-70b-chat-hf", "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF", "Qwen/Qwen2.5-Coder-32B-Instruct", "microsoft/WizardLM-2-8x22B", "google/gemma-2-27b-it", "google/gemma-2-9b-it", "databricks/dbrx-instruct", "mistralai/Mixtral-8x7B-Instruct-v0.1", "mistralai/Mixtral-8x22B-Instruct-v0.1", "Qwen/Qwen2.5-7B-Instruct-Turbo", "Qwen/Qwen2.5-72B-Instruct-Turbo", "Qwen/Qwen2-72B-Instruct", "deepseek-ai/deepseek-llm-67b-chat", "google/gemma-2b-it", "Gryphe/MythoMax-L2-13b", "meta-llama/Llama-2-13b-chat-hf", "mistralai/Mistral-7B-Instruct-v0.1", "mistralai/Mistral-7B-Instruct-v0.2", "mistralai/Mistral-7B-Instruct-v0.3", "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", "togethercomputer/StripedHyena-Nous-7B", "upstage/SOLAR-10.7B-Instruct-v1.0", ], default_model="meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo", src=together_gradio.registry, accept_token=not os.getenv("TOGETHER_API_KEY"), multimodal=True, ) if __name__ == "__main__": demo.launch()