added support for openllm
Browse files
app_modules/llm_loader.py
CHANGED
@@ -8,7 +8,7 @@ import torch
|
|
8 |
from langchain.callbacks.base import BaseCallbackHandler
|
9 |
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
|
10 |
from langchain.chat_models import ChatOpenAI
|
11 |
-
from langchain.llms import GPT4All, HuggingFacePipeline, LlamaCpp
|
12 |
from langchain.schema import LLMResult
|
13 |
from transformers import (
|
14 |
AutoConfig,
|
@@ -156,6 +156,14 @@ class LLMLoader:
|
|
156 |
verbose=True,
|
157 |
temperature=0,
|
158 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
159 |
elif self.llm_model_type.startswith("gpt4all"):
|
160 |
MODEL_PATH = ensure_model_is_downloaded(self.llm_model_type)
|
161 |
self.llm = GPT4All(
|
|
|
8 |
from langchain.callbacks.base import BaseCallbackHandler
|
9 |
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
|
10 |
from langchain.chat_models import ChatOpenAI
|
11 |
+
from langchain.llms import GPT4All, HuggingFacePipeline, LlamaCpp, OpenLLM
|
12 |
from langchain.schema import LLMResult
|
13 |
from transformers import (
|
14 |
AutoConfig,
|
|
|
156 |
verbose=True,
|
157 |
temperature=0,
|
158 |
)
|
159 |
+
elif self.llm_model_type == "openllm":
|
160 |
+
server_url = os.environ.get("OPENLLM_SERVER_URL")
|
161 |
+
print(f" server url: {server_url}")
|
162 |
+
self.llm = OpenLLM(
|
163 |
+
server_url=server_url,
|
164 |
+
callbacks=callbacks,
|
165 |
+
verbose=True,
|
166 |
+
)
|
167 |
elif self.llm_model_type.startswith("gpt4all"):
|
168 |
MODEL_PATH = ensure_model_is_downloaded(self.llm_model_type)
|
169 |
self.llm = GPT4All(
|