moriire commited on
Commit
d02d8d7
1 Parent(s): f9aab03

Update app/llm.py

Browse files
Files changed (1) hide show
  1. app/llm.py +3 -3
app/llm.py CHANGED
@@ -39,8 +39,8 @@ llm_chat = llama_cpp.Llama.from_pretrained(
39
  #chat_format="llama-2"
40
  )
41
  llm_generate = llama_cpp.Llama.from_pretrained(
42
- repo_id="Qwen/Qwen1.5-0.5B-Chat-GGUF",
43
- filename="*q4_0.gguf",
44
  tokenizer=llama_cpp.llama_tokenizer.LlamaHFTokenizer.from_pretrained("Qwen/Qwen1.5-0.5B"),
45
  verbose=False,
46
  n_ctx=4096,
@@ -75,7 +75,7 @@ def health():
75
  # Chat Completion API
76
  @llm_router.post("/chat/", tags=["llm"])
77
  async def chat(chatm:ChatModel):#, user: schemas.BaseUser = fastapi.Depends(current_active_user)):
78
- #chatm.system = chatm.system.format(user.email)
79
  try:
80
  st = time()
81
  output = llm_chat.create_chat_completion(
 
39
  #chat_format="llama-2"
40
  )
41
  llm_generate = llama_cpp.Llama.from_pretrained(
42
+ repo_id="moriire/healthcare-GGUF",
43
+ filename="*.gguf",
44
  tokenizer=llama_cpp.llama_tokenizer.LlamaHFTokenizer.from_pretrained("Qwen/Qwen1.5-0.5B"),
45
  verbose=False,
46
  n_ctx=4096,
 
75
  # Chat Completion API
76
  @llm_router.post("/chat/", tags=["llm"])
77
  async def chat(chatm:ChatModel):#, user: schemas.BaseUser = fastapi.Depends(current_active_user)):
78
+ chatm.system = chatm.system.format("")#user.email)
79
  try:
80
  st = time()
81
  output = llm_chat.create_chat_completion(