Update app.py
Browse files
app.py
CHANGED
@@ -4,6 +4,7 @@ from fastapi.responses import StreamingResponse
|
|
4 |
import httpx
|
5 |
import json
|
6 |
import os
|
|
|
7 |
from helper import create_jwt
|
8 |
|
9 |
# 设置日志
|
@@ -41,8 +42,6 @@ async def chat(request: Request):
|
|
41 |
model = payload.get('model', 'claude-3-5-sonnet-20240620')
|
42 |
|
43 |
# 准备 LLM API 请求
|
44 |
-
url = "https://llm.zed.dev/completion?"
|
45 |
-
|
46 |
llm_payload = {
|
47 |
"provider": "anthropic",
|
48 |
"model": model,
|
@@ -58,10 +57,9 @@ async def chat(request: Request):
|
|
58 |
logger.debug(f"LLM payload: {llm_payload}")
|
59 |
|
60 |
headers = {
|
61 |
-
'Host': 'llm.zed.dev',
|
62 |
'accept': '*/*',
|
63 |
'content-type': 'application/json',
|
64 |
-
'authorization': f'Bearer {jwt_token}',
|
65 |
'user-agent': 'Zed/0.149.3 (macos; aarch64)'
|
66 |
}
|
67 |
logger.debug(f"Request headers: {headers}")
|
@@ -74,7 +72,7 @@ async def chat(request: Request):
|
|
74 |
async def generate():
|
75 |
async with httpx.AsyncClient(proxies=proxies) as client:
|
76 |
try:
|
77 |
-
async with client.stream('POST',
|
78 |
logger.debug(f"LLM API response status: {response.status_code}")
|
79 |
logger.debug(f"LLM API response headers: {response.headers}")
|
80 |
if response.status_code != 200:
|
|
|
4 |
import httpx
|
5 |
import json
|
6 |
import os
|
7 |
+
|
8 |
from helper import create_jwt
|
9 |
|
10 |
# 设置日志
|
|
|
42 |
model = payload.get('model', 'claude-3-5-sonnet-20240620')
|
43 |
|
44 |
# 准备 LLM API 请求
|
|
|
|
|
45 |
llm_payload = {
|
46 |
"provider": "anthropic",
|
47 |
"model": model,
|
|
|
57 |
logger.debug(f"LLM payload: {llm_payload}")
|
58 |
|
59 |
headers = {
|
|
|
60 |
'accept': '*/*',
|
61 |
'content-type': 'application/json',
|
62 |
+
'authorization': f'Bearer {jwt_token}',
|
63 |
'user-agent': 'Zed/0.149.3 (macos; aarch64)'
|
64 |
}
|
65 |
logger.debug(f"Request headers: {headers}")
|
|
|
72 |
async def generate():
|
73 |
async with httpx.AsyncClient(proxies=proxies) as client:
|
74 |
try:
|
75 |
+
async with client.stream('POST', 'https://llm.zed.dev/completion?', headers=headers, json=llm_payload) as response:
|
76 |
logger.debug(f"LLM API response status: {response.status_code}")
|
77 |
logger.debug(f"LLM API response headers: {response.headers}")
|
78 |
if response.status_code != 200:
|