chat-ui_test / src /lib /server /modelEndpoint.ts
coyotte508's picture
coyotte508 HF staff
✨ Add model id to be able to switch models while keeping conversations valid (#181)
cf7ac8d unverified
raw
history blame
798 Bytes
import { HF_ACCESS_TOKEN } from "$env/static/private";
import { sum } from "$lib/utils/sum";
import type { BackendModel } from "./models";
/**
* Find a random load-balanced endpoint
*/
export function modelEndpoint(model: BackendModel): {
url: string;
authorization: string;
weight: number;
} {
if (!model.endpoints) {
return {
url: `https://api-inference.huggingface.co/models/${model.name}`,
authorization: `Bearer ${HF_ACCESS_TOKEN}`,
weight: 1,
};
}
const endpoints = model.endpoints;
const totalWeight = sum(endpoints.map((e) => e.weight));
let random = Math.random() * totalWeight;
for (const endpoint of endpoints) {
if (random < endpoint.weight) {
return endpoint;
}
random -= endpoint.weight;
}
throw new Error("Invalid config, no endpoint found");
}