| | import { env } from "$env/dynamic/private"; |
| | import { buildPrompt } from "$lib/buildPrompt"; |
| | import type { TextGenerationStreamOutput } from "@huggingface/inference"; |
| | import type { Endpoint } from "../endpoints"; |
| | import { z } from "zod"; |
| | import { logger } from "$lib/server/logger"; |
| |
|
| | export const endpointLlamacppParametersSchema = z.object({ |
| | weight: z.number().int().positive().default(1), |
| | model: z.any(), |
| | type: z.literal("llamacpp"), |
| | url: z.string().url().default("http://127.0.0.1:8080"), |
| | baseURL: z.string().url().optional(), |
| | accessToken: z.string().default(env.HF_TOKEN ?? env.HF_ACCESS_TOKEN), |
| | }); |
| |
|
| | export function endpointLlamacpp( |
| | input: z.input<typeof endpointLlamacppParametersSchema> |
| | ): Endpoint { |
| | const { baseURL, url, model } = endpointLlamacppParametersSchema.parse(input); |
| | return async ({ messages, preprompt, continueMessage, generateSettings }) => { |
| | const prompt = await buildPrompt({ |
| | messages, |
| | continueMessage, |
| | preprompt, |
| | model, |
| | }); |
| |
|
| | const parameters = { ...model.parameters, ...generateSettings }; |
| |
|
| | const r = await fetch(`${baseURL ?? url}/completion`, { |
| | method: "POST", |
| | headers: { |
| | "Content-Type": "application/json", |
| | }, |
| | body: JSON.stringify({ |
| | prompt, |
| | stream: true, |
| | temperature: parameters.temperature, |
| | top_p: parameters.top_p, |
| | top_k: parameters.top_k, |
| | stop: parameters.stop, |
| | repeat_penalty: parameters.repetition_penalty, |
| | n_predict: parameters.max_new_tokens, |
| | cache_prompt: true, |
| | }), |
| | }); |
| |
|
| | if (!r.ok) { |
| | throw new Error(`Failed to generate text: ${await r.text()}`); |
| | } |
| |
|
| | const encoder = new TextDecoderStream(); |
| | const reader = r.body?.pipeThrough(encoder).getReader(); |
| |
|
| | return (async function* () { |
| | let stop = false; |
| | let generatedText = ""; |
| | let tokenId = 0; |
| | let accumulatedData = ""; |
| |
|
| | while (!stop) { |
| | |
| | const out = (await reader?.read()) ?? { done: false, value: undefined }; |
| |
|
| | |
| | if (out.done) { |
| | reader?.cancel(); |
| | return; |
| | } |
| |
|
| | if (!out.value) { |
| | return; |
| | } |
| |
|
| | |
| | accumulatedData += out.value; |
| |
|
| | |
| | while (accumulatedData.includes("\n")) { |
| | |
| | const endIndex = accumulatedData.indexOf("\n"); |
| | let jsonString = accumulatedData.substring(0, endIndex).trim(); |
| |
|
| | |
| | accumulatedData = accumulatedData.substring(endIndex + 1); |
| |
|
| | if (jsonString.startsWith("data: ")) { |
| | jsonString = jsonString.slice(6); |
| | let data = null; |
| |
|
| | try { |
| | data = JSON.parse(jsonString); |
| | } catch (e) { |
| | logger.error(e, "Failed to parse JSON"); |
| | logger.error(jsonString, "Problematic JSON string:"); |
| | continue; |
| | } |
| |
|
| | |
| | if (data.content || data.stop) { |
| | generatedText += data.content; |
| | const output: TextGenerationStreamOutput = { |
| | token: { |
| | id: tokenId++, |
| | text: data.content ?? "", |
| | logprob: 0, |
| | special: false, |
| | }, |
| | generated_text: data.stop ? generatedText : null, |
| | details: null, |
| | }; |
| | if (data.stop) { |
| | stop = true; |
| | output.token.special = true; |
| | reader?.cancel(); |
| | } |
| | yield output; |
| | } |
| | } |
| | } |
| | } |
| | })(); |
| | }; |
| | } |
| |
|
| | export default endpointLlamacpp; |
| |
|