Louis 6aae985a55
re-#143: use OpenAI decoder and mutate final result from client (#164)
* chore: use OpenAI parser

* chore: access host's services

* chore: take out llm service - GGUF model for the latest llama.cpp support
2023-09-12 16:29:26 +07:00

27 lines
568 B
TypeScript

import { OpenAI } from "openai-streams";
export async function POST(req: Request) {
const { messages } = await req.json();
if (!messages) {
return new Response(null, {
status: 400,
statusText: "Did not include `messages` parameter",
});
}
const completionsStream = await OpenAI(
"chat",
{
model: "gpt-3.5-turbo",
stream: true,
messages,
max_tokens: 500,
},
{
apiBase: process.env.OPENAPI_ENDPOINT,
apiKey: process.env.OPENAPI_KEY,
}
);
return new Response(completionsStream);
}