Path: blob/master/src/packages/next/pages/api/v2/llm/evaluate.ts
1451 views
// This is the new endpoint for querying any LLM1// Previously, this has been in openai/chatgpt23import { evaluate } from "@cocalc/server/llm/index";4import { analytics_cookie_name } from "@cocalc/util/misc";5import getAccountId from "lib/account/get-account";6import getParams from "lib/api/get-params";78export default async function handle(req, res) {9try {10const result = await doIt(req);11res.json({ ...result, success: true });12} catch (err) {13res.json({ error: `${err.message}` });14return;15}16}1718async function doIt(req) {19const { input, system, history, model, tag } = getParams(req);20const account_id = await getAccountId(req);21const analytics_cookie = req.cookies[analytics_cookie_name];22return {23output: await evaluate({24account_id,25analytics_cookie,26input,27system,28history,29model,30tag,31}),32};33}343536