Path: blob/master/src/packages/next/components/openai/vendor-status-check.tsx
1450 views
import {1LLMServiceName,2getLLMServiceStatusCheckMD,3} from "@cocalc/util/db-schema/llm-utils";4import { unreachable } from "@cocalc/util/misc";5import A from "components/misc/A";67import type { JSX } from "react";89export function LLMServiceStatusCheck({10service,11}: {12service: LLMServiceName;13}): JSX.Element {14switch (service) {15case "openai":16return (17<>18OpenAI <A href="https://status.openai.com/">status</A> and{" "}19<A href="https://downdetector.com/status/openai/">downdetector</A>.20</>21);2223case "google":24return (25<>26Google <A href="https://status.cloud.google.com">status</A> and{" "}27<A href="https://downdetector.com/status/google-cloud">28downdetector29</A>30.31</>32);3334case "ollama":35return (36<>37This Ollama based API endpoint does not have a status page. If you are38experiencing issues you have to check with the API service directly or39try again later.40</>41);4243case "custom_openai":44return (45<>46This Custom OpenAI API endpoint does not have a status page. If you47are experiencing issues you have to check with the API service48directly or try again later.49</>50);5152case "mistralai":53return (54<>55This Mistral based API endpoint does not have a status page. If you56are experiencing issues, use another model or try again later.57</>58);5960case "anthropic":61return (62<>63Anthropic <A href="https://status.anthropic.com/">status</A>.64</>65);6667case "user":68return <>{getLLMServiceStatusCheckMD("user")}</>;6970default:71unreachable(service);72}73return <></>;74}757677