Disable custom models for the moment

We're running into GPU constraints and need to turn off custom models until we find a better provider or can hot-swap them.
This commit is contained in:
Kyle Corbitt
2023-08-25 16:01:51 -07:00
parent e0d0cc0df1
commit ead981b900
2 changed files with 17 additions and 3 deletions

View File

@@ -147,9 +147,10 @@ export default function OutputCell({
<ResponseLog <ResponseLog
time={response.receivedAt} time={response.receivedAt}
title="Response received from API" title="Response received from API"
message={`statusCode: ${response.statusCode ?? ""}\n ${ message={[
response.errorMessage ?? "" response.statusCode ? `Status: ${response.statusCode}\n` : "",
}`} response.errorMessage ?? "",
].join("")}
/> />
)} )}
</Fragment> </Fragment>

View File

@@ -17,10 +17,23 @@ const modelEndpoints: Record<OpenpipeChatInput["model"], string> = {
"NousResearch/Nous-Hermes-llama-2-7b": "https://ua1bpc6kv3dgge-8000.proxy.runpod.net/v1", "NousResearch/Nous-Hermes-llama-2-7b": "https://ua1bpc6kv3dgge-8000.proxy.runpod.net/v1",
}; };
const CUSTOM_MODELS_ENABLED = false;
export async function getCompletion( export async function getCompletion(
input: OpenpipeChatInput, input: OpenpipeChatInput,
onStream: ((partialOutput: OpenpipeChatOutput) => void) | null, onStream: ((partialOutput: OpenpipeChatOutput) => void) | null,
): Promise<CompletionResponse<OpenpipeChatOutput>> { ): Promise<CompletionResponse<OpenpipeChatOutput>> {
// Temporarily disable these models because of GPU constraints
if (!CUSTOM_MODELS_ENABLED) {
return {
type: "error",
message:
"We've disabled this model temporarily because of GPU capacity constraints. Check back later.",
autoRetry: false,
};
}
const { model, messages, ...rest } = input; const { model, messages, ...rest } = input;
const templatedPrompt = frontendModelProvider.models[model].templatePrompt?.(messages); const templatedPrompt = frontendModelProvider.models[model].templatePrompt?.(messages);