From 9ca84edb9aa1b4b8d59e242358e0d8bd931b5a2b Mon Sep 17 00:00:00 2001 From: Danny Avila <110412045+danny-avila@users.noreply.github.com> Date: Fri, 10 Nov 2023 10:33:56 -0500 Subject: [PATCH] fix(openai/completions): use old method for instruct/davinci/text gen models (#1166) --- api/app/clients/OpenAIClient.js | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/api/app/clients/OpenAIClient.js b/api/app/clients/OpenAIClient.js index 124d05e68f..c7c973e482 100644 --- a/api/app/clients/OpenAIClient.js +++ b/api/app/clients/OpenAIClient.js @@ -367,7 +367,7 @@ If your reverse proxy is compatible to OpenAI specs in every other way, it may s let streamResult = null; this.modelOptions.user = this.user; const invalidBaseUrl = this.completionsUrl && extractBaseURL(this.completionsUrl) === null; - const useOldMethod = !!(this.azure || invalidBaseUrl); + const useOldMethod = !!(this.azure || invalidBaseUrl || !this.isChatCompletion); if (typeof opts.onProgress === 'function' && useOldMethod) { await this.getCompletion( payload, @@ -702,9 +702,10 @@ ${convo} if (typeof onProgress === 'function') { modelOptions.stream = true; } - if (this.isChatGptModel) { + if (this.isChatCompletion) { modelOptions.messages = payload; } else { + // TODO: unreachable code. Need to implement completions call for non-chat models modelOptions.prompt = payload; }