From 5eb0a3ad90f638610bb480203ea2b9c51fab1e11 Mon Sep 17 00:00:00 2001 From: Danny Avila Date: Wed, 4 Feb 2026 11:02:27 +0100 Subject: [PATCH] =?UTF-8?q?=E2=9A=A0=EF=B8=8F=20chore:=20Remove=20Deprecat?= =?UTF-8?q?ed=20`forcePrompt`=20setting=20(#11622)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit - Removed `forcePrompt` parameter from various configuration files including `librechat.example.yaml`, `initialize.js`, `values.yaml`, and `initialize.ts`. - This change simplifies the configuration by eliminating unused options, enhancing clarity and maintainability across the codebase. --- api/server/services/Endpoints/azureAssistants/initialize.js | 1 - helm/librechat/values.yaml | 1 - librechat.example.yaml | 4 ---- packages/api/src/endpoints/custom/initialize.ts | 1 - packages/api/src/endpoints/openai/config.spec.ts | 3 --- packages/data-provider/src/config.ts | 2 -- 6 files changed, 12 deletions(-) diff --git a/api/server/services/Endpoints/azureAssistants/initialize.js b/api/server/services/Endpoints/azureAssistants/initialize.js index 6a9118ea8a..e81f0bcd8a 100644 --- a/api/server/services/Endpoints/azureAssistants/initialize.js +++ b/api/server/services/Endpoints/azureAssistants/initialize.js @@ -128,7 +128,6 @@ const initializeClient = async ({ req, res, version, endpointOption, initAppClie const groupName = modelGroupMap[modelName].group; clientOptions.addParams = azureConfig.groupMap[groupName].addParams; clientOptions.dropParams = azureConfig.groupMap[groupName].dropParams; - clientOptions.forcePrompt = azureConfig.groupMap[groupName].forcePrompt; clientOptions.reverseProxyUrl = baseURL ?? clientOptions.reverseProxyUrl; clientOptions.headers = opts.defaultHeaders; diff --git a/helm/librechat/values.yaml b/helm/librechat/values.yaml index a15b681de2..c6461ade61 100755 --- a/helm/librechat/values.yaml +++ b/helm/librechat/values.yaml @@ -97,7 +97,6 @@ librechat: # titleModel: "gpt-3.5-turbo" # summarize: false # summaryModel: "gpt-3.5-turbo" - # forcePrompt: false # modelDisplayLabel: "OpenRouter" # name of existing Yaml configmap, key must be librechat.yaml diff --git a/librechat.example.yaml b/librechat.example.yaml index 6d2b587dd8..aa5b4e8cd8 100644 --- a/librechat.example.yaml +++ b/librechat.example.yaml @@ -374,9 +374,6 @@ endpoints: # Summary Model: Specify the model to use if summarization is enabled. # summaryModel: "mistral-tiny" # Defaults to "gpt-3.5-turbo" if omitted. - # Force Prompt setting: If true, sends a `prompt` parameter instead of `messages`. - # forcePrompt: false - # The label displayed for the AI model in messages. modelDisplayLabel: 'Mistral' # Default is "AI" when not set. @@ -436,7 +433,6 @@ endpoints: titleModel: 'current_model' summarize: false summaryModel: 'current_model' - forcePrompt: false modelDisplayLabel: 'Portkey' iconURL: https://images.crunchbase.com/image/upload/c_pad,f_auto,q_auto:eco,dpr_1/rjqy7ghvjoiu4cd1xjbf diff --git a/packages/api/src/endpoints/custom/initialize.ts b/packages/api/src/endpoints/custom/initialize.ts index 4550fa9f5b..7930b1c12f 100644 --- a/packages/api/src/endpoints/custom/initialize.ts +++ b/packages/api/src/endpoints/custom/initialize.ts @@ -31,7 +31,6 @@ function buildCustomOptions( customParams: endpointConfig.customParams, titleConvo: endpointConfig.titleConvo, titleModel: endpointConfig.titleModel, - forcePrompt: endpointConfig.forcePrompt, summaryModel: endpointConfig.summaryModel, modelDisplayLabel: endpointConfig.modelDisplayLabel, titleMethod: endpointConfig.titleMethod ?? 'completion', diff --git a/packages/api/src/endpoints/openai/config.spec.ts b/packages/api/src/endpoints/openai/config.spec.ts index 405873490f..93864649f9 100644 --- a/packages/api/src/endpoints/openai/config.spec.ts +++ b/packages/api/src/endpoints/openai/config.spec.ts @@ -1300,7 +1300,6 @@ describe('getOpenAIConfig', () => { max_completion_tokens: 4000, }, dropParams: ['frequency_penalty'], - forcePrompt: false, modelOptions: { model: modelName, user: 'azure-user-123', @@ -1395,7 +1394,6 @@ describe('getOpenAIConfig', () => { dropParams: ['presence_penalty'], titleConvo: true, titleModel: 'gpt-3.5-turbo', - forcePrompt: false, summaryModel: 'gpt-3.5-turbo', modelDisplayLabel: 'Custom GPT-4', titleMethod: 'completion', @@ -1414,7 +1412,6 @@ describe('getOpenAIConfig', () => { customParams: {}, titleConvo: endpointConfig.titleConvo, titleModel: endpointConfig.titleModel, - forcePrompt: endpointConfig.forcePrompt, summaryModel: endpointConfig.summaryModel, modelDisplayLabel: endpointConfig.modelDisplayLabel, titleMethod: endpointConfig.titleMethod, diff --git a/packages/data-provider/src/config.ts b/packages/data-provider/src/config.ts index 03073e32f8..6402f30670 100644 --- a/packages/data-provider/src/config.ts +++ b/packages/data-provider/src/config.ts @@ -122,7 +122,6 @@ export const azureBaseSchema = z.object({ assistants: z.boolean().optional(), addParams: z.record(z.any()).optional(), dropParams: z.array(z.string()).optional(), - forcePrompt: z.boolean().optional(), version: z.string().optional(), baseURL: z.string().optional(), additionalHeaders: z.record(z.any()).optional(), @@ -321,7 +320,6 @@ export const endpointSchema = baseEndpointSchema.merge( summarize: z.boolean().optional(), summaryModel: z.string().optional(), iconURL: z.string().optional(), - forcePrompt: z.boolean().optional(), modelDisplayLabel: z.string().optional(), headers: z.record(z.any()).optional(), addParams: z.record(z.any()).optional(),