mirror of
https://github.com/danny-avila/LibreChat.git
synced 2025-12-20 02:10:15 +01:00
* WIP: app.locals refactoring
WIP: appConfig
fix: update memory configuration retrieval to use getAppConfig based on user role
fix: update comment for AppConfig interface to clarify purpose
🏷️ refactor: Update tests to use getAppConfig for endpoint configurations
ci: Update AppService tests to initialize app config instead of app.locals
ci: Integrate getAppConfig into remaining tests
refactor: Update multer storage destination to use promise-based getAppConfig and improve error handling in tests
refactor: Rename initializeAppConfig to setAppConfig and update related tests
ci: Mock getAppConfig in various tests to provide default configurations
refactor: Update convertMCPToolsToPlugins to use mcpManager for server configuration and adjust related tests
chore: rename `Config/getAppConfig` -> `Config/app`
fix: streamline OpenAI image tools configuration by removing direct appConfig dependency and using function parameters
chore: correct parameter documentation for imageOutputType in ToolService.js
refactor: remove `getCustomConfig` dependency in config route
refactor: update domain validation to use appConfig for allowed domains
refactor: use appConfig registration property
chore: remove app parameter from AppService invocation
refactor: update AppConfig interface to correct registration and turnstile configurations
refactor: remove getCustomConfig dependency and use getAppConfig in PluginController, multer, and MCP services
refactor: replace getCustomConfig with getAppConfig in STTService, TTSService, and related files
refactor: replace getCustomConfig with getAppConfig in Conversation and Message models, update tempChatRetention functions to use AppConfig type
refactor: update getAppConfig calls in Conversation and Message models to include user role for temporary chat expiration
ci: update related tests
refactor: update getAppConfig call in getCustomConfigSpeech to include user role
fix: update appConfig usage to access allowedDomains from actions instead of registration
refactor: enhance AppConfig to include fileStrategies and update related file strategy logic
refactor: update imports to use normalizeEndpointName from @librechat/api and remove redundant definitions
chore: remove deprecated unused RunManager
refactor: get balance config primarily from appConfig
refactor: remove customConfig dependency for appConfig and streamline loadConfigModels logic
refactor: remove getCustomConfig usage and use app config in file citations
refactor: consolidate endpoint loading logic into loadEndpoints function
refactor: update appConfig access to use endpoints structure across various services
refactor: implement custom endpoints configuration and streamline endpoint loading logic
refactor: update getAppConfig call to include user role parameter
refactor: streamline endpoint configuration and enhance appConfig usage across services
refactor: replace getMCPAuthMap with getUserMCPAuthMap and remove unused getCustomConfig file
refactor: add type annotation for loadedEndpoints in loadEndpoints function
refactor: move /services/Files/images/parse to TS API
chore: add missing FILE_CITATIONS permission to IRole interface
refactor: restructure toolkits to TS API
refactor: separate manifest logic into its own module
refactor: consolidate tool loading logic into a new tools module for startup logic
refactor: move interface config logic to TS API
refactor: migrate checkEmailConfig to TypeScript and update imports
refactor: add FunctionTool interface and availableTools to AppConfig
refactor: decouple caching and DB operations from AppService, make part of consolidated `getAppConfig`
WIP: fix tests
* fix: rebase conflicts
* refactor: remove app.locals references
* refactor: replace getBalanceConfig with getAppConfig in various strategies and middleware
* refactor: replace appConfig?.balance with getBalanceConfig in various controllers and clients
* test: add balance configuration to titleConvo method in AgentClient tests
* chore: remove unused `openai-chat-tokens` package
* chore: remove unused imports in initializeMCPs.js
* refactor: update balance configuration to use getAppConfig instead of getBalanceConfig
* refactor: integrate configMiddleware for centralized configuration handling
* refactor: optimize email domain validation by removing unnecessary async calls
* refactor: simplify multer storage configuration by removing async calls
* refactor: reorder imports for better readability in user.js
* refactor: replace getAppConfig calls with req.config for improved performance
* chore: replace getAppConfig calls with req.config in tests for centralized configuration handling
* chore: remove unused override config
* refactor: add configMiddleware to endpoint route and replace getAppConfig with req.config
* chore: remove customConfig parameter from TTSService constructor
* refactor: pass appConfig from request to processFileCitations for improved configuration handling
* refactor: remove configMiddleware from endpoint route and retrieve appConfig directly in getEndpointsConfig if not in `req.config`
* test: add mockAppConfig to processFileCitations tests for improved configuration handling
* fix: pass req.config to hasCustomUserVars and call without await after synchronous refactor
* fix: type safety in useExportConversation
* refactor: retrieve appConfig using getAppConfig in PluginController and remove configMiddleware from plugins route, to avoid always retrieving when plugins are cached
* chore: change `MongoUser` typedef to `IUser`
* fix: Add `user` and `config` fields to ServerRequest and update JSDoc type annotations from Express.Request to ServerRequest
* fix: remove unused setAppConfig mock from Server configuration tests
176 lines
5.2 KiB
TypeScript
176 lines
5.2 KiB
TypeScript
import { ErrorTypes, EModelEndpoint, mapModelToAzureConfig } from 'librechat-data-provider';
|
|
import type {
|
|
InitializeOpenAIOptionsParams,
|
|
OpenAIOptionsResult,
|
|
OpenAIConfigOptions,
|
|
UserKeyValues,
|
|
} from '~/types';
|
|
import { createHandleLLMNewToken } from '~/utils/generators';
|
|
import { getAzureCredentials } from '~/utils/azure';
|
|
import { isUserProvided } from '~/utils/common';
|
|
import { resolveHeaders } from '~/utils/env';
|
|
import { getOpenAIConfig } from './llm';
|
|
|
|
/**
|
|
* Initializes OpenAI options for agent usage. This function always returns configuration
|
|
* options and never creates a client instance (equivalent to optionsOnly=true behavior).
|
|
*
|
|
* @param params - Configuration parameters
|
|
* @returns Promise resolving to OpenAI configuration options
|
|
* @throws Error if API key is missing or user key has expired
|
|
*/
|
|
export const initializeOpenAI = async ({
|
|
req,
|
|
appConfig,
|
|
overrideModel,
|
|
endpointOption,
|
|
overrideEndpoint,
|
|
getUserKeyValues,
|
|
checkUserKeyExpiry,
|
|
}: InitializeOpenAIOptionsParams): Promise<OpenAIOptionsResult> => {
|
|
const { PROXY, OPENAI_API_KEY, AZURE_API_KEY, OPENAI_REVERSE_PROXY, AZURE_OPENAI_BASEURL } =
|
|
process.env;
|
|
|
|
const { key: expiresAt } = req.body;
|
|
const modelName = overrideModel ?? req.body.model;
|
|
const endpoint = overrideEndpoint ?? req.body.endpoint;
|
|
|
|
if (!endpoint) {
|
|
throw new Error('Endpoint is required');
|
|
}
|
|
|
|
const credentials = {
|
|
[EModelEndpoint.openAI]: OPENAI_API_KEY,
|
|
[EModelEndpoint.azureOpenAI]: AZURE_API_KEY,
|
|
};
|
|
|
|
const baseURLOptions = {
|
|
[EModelEndpoint.openAI]: OPENAI_REVERSE_PROXY,
|
|
[EModelEndpoint.azureOpenAI]: AZURE_OPENAI_BASEURL,
|
|
};
|
|
|
|
const userProvidesKey = isUserProvided(credentials[endpoint as keyof typeof credentials]);
|
|
const userProvidesURL = isUserProvided(baseURLOptions[endpoint as keyof typeof baseURLOptions]);
|
|
|
|
let userValues: UserKeyValues | null = null;
|
|
if (expiresAt && (userProvidesKey || userProvidesURL)) {
|
|
checkUserKeyExpiry(expiresAt, endpoint);
|
|
userValues = await getUserKeyValues({ userId: req.user.id, name: endpoint });
|
|
}
|
|
|
|
let apiKey = userProvidesKey
|
|
? userValues?.apiKey
|
|
: credentials[endpoint as keyof typeof credentials];
|
|
const baseURL = userProvidesURL
|
|
? userValues?.baseURL
|
|
: baseURLOptions[endpoint as keyof typeof baseURLOptions];
|
|
|
|
const clientOptions: OpenAIConfigOptions = {
|
|
proxy: PROXY ?? undefined,
|
|
reverseProxyUrl: baseURL || undefined,
|
|
streaming: true,
|
|
};
|
|
|
|
const isAzureOpenAI = endpoint === EModelEndpoint.azureOpenAI;
|
|
const azureConfig = isAzureOpenAI && appConfig.endpoints?.[EModelEndpoint.azureOpenAI];
|
|
|
|
if (isAzureOpenAI && azureConfig) {
|
|
const { modelGroupMap, groupMap } = azureConfig;
|
|
const {
|
|
azureOptions,
|
|
baseURL: configBaseURL,
|
|
headers = {},
|
|
serverless,
|
|
} = mapModelToAzureConfig({
|
|
modelName: modelName || '',
|
|
modelGroupMap,
|
|
groupMap,
|
|
});
|
|
|
|
clientOptions.reverseProxyUrl = configBaseURL ?? clientOptions.reverseProxyUrl;
|
|
clientOptions.headers = resolveHeaders({
|
|
headers: { ...headers, ...(clientOptions.headers ?? {}) },
|
|
user: req.user,
|
|
});
|
|
|
|
const groupName = modelGroupMap[modelName || '']?.group;
|
|
if (groupName && groupMap[groupName]) {
|
|
clientOptions.addParams = groupMap[groupName]?.addParams;
|
|
clientOptions.dropParams = groupMap[groupName]?.dropParams;
|
|
}
|
|
|
|
apiKey = azureOptions.azureOpenAIApiKey;
|
|
clientOptions.azure = !serverless ? azureOptions : undefined;
|
|
|
|
if (serverless === true) {
|
|
clientOptions.defaultQuery = azureOptions.azureOpenAIApiVersion
|
|
? { 'api-version': azureOptions.azureOpenAIApiVersion }
|
|
: undefined;
|
|
|
|
if (!clientOptions.headers) {
|
|
clientOptions.headers = {};
|
|
}
|
|
clientOptions.headers['api-key'] = apiKey;
|
|
}
|
|
} else if (isAzureOpenAI) {
|
|
clientOptions.azure =
|
|
userProvidesKey && userValues?.apiKey ? JSON.parse(userValues.apiKey) : getAzureCredentials();
|
|
apiKey = clientOptions.azure?.azureOpenAIApiKey;
|
|
}
|
|
|
|
if (userProvidesKey && !apiKey) {
|
|
throw new Error(
|
|
JSON.stringify({
|
|
type: ErrorTypes.NO_USER_KEY,
|
|
}),
|
|
);
|
|
}
|
|
|
|
if (!apiKey) {
|
|
throw new Error(`${endpoint} API Key not provided.`);
|
|
}
|
|
|
|
const modelOptions = {
|
|
...endpointOption.model_parameters,
|
|
model: modelName,
|
|
user: req.user.id,
|
|
};
|
|
|
|
const finalClientOptions: OpenAIConfigOptions = {
|
|
...clientOptions,
|
|
modelOptions,
|
|
};
|
|
|
|
const options = getOpenAIConfig(apiKey, finalClientOptions, endpoint);
|
|
|
|
const openAIConfig = appConfig.endpoints?.[EModelEndpoint.openAI];
|
|
const allConfig = appConfig.endpoints?.all;
|
|
const azureRate = modelName?.includes('gpt-4') ? 30 : 17;
|
|
|
|
let streamRate: number | undefined;
|
|
|
|
if (isAzureOpenAI && azureConfig) {
|
|
streamRate = azureConfig.streamRate ?? azureRate;
|
|
} else if (!isAzureOpenAI && openAIConfig) {
|
|
streamRate = openAIConfig.streamRate;
|
|
}
|
|
|
|
if (allConfig?.streamRate) {
|
|
streamRate = allConfig.streamRate;
|
|
}
|
|
|
|
if (streamRate) {
|
|
options.llmConfig.callbacks = [
|
|
{
|
|
handleLLMNewToken: createHandleLLMNewToken(streamRate),
|
|
},
|
|
];
|
|
}
|
|
|
|
const result: OpenAIOptionsResult = {
|
|
...options,
|
|
streamRate,
|
|
};
|
|
|
|
return result;
|
|
};
|