mirror of
https://github.com/danny-avila/LibreChat.git
synced 2025-12-18 01:10:14 +01:00
* feat: Refactor ModelEndHandler to collect usage metadata only if it exists * feat: google tool end handling, custom anthropic class for better token ux * refactor: differentiate between client <> request options * feat: initial support for google agents * feat: only cache messages with non-empty text * feat: Cache non-empty messages in chatV2 controller * fix: anthropic llm client options llmConfig * refactor: streamline client options handling in LLM configuration * fix: VertexAI Agent Auth & Tool Handling * fix: additional fields for llmConfig, however customHeaders are not supported by langchain, requires PR * feat: set default location for vertexai LLM configuration * fix: outdated OpenAI Client options for getLLMConfig * chore: agent provider options typing * chore: add note about currently unsupported customHeaders in langchain GenAI client * fix: skip transaction creation when rawAmount is NaN
83 lines
2.1 KiB
JavaScript
83 lines
2.1 KiB
JavaScript
const { EModelEndpoint, AuthKeys } = require('librechat-data-provider');
|
|
const { getUserKey, checkUserKeyExpiry } = require('~/server/services/UserService');
|
|
const { getLLMConfig } = require('~/server/services/Endpoints/google/llm');
|
|
const { isEnabled } = require('~/server/utils');
|
|
const { GoogleClient } = require('~/app');
|
|
|
|
const initializeClient = async ({ req, res, endpointOption, overrideModel, optionsOnly }) => {
|
|
const {
|
|
GOOGLE_KEY,
|
|
GOOGLE_REVERSE_PROXY,
|
|
GOOGLE_AUTH_HEADER,
|
|
PROXY,
|
|
} = process.env;
|
|
const isUserProvided = GOOGLE_KEY === 'user_provided';
|
|
const { key: expiresAt } = req.body;
|
|
|
|
let userKey = null;
|
|
if (expiresAt && isUserProvided) {
|
|
checkUserKeyExpiry(expiresAt, EModelEndpoint.google);
|
|
userKey = await getUserKey({ userId: req.user.id, name: EModelEndpoint.google });
|
|
}
|
|
|
|
let serviceKey = {};
|
|
try {
|
|
serviceKey = require('~/data/auth.json');
|
|
} catch (e) {
|
|
// Do nothing
|
|
}
|
|
|
|
const credentials = isUserProvided
|
|
? userKey
|
|
: {
|
|
[AuthKeys.GOOGLE_SERVICE_KEY]: serviceKey,
|
|
[AuthKeys.GOOGLE_API_KEY]: GOOGLE_KEY,
|
|
};
|
|
|
|
let clientOptions = {};
|
|
|
|
/** @type {undefined | TBaseEndpoint} */
|
|
const allConfig = req.app.locals.all;
|
|
/** @type {undefined | TBaseEndpoint} */
|
|
const googleConfig = req.app.locals[EModelEndpoint.google];
|
|
|
|
if (googleConfig) {
|
|
clientOptions.streamRate = googleConfig.streamRate;
|
|
}
|
|
|
|
if (allConfig) {
|
|
clientOptions.streamRate = allConfig.streamRate;
|
|
}
|
|
|
|
clientOptions = {
|
|
req,
|
|
res,
|
|
reverseProxyUrl: GOOGLE_REVERSE_PROXY ?? null,
|
|
authHeader: isEnabled(GOOGLE_AUTH_HEADER) ?? null,
|
|
proxy: PROXY ?? null,
|
|
...clientOptions,
|
|
...endpointOption,
|
|
};
|
|
|
|
if (optionsOnly) {
|
|
clientOptions = Object.assign(
|
|
{
|
|
modelOptions: endpointOption.model_parameters,
|
|
},
|
|
clientOptions,
|
|
);
|
|
if (overrideModel) {
|
|
clientOptions.modelOptions.model = overrideModel;
|
|
}
|
|
return getLLMConfig(credentials, clientOptions);
|
|
}
|
|
|
|
const client = new GoogleClient(credentials, clientOptions);
|
|
|
|
return {
|
|
client,
|
|
credentials,
|
|
};
|
|
};
|
|
|
|
module.exports = initializeClient;
|