mirror of
https://github.com/danny-avila/LibreChat.git
synced 2025-09-22 06:00:56 +02:00

* feat: update PaLM icons * feat: add additional google models * POC: formatting inputs for Vertex AI streaming * refactor: move endpoints services outside of /routes dir to /services/Endpoints * refactor: shorten schemas import * refactor: rename PALM to GOOGLE * feat: make Google editable endpoint * feat: reusable Ask and Edit controllers based off Anthropic * chore: organize imports/logic * fix(parseConvo): include examples in googleSchema * fix: google only allows odd number of messages to be sent * fix: pass proxy to AnthropicClient * refactor: change `google` altName to `Google` * refactor: update getModelMaxTokens and related functions to handle maxTokensMap with nested endpoint model key/values * refactor: google Icon and response sender changes (Codey and Google logo instead of PaLM in all cases) * feat: google support for maxTokensMap * feat: google updated endpoints with Ask/Edit controllers, buildOptions, and initializeClient * feat(GoogleClient): now builds prompt for text models and supports real streaming from Vertex AI through langchain * chore(GoogleClient): remove comments, left before for reference in git history * docs: update google instructions (WIP) * docs(apis_and_tokens.md): add images to google instructions * docs: remove typo apis_and_tokens.md * Update apis_and_tokens.md * feat(Google): use default settings map, fully support context for both text and chat models, fully support examples for chat models * chore: update more PaLM references to Google * chore: move playwright out of workflows to avoid failing tests
27 lines
1 KiB
JavaScript
27 lines
1 KiB
JavaScript
const { processFiles } = require('~/server/services/Files');
|
|
const openAI = require('~/server/services/Endpoints/openAI');
|
|
const google = require('~/server/services/Endpoints/google');
|
|
const anthropic = require('~/server/services/Endpoints/anthropic');
|
|
const gptPlugins = require('~/server/services/Endpoints/gptPlugins');
|
|
const { parseConvo, EModelEndpoint } = require('~/server/services/Endpoints');
|
|
|
|
const buildFunction = {
|
|
[EModelEndpoint.openAI]: openAI.buildOptions,
|
|
[EModelEndpoint.google]: google.buildOptions,
|
|
[EModelEndpoint.azureOpenAI]: openAI.buildOptions,
|
|
[EModelEndpoint.anthropic]: anthropic.buildOptions,
|
|
[EModelEndpoint.gptPlugins]: gptPlugins.buildOptions,
|
|
};
|
|
|
|
function buildEndpointOption(req, res, next) {
|
|
const { endpoint } = req.body;
|
|
const parsedBody = parseConvo(endpoint, req.body);
|
|
req.body.endpointOption = buildFunction[endpoint](endpoint, parsedBody);
|
|
if (req.body.files) {
|
|
// hold the promise
|
|
req.body.endpointOption.attachments = processFiles(req.body.files);
|
|
}
|
|
next();
|
|
}
|
|
|
|
module.exports = buildEndpointOption;
|