LibreChat/api/server/services/Endpoints
Danny Avila 40685f6eb4
🚀 refactor: Enhance Custom Endpoints, Message Logic, and Payload Handling (#2895)
* chore: use node-fetch for OpenAIClient fetch key for non-crashing usage of AbortController in Bun runtime

* chore: variable order

* fix(useSSE): prevent finalHandler call in abortConversation to update messages/conversation after user navigated away

* chore: params order

* refactor: organize intermediate message logic and ensure correct variables are passed

* fix: Add stt and tts routes before upload limiters, prevent bans

* fix(abortRun): temp fix to delete unfinished messages to avoid message thread parent relationship issues

* refactor: Update AnthropicClient to use node-fetch for fetch key and add proxy support

* fix(gptPlugins): ensure parentMessageId/messageId relationship is maintained

* feat(BaseClient): custom fetch function to analyze/edit payloads just before sending (also prevents abortController crash on Bun runtime)

* feat: `directEndpoint` and `titleMessageRole` custom endpoint options

* chore: Bump version to 0.6.6 in data-provider package.json
2024-05-28 14:52:12 -04:00
..
anthropic 🪙 feat: Configure Max Context and Output Tokens (#2648) 2024-05-09 13:27:13 -04:00
assistants 🤖 feat: OpenAI Assistants v2 (initial support) (#2781) 2024-05-19 12:56:55 -04:00
azureAssistants 🤖 feat: OpenAI Assistants v2 (initial support) (#2781) 2024-05-19 12:56:55 -04:00
custom 🚀 refactor: Enhance Custom Endpoints, Message Logic, and Payload Handling (#2895) 2024-05-28 14:52:12 -04:00
google 🤖 feat: Model Specs & Save Tools per Convo/Preset (#2578) 2024-04-30 22:11:48 -04:00
gptPlugins 🪙 feat: Configure Max Context and Output Tokens (#2648) 2024-05-09 13:27:13 -04:00
openAI 🪙 feat: Configure Max Context and Output Tokens (#2648) 2024-05-09 13:27:13 -04:00