LibreChat/api/server/services/Endpoints/gptPlugins
Danny Avila 5d40d0a37a
⚙️ feat: Adjust Rate of Stream Progress (#3244)
* chore: bump data-provider and add MESSAGES CacheKey

* refactor: avoid saving messages while streaming, save partial text to cache instead

* fix(ci): processChunks

* chore: logging aborted request to debug

* feat: set stream rate for token processing

* chore: specify default stream rate

* fix(ci): Update AppService.js to use optional chaining for endpointLocals assignment

* refactor: abstract the error handler

* feat: streamRate for assistants; refactor: update default rate for token

* refactor: update error handling in assistants/errors.js

* refactor: update error handling in assistants/errors.js
2024-07-17 10:47:17 -04:00
..
buildOptions.js 🪙 feat: Configure Max Context and Output Tokens (#2648) 2024-05-09 13:27:13 -04:00
index.js feat(Google): Support all Text/Chat Models, Response streaming, PaLM -> Google 🤖 (#1316) 2023-12-10 14:54:13 -05:00
initializeClient.js ⚙️ feat: Adjust Rate of Stream Progress (#3244) 2024-07-17 10:47:17 -04:00
initializeClient.spec.js 🧑‍💻 refactor: Display Client-facing Errors (#2476) 2024-04-21 08:31:54 -04:00