LibreChat/api/app/clients/specs
Danny Avila a45b384bbc
💾 feat: Anthropic Prompt Caching (#3670)
* wip: initial cache control implementation, add typing for transactions handling

* feat: first pass of Anthropic Prompt Caching

* feat: standardize stream usage as pass in when calculating token counts

* feat: Add getCacheMultiplier function to calculate cache multiplier for different valueKeys and cacheTypes

* chore: imports order

* refactor: token usage recording in AnthropicClient, no need to "correct" as we have the correct amount

* feat: more accurate token counting using stream usage data

* feat: Improve token counting accuracy with stream usage data

* refactor: ensure more accurate than not token estimations if custom instructions or files are not being resent with every request

* refactor: cleanup updateUserMessageTokenCount to allow transactions to be as accurate as possible even if we shouldn't update user message token counts

* ci: fix tests
2024-08-17 03:24:09 -04:00
..
AnthropicClient.test.js 💾 feat: Anthropic Prompt Caching (#3670) 2024-08-17 03:24:09 -04:00
BaseClient.test.js fix: Wait for Initial Message Save & Correct Latest Message (#3399) 2024-07-20 01:51:59 -04:00
FakeClient.js 🪙 feat: Configure Max Context and Output Tokens (#2648) 2024-05-09 13:27:13 -04:00
OpenAIClient.test.js 🤖 feat: OpenAI Assistants v2 (initial support) (#2781) 2024-05-19 12:56:55 -04:00
OpenAIClient.tokens.js 🛡️ feat: Optimize and Improve Anonymity of SharedLinks (#3543) 2024-08-05 03:34:00 -04:00
PluginsClient.test.js 🛠️ fix: Tool Filtering in PluginsClient (#3266) 2024-07-04 10:34:28 -04:00