mirror of
https://github.com/danny-avila/LibreChat.git
synced 2026-03-09 01:22:36 +01:00
🌍 i18n: Update translation.json with latest translations (#6009)
Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
This commit is contained in:
parent
be280004cf
commit
34f967eff8
2 changed files with 7 additions and 6 deletions
|
|
@ -140,11 +140,9 @@
|
||||||
"com_endpoint_ai": "AI",
|
"com_endpoint_ai": "AI",
|
||||||
"com_endpoint_anthropic_maxoutputtokens": "Maximum number of tokens that can be generated in the response. Specify a lower value for shorter responses and a higher value for longer responses. Note: models may stop before reaching this maximum.",
|
"com_endpoint_anthropic_maxoutputtokens": "Maximum number of tokens that can be generated in the response. Specify a lower value for shorter responses and a higher value for longer responses. Note: models may stop before reaching this maximum.",
|
||||||
"com_endpoint_anthropic_prompt_cache": "Prompt caching allows reusing large context or instructions across API calls, reducing costs and latency",
|
"com_endpoint_anthropic_prompt_cache": "Prompt caching allows reusing large context or instructions across API calls, reducing costs and latency",
|
||||||
"com_endpoint_thinking": "Thinking",
|
|
||||||
"com_endpoint_thinking_budget": "Thinking Budget",
|
|
||||||
"com_endpoint_anthropic_thinking": "Enables internal reasoning for supported Claude models (3.7 Sonnet). Note: requires \"Thinking Budget\" to be set and lower than \"Max Output Tokens\"",
|
|
||||||
"com_endpoint_anthropic_thinking_budget": "Determines the max number of tokens Claude is allowed to use for its internal reasoning process. Larger budgets can improve response quality by enabling more thorough analysis for complex problems, although Claude may not use the entire budget allocated, especially at ranges above 32K. This setting must be lower than \"Max Output Tokens.\"",
|
|
||||||
"com_endpoint_anthropic_temp": "Ranges from 0 to 1. Use temp closer to 0 for analytical / multiple choice, and closer to 1 for creative and generative tasks. We recommend altering this or Top P but not both.",
|
"com_endpoint_anthropic_temp": "Ranges from 0 to 1. Use temp closer to 0 for analytical / multiple choice, and closer to 1 for creative and generative tasks. We recommend altering this or Top P but not both.",
|
||||||
|
"com_endpoint_anthropic_thinking": "Enables internal reasoning for supported Claude models (3.7 Sonnet). Note: requires \"Thinking Budget\" to be set and lower than \"Max Output Tokens\"",
|
||||||
|
"com_endpoint_anthropic_thinking_budget": "Determines the max number of tokens Claude is allowed use for its internal reasoning process. Larger budgets can improve response quality by enabling more thorough analysis for complex problems, although Claude may not use the entire budget allocated, especially at ranges above 32K. This setting must be lower than \"Max Output Tokens.\"",
|
||||||
"com_endpoint_anthropic_topk": "Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature).",
|
"com_endpoint_anthropic_topk": "Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature).",
|
||||||
"com_endpoint_anthropic_topp": "Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value.",
|
"com_endpoint_anthropic_topp": "Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value.",
|
||||||
"com_endpoint_assistant": "Assistant",
|
"com_endpoint_assistant": "Assistant",
|
||||||
|
|
@ -250,6 +248,8 @@
|
||||||
"com_endpoint_stop": "Stop Sequences",
|
"com_endpoint_stop": "Stop Sequences",
|
||||||
"com_endpoint_stop_placeholder": "Separate values by pressing `Enter`",
|
"com_endpoint_stop_placeholder": "Separate values by pressing `Enter`",
|
||||||
"com_endpoint_temperature": "Temperature",
|
"com_endpoint_temperature": "Temperature",
|
||||||
|
"com_endpoint_thinking": "Thinking",
|
||||||
|
"com_endpoint_thinking_budget": "Thinking Budget",
|
||||||
"com_endpoint_top_k": "Top K",
|
"com_endpoint_top_k": "Top K",
|
||||||
"com_endpoint_top_p": "Top P",
|
"com_endpoint_top_p": "Top P",
|
||||||
"com_endpoint_use_active_assistant": "Use Active Assistant",
|
"com_endpoint_use_active_assistant": "Use Active Assistant",
|
||||||
|
|
@ -357,12 +357,12 @@
|
||||||
"com_nav_lang_estonian": "Eesti keel",
|
"com_nav_lang_estonian": "Eesti keel",
|
||||||
"com_nav_lang_finnish": "Suomi",
|
"com_nav_lang_finnish": "Suomi",
|
||||||
"com_nav_lang_french": "Français ",
|
"com_nav_lang_french": "Français ",
|
||||||
|
"com_nav_lang_georgian": "ქართული",
|
||||||
"com_nav_lang_german": "Deutsch",
|
"com_nav_lang_german": "Deutsch",
|
||||||
"com_nav_lang_hebrew": "עברית",
|
"com_nav_lang_hebrew": "עברית",
|
||||||
"com_nav_lang_indonesia": "Indonesia",
|
"com_nav_lang_indonesia": "Indonesia",
|
||||||
"com_nav_lang_italian": "Italiano",
|
"com_nav_lang_italian": "Italiano",
|
||||||
"com_nav_lang_japanese": "日本語",
|
"com_nav_lang_japanese": "日本語",
|
||||||
"com_nav_lang_georgian": "ქართული",
|
|
||||||
"com_nav_lang_korean": "한국어",
|
"com_nav_lang_korean": "한국어",
|
||||||
"com_nav_lang_polish": "Polski",
|
"com_nav_lang_polish": "Polski",
|
||||||
"com_nav_lang_portuguese": "Português",
|
"com_nav_lang_portuguese": "Português",
|
||||||
|
|
@ -834,4 +834,4 @@
|
||||||
"com_ui_zoom": "Zoom",
|
"com_ui_zoom": "Zoom",
|
||||||
"com_user_message": "You",
|
"com_user_message": "You",
|
||||||
"com_warning_resubmit_unsupported": "Resubmitting the AI message is not supported for this endpoint."
|
"com_warning_resubmit_unsupported": "Resubmitting the AI message is not supported for this endpoint."
|
||||||
}
|
}
|
||||||
|
|
@ -720,6 +720,7 @@
|
||||||
"com_ui_stop": "停止",
|
"com_ui_stop": "停止",
|
||||||
"com_ui_storage": "存储",
|
"com_ui_storage": "存储",
|
||||||
"com_ui_submit": "提交",
|
"com_ui_submit": "提交",
|
||||||
|
"com_ui_teach_or_explain": "学习中",
|
||||||
"com_ui_temporary_chat": "临时对话",
|
"com_ui_temporary_chat": "临时对话",
|
||||||
"com_ui_terms_and_conditions": "条款和条件",
|
"com_ui_terms_and_conditions": "条款和条件",
|
||||||
"com_ui_terms_of_service": "服务政策",
|
"com_ui_terms_of_service": "服务政策",
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue