|
|
|
@ -209,6 +209,7 @@
|
|
|
|
"com_endpoint_google_thinking_budget": "Guides the number of thinking tokens the model uses. The actual amount may exceed or fall below this value depending on the prompt.\n\nThis setting is only supported by certain models (2.5 series). Gemini 2.5 Pro supports 128-32,768 tokens. Gemini 2.5 Flash supports 0-24,576 tokens. Gemini 2.5 Flash Lite supports 512-24,576 tokens.\n\nLeave blank or set to \"-1\" to let the model automatically decide when and how much to think. By default, Gemini 2.5 Flash Lite does not think.",
|
|
|
|
"com_endpoint_google_thinking_budget": "Guides the number of thinking tokens the model uses. The actual amount may exceed or fall below this value depending on the prompt.\n\nThis setting is only supported by certain models (2.5 series). Gemini 2.5 Pro supports 128-32,768 tokens. Gemini 2.5 Flash supports 0-24,576 tokens. Gemini 2.5 Flash Lite supports 512-24,576 tokens.\n\nLeave blank or set to \"-1\" to let the model automatically decide when and how much to think. By default, Gemini 2.5 Flash Lite does not think.",
|
|
|
|
"com_endpoint_google_topk": "Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature).",
|
|
|
|
"com_endpoint_google_topk": "Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature).",
|
|
|
|
"com_endpoint_google_topp": "Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value.",
|
|
|
|
"com_endpoint_google_topp": "Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value.",
|
|
|
|
|
|
|
|
"com_endpoint_google_use_search_grounding": "Use Google's search grounding feature to enhance responses with real-time web search results. This enables models to access current information and provide more accurate, up-to-date answers.",
|
|
|
|
"com_endpoint_instructions_assistants": "Override Instructions",
|
|
|
|
"com_endpoint_instructions_assistants": "Override Instructions",
|
|
|
|
"com_endpoint_instructions_assistants_placeholder": "Overrides the instructions of the assistant. This is useful for modifying the behavior on a per-run basis.",
|
|
|
|
"com_endpoint_instructions_assistants_placeholder": "Overrides the instructions of the assistant. This is useful for modifying the behavior on a per-run basis.",
|
|
|
|
"com_endpoint_max_output_tokens": "Max Output Tokens",
|
|
|
|
"com_endpoint_max_output_tokens": "Max Output Tokens",
|
|
|
|
@ -225,7 +226,7 @@
|
|
|
|
"com_endpoint_openai_max_tokens": "Optional 'max_tokens' field, representing the maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the models context length. You may experience errors if this number exceeds the max context tokens.",
|
|
|
|
"com_endpoint_openai_max_tokens": "Optional 'max_tokens' field, representing the maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the models context length. You may experience errors if this number exceeds the max context tokens.",
|
|
|
|
"com_endpoint_openai_pres": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
|
|
|
|
"com_endpoint_openai_pres": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
|
|
|
|
"com_endpoint_openai_prompt_prefix_placeholder": "Set custom instructions to include in System Message. Default: none",
|
|
|
|
"com_endpoint_openai_prompt_prefix_placeholder": "Set custom instructions to include in System Message. Default: none",
|
|
|
|
"com_endpoint_openai_reasoning_effort": "Reasoning models only: constrains effort on reasoning for reasoning models. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.",
|
|
|
|
"com_endpoint_openai_reasoning_effort": "o1 and o3 models only: constrains effort on reasoning for reasoning models. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.",
|
|
|
|
"com_endpoint_openai_reasoning_summary": "Responses API only: A summary of the reasoning performed by the model. This can be useful for debugging and understanding the model's reasoning process. Set to none,auto, concise, or detailed.",
|
|
|
|
"com_endpoint_openai_reasoning_summary": "Responses API only: A summary of the reasoning performed by the model. This can be useful for debugging and understanding the model's reasoning process. Set to none,auto, concise, or detailed.",
|
|
|
|
"com_endpoint_openai_resend": "Resend all previously attached images. Note: this can significantly increase token cost and you may experience errors with many image attachments.",
|
|
|
|
"com_endpoint_openai_resend": "Resend all previously attached images. Note: this can significantly increase token cost and you may experience errors with many image attachments.",
|
|
|
|
"com_endpoint_openai_resend_files": "Resend all previously attached files. Note: this will increase token cost and you may experience errors with many attachments.",
|
|
|
|
"com_endpoint_openai_resend_files": "Resend all previously attached files. Note: this will increase token cost and you may experience errors with many attachments.",
|
|
|
|
@ -233,7 +234,6 @@
|
|
|
|
"com_endpoint_openai_temp": "Higher values = more random, while lower values = more focused and deterministic. We recommend altering this or Top P but not both.",
|
|
|
|
"com_endpoint_openai_temp": "Higher values = more random, while lower values = more focused and deterministic. We recommend altering this or Top P but not both.",
|
|
|
|
"com_endpoint_openai_topp": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We recommend altering this or temperature but not both.",
|
|
|
|
"com_endpoint_openai_topp": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We recommend altering this or temperature but not both.",
|
|
|
|
"com_endpoint_openai_use_responses_api": "Use the Responses API instead of Chat Completions, which includes extended features from OpenAI. Required for o1-pro, o3-pro, and to enable reasoning summaries.",
|
|
|
|
"com_endpoint_openai_use_responses_api": "Use the Responses API instead of Chat Completions, which includes extended features from OpenAI. Required for o1-pro, o3-pro, and to enable reasoning summaries.",
|
|
|
|
"com_endpoint_google_use_search_grounding": "Use Google's search grounding feature to enhance responses with real-time web search results. This enables models to access current information and provide more accurate, up-to-date answers.",
|
|
|
|
|
|
|
|
"com_endpoint_output": "Output",
|
|
|
|
"com_endpoint_output": "Output",
|
|
|
|
"com_endpoint_plug_image_detail": "Image Detail",
|
|
|
|
"com_endpoint_plug_image_detail": "Image Detail",
|
|
|
|
"com_endpoint_plug_resend_files": "Resend Files",
|
|
|
|
"com_endpoint_plug_resend_files": "Resend Files",
|
|
|
|
@ -290,10 +290,10 @@
|
|
|
|
"com_error_files_upload": "An error occurred while uploading the file.",
|
|
|
|
"com_error_files_upload": "An error occurred while uploading the file.",
|
|
|
|
"com_error_files_upload_canceled": "The file upload request was canceled. Note: the file upload may still be processing and will need to be manually deleted.",
|
|
|
|
"com_error_files_upload_canceled": "The file upload request was canceled. Note: the file upload may still be processing and will need to be manually deleted.",
|
|
|
|
"com_error_files_validation": "An error occurred while validating the file.",
|
|
|
|
"com_error_files_validation": "An error occurred while validating the file.",
|
|
|
|
|
|
|
|
"com_error_google_tool_conflict": "Usage of built-in Google tools are not supported with external tools. Please disable either the built-in tools or the external tools.",
|
|
|
|
"com_error_heic_conversion": "Failed to convert HEIC image to JPEG. Please try converting the image manually or use a different format.",
|
|
|
|
"com_error_heic_conversion": "Failed to convert HEIC image to JPEG. Please try converting the image manually or use a different format.",
|
|
|
|
"com_error_input_length": "The latest message token count is too long, exceeding the token limit, or your token limit parameters are misconfigured, adversely affecting the context window. More info: {{0}}. Please shorten your message, adjust the max context size from the conversation parameters, or fork the conversation to continue.",
|
|
|
|
"com_error_input_length": "The latest message token count is too long, exceeding the token limit, or your token limit parameters are misconfigured, adversely affecting the context window. More info: {{0}}. Please shorten your message, adjust the max context size from the conversation parameters, or fork the conversation to continue.",
|
|
|
|
"com_error_invalid_agent_provider": "The \"{{0}}\" provider is not available for use with Agents. Please go to your agent's settings and select a currently available provider.",
|
|
|
|
"com_error_invalid_agent_provider": "The \"{{0}}\" provider is not available for use with Agents. Please go to your agent's settings and select a currently available provider.",
|
|
|
|
"com_error_google_tool_conflict": "Usage of built-in Google tools are not supported with external tools. Please disable either the built-in tools or the external tools.",
|
|
|
|
|
|
|
|
"com_error_invalid_user_key": "Invalid key provided. Please provide a valid key and try again.",
|
|
|
|
"com_error_invalid_user_key": "Invalid key provided. Please provide a valid key and try again.",
|
|
|
|
"com_error_moderation": "It appears that the content submitted has been flagged by our moderation system for not aligning with our community guidelines. We're unable to proceed with this specific topic. If you have any other questions or topics you'd like to explore, please edit your message, or create a new conversation.",
|
|
|
|
"com_error_moderation": "It appears that the content submitted has been flagged by our moderation system for not aligning with our community guidelines. We're unable to proceed with this specific topic. If you have any other questions or topics you'd like to explore, please edit your message, or create a new conversation.",
|
|
|
|
"com_error_no_base_url": "No base URL found. Please provide one and try again.",
|
|
|
|
"com_error_no_base_url": "No base URL found. Please provide one and try again.",
|
|
|
|
@ -641,6 +641,7 @@
|
|
|
|
"com_ui_command_placeholder": "Optional: Enter a command for the prompt or name will be used",
|
|
|
|
"com_ui_command_placeholder": "Optional: Enter a command for the prompt or name will be used",
|
|
|
|
"com_ui_command_usage_placeholder": "Select a Prompt by command or name",
|
|
|
|
"com_ui_command_usage_placeholder": "Select a Prompt by command or name",
|
|
|
|
"com_ui_complete_setup": "Complete Setup",
|
|
|
|
"com_ui_complete_setup": "Complete Setup",
|
|
|
|
|
|
|
|
"com_ui_concise": "Concise",
|
|
|
|
"com_ui_configure_mcp_variables_for": "Configure Variables for {{0}}",
|
|
|
|
"com_ui_configure_mcp_variables_for": "Configure Variables for {{0}}",
|
|
|
|
"com_ui_confirm_action": "Confirm Action",
|
|
|
|
"com_ui_confirm_action": "Confirm Action",
|
|
|
|
"com_ui_confirm_admin_use_change": "Changing this setting will block access for admins, including yourself. Are you sure you want to proceed?",
|
|
|
|
"com_ui_confirm_admin_use_change": "Changing this setting will block access for admins, including yourself. Are you sure you want to proceed?",
|
|
|
|
@ -706,6 +707,7 @@
|
|
|
|
"com_ui_description": "Description",
|
|
|
|
"com_ui_description": "Description",
|
|
|
|
"com_ui_description_placeholder": "Optional: Enter a description to display for the prompt",
|
|
|
|
"com_ui_description_placeholder": "Optional: Enter a description to display for the prompt",
|
|
|
|
"com_ui_deselect_all": "Deselect All",
|
|
|
|
"com_ui_deselect_all": "Deselect All",
|
|
|
|
|
|
|
|
"com_ui_detailed": "Detailed",
|
|
|
|
"com_ui_disabling": "Disabling...",
|
|
|
|
"com_ui_disabling": "Disabling...",
|
|
|
|
"com_ui_download": "Download",
|
|
|
|
"com_ui_download": "Download",
|
|
|
|
"com_ui_download_artifact": "Download Artifact",
|
|
|
|
"com_ui_download_artifact": "Download Artifact",
|
|
|
|
@ -800,6 +802,7 @@
|
|
|
|
"com_ui_happy_birthday": "It's my 1st birthday!",
|
|
|
|
"com_ui_happy_birthday": "It's my 1st birthday!",
|
|
|
|
"com_ui_hide_image_details": "Hide Image Details",
|
|
|
|
"com_ui_hide_image_details": "Hide Image Details",
|
|
|
|
"com_ui_hide_qr": "Hide QR Code",
|
|
|
|
"com_ui_hide_qr": "Hide QR Code",
|
|
|
|
|
|
|
|
"com_ui_high": "High",
|
|
|
|
"com_ui_host": "Host",
|
|
|
|
"com_ui_host": "Host",
|
|
|
|
"com_ui_icon": "Icon",
|
|
|
|
"com_ui_icon": "Icon",
|
|
|
|
"com_ui_idea": "Ideas",
|
|
|
|
"com_ui_idea": "Ideas",
|
|
|
|
@ -828,10 +831,6 @@
|
|
|
|
"com_ui_locked": "Locked",
|
|
|
|
"com_ui_locked": "Locked",
|
|
|
|
"com_ui_logo": "{{0}} Logo",
|
|
|
|
"com_ui_logo": "{{0}} Logo",
|
|
|
|
"com_ui_low": "Low",
|
|
|
|
"com_ui_low": "Low",
|
|
|
|
"com_ui_concise": "Concise",
|
|
|
|
|
|
|
|
"com_ui_detailed": "Detailed",
|
|
|
|
|
|
|
|
"com_ui_high": "High",
|
|
|
|
|
|
|
|
"com_ui_medium": "Medium",
|
|
|
|
|
|
|
|
"com_ui_manage": "Manage",
|
|
|
|
"com_ui_manage": "Manage",
|
|
|
|
"com_ui_max_tags": "Maximum number allowed is {{0}}, using latest values.",
|
|
|
|
"com_ui_max_tags": "Maximum number allowed is {{0}}, using latest values.",
|
|
|
|
"com_ui_mcp_dialog_desc": "Please enter the necessary information below.",
|
|
|
|
"com_ui_mcp_dialog_desc": "Please enter the necessary information below.",
|
|
|
|
@ -839,6 +838,7 @@
|
|
|
|
"com_ui_mcp_server_not_found": "Server not found.",
|
|
|
|
"com_ui_mcp_server_not_found": "Server not found.",
|
|
|
|
"com_ui_mcp_servers": "MCP Servers",
|
|
|
|
"com_ui_mcp_servers": "MCP Servers",
|
|
|
|
"com_ui_mcp_url": "MCP Server URL",
|
|
|
|
"com_ui_mcp_url": "MCP Server URL",
|
|
|
|
|
|
|
|
"com_ui_medium": "Medium",
|
|
|
|
"com_ui_memories": "Memories",
|
|
|
|
"com_ui_memories": "Memories",
|
|
|
|
"com_ui_memories_allow_create": "Allow creating Memories",
|
|
|
|
"com_ui_memories_allow_create": "Allow creating Memories",
|
|
|
|
"com_ui_memories_allow_opt_out": "Allow users to opt out of Memories",
|
|
|
|
"com_ui_memories_allow_opt_out": "Allow users to opt out of Memories",
|
|
|
|
|