diff --git a/client/src/components/Endpoints/Plugins/AgentSettings.jsx b/client/src/components/Endpoints/Plugins/AgentSettings.jsx index 2dddc9a476..194cef1f2c 100644 --- a/client/src/components/Endpoints/Plugins/AgentSettings.jsx +++ b/client/src/components/Endpoints/Plugins/AgentSettings.jsx @@ -74,7 +74,7 @@ function Settings(props) { - + @@ -86,7 +86,7 @@ function Settings(props) { - + diff --git a/client/src/components/Endpoints/Plugins/OptionHover.jsx b/client/src/components/Endpoints/Plugins/OptionHover.jsx index 85e04f9863..ddafe023a8 100644 --- a/client/src/components/Endpoints/Plugins/OptionHover.jsx +++ b/client/src/components/Endpoints/Plugins/OptionHover.jsx @@ -2,6 +2,8 @@ import { HoverCardPortal, HoverCardContent } from '~/components'; const types = { temp: 'Higher values = more random, while lower values = more focused and deterministic. We recommend altering this or Top P but not both.', + func: 'Enable use of Plugins as OpenAI Functions', + skip: 'Enable skipping the completion step, which reviews the final answer and generated steps', max: "The max tokens to generate. The total length of input tokens and generated tokens is limited by the model's context length.", topp: 'An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We recommend altering this or temperature but not both.', freq: "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.",