diff --git a/client/src/components/Endpoints/Plugins/AgentSettings.jsx b/client/src/components/Endpoints/Plugins/AgentSettings.jsx
index 2dddc9a476..194cef1f2c 100644
--- a/client/src/components/Endpoints/Plugins/AgentSettings.jsx
+++ b/client/src/components/Endpoints/Plugins/AgentSettings.jsx
@@ -74,7 +74,7 @@ function Settings(props) {
-
+
@@ -86,7 +86,7 @@ function Settings(props) {
-
+
diff --git a/client/src/components/Endpoints/Plugins/OptionHover.jsx b/client/src/components/Endpoints/Plugins/OptionHover.jsx
index 85e04f9863..ddafe023a8 100644
--- a/client/src/components/Endpoints/Plugins/OptionHover.jsx
+++ b/client/src/components/Endpoints/Plugins/OptionHover.jsx
@@ -2,6 +2,8 @@ import { HoverCardPortal, HoverCardContent } from '~/components';
const types = {
temp: 'Higher values = more random, while lower values = more focused and deterministic. We recommend altering this or Top P but not both.',
+ func: 'Enable use of Plugins as OpenAI Functions',
+ skip: 'Enable skipping the completion step, which reviews the final answer and generated steps',
max: "The max tokens to generate. The total length of input tokens and generated tokens is limited by the model's context length.",
topp: 'An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We recommend altering this or temperature but not both.',
freq: "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.",