From 9778e73087b59492a45eaf7f5e828c948c29e6be Mon Sep 17 00:00:00 2001 From: Daniel Avila Date: Thu, 15 Jun 2023 23:52:40 -0400 Subject: [PATCH] style(OptionHover.jsx): add descriptions for new types feat(AgentSettings.jsx): change OptionHover type from 'temp' to 'func' and from 'temp' to 'skip' --- client/src/components/Endpoints/Plugins/AgentSettings.jsx | 4 ++-- client/src/components/Endpoints/Plugins/OptionHover.jsx | 2 ++ 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/client/src/components/Endpoints/Plugins/AgentSettings.jsx b/client/src/components/Endpoints/Plugins/AgentSettings.jsx index 2dddc9a476..194cef1f2c 100644 --- a/client/src/components/Endpoints/Plugins/AgentSettings.jsx +++ b/client/src/components/Endpoints/Plugins/AgentSettings.jsx @@ -74,7 +74,7 @@ function Settings(props) { - + @@ -86,7 +86,7 @@ function Settings(props) { - + diff --git a/client/src/components/Endpoints/Plugins/OptionHover.jsx b/client/src/components/Endpoints/Plugins/OptionHover.jsx index 85e04f9863..ddafe023a8 100644 --- a/client/src/components/Endpoints/Plugins/OptionHover.jsx +++ b/client/src/components/Endpoints/Plugins/OptionHover.jsx @@ -2,6 +2,8 @@ import { HoverCardPortal, HoverCardContent } from '~/components'; const types = { temp: 'Higher values = more random, while lower values = more focused and deterministic. We recommend altering this or Top P but not both.', + func: 'Enable use of Plugins as OpenAI Functions', + skip: 'Enable skipping the completion step, which reviews the final answer and generated steps', max: "The max tokens to generate. The total length of input tokens and generated tokens is limited by the model's context length.", topp: 'An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We recommend altering this or temperature but not both.', freq: "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.",