diff --git a/api/app/langchain/agents/Functions/initializeFunctionsAgent.js b/api/app/langchain/agents/Functions/initializeFunctionsAgent.js index da813e7ef1..dfc33c9262 100644 --- a/api/app/langchain/agents/Functions/initializeFunctionsAgent.js +++ b/api/app/langchain/agents/Functions/initializeFunctionsAgent.js @@ -1,5 +1,3 @@ -// const FunctionsAgent = require('./FunctionsAgent'); -// const { AgentExecutor, initializeAgentExecutorWithOptions } = require('langchain/agents'); const { initializeAgentExecutorWithOptions } = require('langchain/agents'); const { BufferMemory, ChatMessageHistory } = require('langchain/memory'); @@ -10,14 +8,7 @@ const initializeFunctionsAgent = async ({ // currentDateString, ...rest }) => { - // const agent = FunctionsAgent.fromLLMAndTools( - // model, - // tools, - // { - // currentDateString, - // }); - - + const memory = new BufferMemory({ chatHistory: new ChatMessageHistory(pastMessages), memoryKey: 'chat_history', @@ -28,15 +19,12 @@ const initializeFunctionsAgent = async ({ returnMessages: true, }); - // return AgentExecutor.fromAgentAndTools({ agent, tools, memory, ...rest }); - return await initializeAgentExecutorWithOptions( tools, model, { agentType: "openai-functions", memory, - maxIterations: 4, ...rest, } ); diff --git a/api/server/routes/ask/askGPTPlugins.js b/api/server/routes/ask/askGPTPlugins.js index a904fd5bc2..10465e71ce 100644 --- a/api/server/routes/ask/askGPTPlugins.js +++ b/api/server/routes/ask/askGPTPlugins.js @@ -39,8 +39,8 @@ router.post('/', requireJwtAuth, async (req, res) => { if (endpoint !== 'gptPlugins') return handleError(res, { text: 'Illegal request' }); const agentOptions = req.body?.agentOptions ?? { - agent: 'classic', - skipCompletion: false, + agent: 'functions', + skipCompletion: true, model: 'gpt-3.5-turbo', temperature: 0, // top_p: 1, diff --git a/client/src/utils/cleanupPreset.js b/client/src/utils/cleanupPreset.js index d9011aead8..ff01b5b8b2 100644 --- a/client/src/utils/cleanupPreset.js +++ b/client/src/utils/cleanupPreset.js @@ -51,8 +51,8 @@ const cleanupPreset = ({ preset: _preset, endpointsConfig = {} }) => { }; } else if (endpoint === 'gptPlugins') { const agentOptions = _preset?.agentOptions ?? { - agent: 'classic', - skipCompletion: false, + agent: 'functions', + skipCompletion: true, model: 'gpt-3.5-turbo', temperature: 0, // top_p: 1, diff --git a/client/src/utils/getDefaultConversation.js b/client/src/utils/getDefaultConversation.js index 91344fb01f..90d9b30363 100644 --- a/client/src/utils/getDefaultConversation.js +++ b/client/src/utils/getDefaultConversation.js @@ -67,8 +67,8 @@ const buildDefaultConversation = ({ }; } else if (endpoint === 'gptPlugins') { const agentOptions = lastConversationSetup?.agentOptions ?? { - agent: 'classic', - skipCompletion: false, + agent: 'functions', + skipCompletion: true, model: 'gpt-3.5-turbo', temperature: 0, // top_p: 1, diff --git a/client/src/utils/handleSubmit.js b/client/src/utils/handleSubmit.js index a8319136bd..76c5d96f96 100644 --- a/client/src/utils/handleSubmit.js +++ b/client/src/utils/handleSubmit.js @@ -88,8 +88,8 @@ const useMessageHandler = () => { responseSender = 'ChatGPT'; } else if (endpoint === 'gptPlugins') { const agentOptions = currentConversation?.agentOptions ?? { - agent: 'classic', - skipCompletion: false, + agent: 'functions', + skipCompletion: true, model: 'gpt-3.5-turbo', temperature: 0, // top_p: 1,