From db870e55c3033f3a92411ec6e9e7c23a2febe896 Mon Sep 17 00:00:00 2001 From: Fuegovic <32828263+fuegovic@users.noreply.github.com> Date: Sat, 9 Mar 2024 08:32:08 -0500 Subject: [PATCH] =?UTF-8?q?=F0=9F=94=96=20chore:=20update=20groq=20models?= =?UTF-8?q?=20(#2031)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docs/install/configuration/ai_endpoints.md | 3 ++- docs/install/configuration/custom_config.md | 5 +++++ librechat.example.yaml | 23 +++++++++++++++++---- 3 files changed, 26 insertions(+), 5 deletions(-) diff --git a/docs/install/configuration/ai_endpoints.md b/docs/install/configuration/ai_endpoints.md index 59ba1bff6a..7990bf616d 100644 --- a/docs/install/configuration/ai_endpoints.md +++ b/docs/install/configuration/ai_endpoints.md @@ -32,7 +32,8 @@ Some of the endpoints are marked as **Known,** which means they might have speci models: default: [ "llama2-70b-4096", - "mixtral-8x7b-32768" + "mixtral-8x7b-32768", + "gemma-7b-it" ] fetch: false titleConvo: true diff --git a/docs/install/configuration/custom_config.md b/docs/install/configuration/custom_config.md index dab98a6893..df8f2177b4 100644 --- a/docs/install/configuration/custom_config.md +++ b/docs/install/configuration/custom_config.md @@ -563,6 +563,11 @@ endpoints: - "Mistral" - "OpenRouter" - "Groq" + - "Anyscale" + - "Fireworks" + - "Perplexity" + - "together.ai" + - "Ollama" ### **models**: diff --git a/librechat.example.yaml b/librechat.example.yaml index 48b00ac772..f0bf6088b0 100644 --- a/librechat.example.yaml +++ b/librechat.example.yaml @@ -2,7 +2,7 @@ # https://docs.librechat.ai/install/configuration/custom_config.html # Configuration version (required) -version: 1.0.3 +version: 1.0.4 # Cache settings: Set to true to enable caching cache: true @@ -48,7 +48,22 @@ endpoints: # supportedIds: ["asst_supportedAssistantId1", "asst_supportedAssistantId2"] # # excludedIds: ["asst_excludedAssistantId"] custom: - # Mistral AI API + # Groq Example + - name: "groq" + apiKey: "${GROQ_API_KEY}" + baseURL: "https://api.groq.com/openai/v1/" + models: + default: [ + "llama2-70b-4096", + "mixtral-8x7b-32768", + "gemma-7b-it" + ] + fetch: false + titleConvo: true + titleModel: "mixtral-8x7b-32768" + modelDisplayLabel: "groq" + + # Mistral AI Example - name: "Mistral" # Unique name for the endpoint # For `apiKey` and `baseURL`, you can use environment variables that you define. # recommended environment variables: @@ -92,8 +107,8 @@ endpoints: # Drop Default params parameters from the request. See default params in guide linked below. # NOTE: For Mistral, it is necessary to drop the following parameters or you will encounter a 422 Error: dropParams: ["stop", "user", "frequency_penalty", "presence_penalty"] - - # OpenRouter.ai Example + + # OpenRouter Example - name: "OpenRouter" # For `apiKey` and `baseURL`, you can use environment variables that you define. # recommended environment variables: