From 06e571e6d2f4269f334555633fcfa4bc3d854ddb Mon Sep 17 00:00:00 2001 From: InspectorCaracal <51038201+InspectorCaracal@users.noreply.github.com> Date: Thu, 29 Aug 2024 11:41:25 -0600 Subject: [PATCH] fix llm-client headers --- evennia/contrib/rpg/llm/llm_client.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/evennia/contrib/rpg/llm/llm_client.py b/evennia/contrib/rpg/llm/llm_client.py index 51fa4ea2c1..91641de4c0 100644 --- a/evennia/contrib/rpg/llm/llm_client.py +++ b/evennia/contrib/rpg/llm/llm_client.py @@ -37,7 +37,7 @@ from evennia.utils.utils import make_iter DEFAULT_LLM_HOST = "http://127.0.0.1:5000" DEFAULT_LLM_PATH = "/api/v1/generate" -DEFAULT_LLM_HEADERS = {"Content-Type": "application/json"} +DEFAULT_LLM_HEADERS = {"Content-Type": ["application/json"]} DEFAULT_LLM_PROMPT_KEYNAME = "prompt" DEFAULT_LLM_API_TYPE = "" # or openai DEFAULT_LLM_REQUEST_BODY = {