From 6216ad77f443793ec08bb80eb287b83ba28f1546 Mon Sep 17 00:00:00 2001 From: Prateek Shourya Date: Fri, 23 May 2025 15:06:58 +0530 Subject: [PATCH] [WEB-4146] fix: AI environment variables configuration in GodMode (#7104) * [WEB-4146] fix: artificial intelligence environment variables configuration * chore: update llm configuration keys --- admin/app/ai/form.tsx | 16 ++++++++-------- apiserver/plane/license/api/views/instance.py | 8 ++++---- packages/types/src/instance/ai.d.ts | 2 +- packages/types/src/instance/base.d.ts | 2 +- .../components/description-editor.tsx | 4 ++-- 5 files changed, 16 insertions(+), 16 deletions(-) diff --git a/admin/app/ai/form.tsx b/admin/app/ai/form.tsx index 4258a99fb..47ab9480e 100644 --- a/admin/app/ai/form.tsx +++ b/admin/app/ai/form.tsx @@ -26,16 +26,16 @@ export const InstanceAIForm: FC = (props) => { formState: { errors, isSubmitting }, } = useForm({ defaultValues: { - OPENAI_API_KEY: config["OPENAI_API_KEY"], - GPT_ENGINE: config["GPT_ENGINE"], + LLM_API_KEY: config["LLM_API_KEY"], + LLM_MODEL: config["LLM_MODEL"], }, }); const aiFormFields: TControllerInputFormField[] = [ { - key: "GPT_ENGINE", + key: "LLM_MODEL", type: "text", - label: "GPT_ENGINE", + label: "LLM Model", description: ( <> Choose an OpenAI engine.{" "} @@ -49,12 +49,12 @@ export const InstanceAIForm: FC = (props) => { ), - placeholder: "gpt-3.5-turbo", - error: Boolean(errors.GPT_ENGINE), + placeholder: "gpt-4o-mini", + error: Boolean(errors.LLM_MODEL), required: false, }, { - key: "OPENAI_API_KEY", + key: "LLM_API_KEY", type: "password", label: "API key", description: ( @@ -71,7 +71,7 @@ export const InstanceAIForm: FC = (props) => { ), placeholder: "sk-asddassdfasdefqsdfasd23das3dasdcasd", - error: Boolean(errors.OPENAI_API_KEY), + error: Boolean(errors.LLM_API_KEY), required: false, }, ]; diff --git a/apiserver/plane/license/api/views/instance.py b/apiserver/plane/license/api/views/instance.py index 0e2b64fc9..c598acfef 100644 --- a/apiserver/plane/license/api/views/instance.py +++ b/apiserver/plane/license/api/views/instance.py @@ -57,7 +57,7 @@ class InstanceEndpoint(BaseAPIView): POSTHOG_API_KEY, POSTHOG_HOST, UNSPLASH_ACCESS_KEY, - OPENAI_API_KEY, + LLM_API_KEY, IS_INTERCOM_ENABLED, INTERCOM_APP_ID, ) = get_configuration_value( @@ -112,8 +112,8 @@ class InstanceEndpoint(BaseAPIView): "default": os.environ.get("UNSPLASH_ACCESS_KEY", ""), }, { - "key": "OPENAI_API_KEY", - "default": os.environ.get("OPENAI_API_KEY", ""), + "key": "LLM_API_KEY", + "default": os.environ.get("LLM_API_KEY", ""), }, # Intercom settings { @@ -151,7 +151,7 @@ class InstanceEndpoint(BaseAPIView): data["has_unsplash_configured"] = bool(UNSPLASH_ACCESS_KEY) # Open AI settings - data["has_openai_configured"] = bool(OPENAI_API_KEY) + data["has_llm_configured"] = bool(LLM_API_KEY) # File size settings data["file_size_limit"] = float(os.environ.get("FILE_SIZE_LIMIT", 5242880)) diff --git a/packages/types/src/instance/ai.d.ts b/packages/types/src/instance/ai.d.ts index 0ac34557a..5bfd1a6ba 100644 --- a/packages/types/src/instance/ai.d.ts +++ b/packages/types/src/instance/ai.d.ts @@ -1 +1 @@ -export type TInstanceAIConfigurationKeys = "OPENAI_API_KEY" | "GPT_ENGINE"; +export type TInstanceAIConfigurationKeys = "LLM_API_KEY" | "LLM_MODEL"; diff --git a/packages/types/src/instance/base.d.ts b/packages/types/src/instance/base.d.ts index dc5ee5fc7..79b1e642f 100644 --- a/packages/types/src/instance/base.d.ts +++ b/packages/types/src/instance/base.d.ts @@ -49,7 +49,7 @@ export interface IInstanceConfig { posthog_api_key: string | undefined; posthog_host: string | undefined; has_unsplash_configured: boolean; - has_openai_configured: boolean; + has_llm_configured: boolean; file_size_limit: number | undefined; is_smtp_configured: boolean; app_base_url: string | undefined; diff --git a/web/core/components/issues/issue-modal/components/description-editor.tsx b/web/core/components/issues/issue-modal/components/description-editor.tsx index b19fc62db..9cae8d840 100644 --- a/web/core/components/issues/issue-modal/components/description-editor.tsx +++ b/web/core/components/issues/issue-modal/components/description-editor.tsx @@ -225,7 +225,7 @@ export const IssueDescriptionEditor: React.FC = ob )} />
- {issueName && issueName.trim() !== "" && config?.has_openai_configured && ( + {issueName && issueName.trim() !== "" && config?.has_llm_configured && ( )} - {config?.has_openai_configured && projectId && ( + {config?.has_llm_configured && projectId && ( {