From 4f4473e3d08bcf2ddaad81664b94fe23c99432a2 Mon Sep 17 00:00:00 2001 From: stream Date: Mon, 21 Jul 2025 18:36:20 +0800 Subject: [PATCH] chore: run ruff --- api/core/llm_generator/prompts.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/api/core/llm_generator/prompts.py b/api/core/llm_generator/prompts.py index ace8ac73a6..f2fb4a9c41 100644 --- a/api/core/llm_generator/prompts.py +++ b/api/core/llm_generator/prompts.py @@ -412,8 +412,8 @@ Your output must strictly follow the schema format, do not output any content ou INSTRUCTION_GENERATE_TEMPLATE_PROMPT = """ The output of this prompt is not as expected. You should edit the prompt according to the IDEAL OUTPUT. -""" # noqa: E501 +""" INSTRUCTION_GENERATE_TEMPLATE_CODE = """ Please fix the errors in the error message. -""" # noqa: E501 +"""