diff --git a/api/core/llm_generator/llm_generator.py b/api/core/llm_generator/llm_generator.py index 8f9b6aa19f..19c8c29ef0 100644 --- a/api/core/llm_generator/llm_generator.py +++ b/api/core/llm_generator/llm_generator.py @@ -476,11 +476,8 @@ class LLMGenerator: ) return { - "variables": response.variables, + **response.model_dump(), "code_language": language, - "code": response.code, - "outputs": response.outputs, - "message": response.explanation, "error": "", } @@ -529,7 +526,7 @@ class LLMGenerator: ) prompt_messages: list[PromptMessage] = [ - SystemPromptMessage(content=system_prompt), + UserPromptMessage(content=system_prompt), ] # Get model instance - use default if model_config not provided diff --git a/api/core/llm_generator/output_models.py b/api/core/llm_generator/output_models.py index 56be4a6ed2..61430867e8 100644 --- a/api/core/llm_generator/output_models.py +++ b/api/core/llm_generator/output_models.py @@ -55,7 +55,7 @@ class CodeNodeStructuredOutput(BaseModel): outputs: list[CodeNodeOutputItem] = Field( description="Output variable definitions specifying name and type for each return value" ) - explanation: str = Field(description="Brief explanation of what the generated code does") + message: str = Field(description="Brief explanation of what the generated code does") class InstructionModifyOutput(BaseModel):