diff --git a/api/core/llm_generator/llm_generator.py b/api/core/llm_generator/llm_generator.py index ec6c537af0..605767bda7 100644 --- a/api/core/llm_generator/llm_generator.py +++ b/api/core/llm_generator/llm_generator.py @@ -589,12 +589,9 @@ class LLMGenerator: "instruction": filled_instruction, } ) - llm_result = cast( - LLMResult, - model_instance.invoke_llm( - prompt_messages=[UserPromptMessage(content=formatted_prompt)], - model_parameters=memory_spec.model.completion_params, - stream=False, - ) + llm_result = model_instance.invoke_llm( + prompt_messages=[UserPromptMessage(content=formatted_prompt)], + model_parameters=memory_spec.model.completion_params, + stream=False, ) return llm_result.message.get_text_content()