From 9dee0a5fc9992df984125ea6046e3e713667d36b Mon Sep 17 00:00:00 2001 From: John Wang Date: Tue, 23 May 2023 19:48:04 +0800 Subject: [PATCH] fix: stop words not empty --- api/core/completion.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/api/core/completion.py b/api/core/completion.py index c5e707d1de..5e559ac7c7 100644 --- a/api/core/completion.py +++ b/api/core/completion.py @@ -123,7 +123,7 @@ class Completion: def get_main_llm_prompt(cls, mode: str, llm: BaseLanguageModel, pre_prompt: str, query: str, inputs: dict, chain_output: Optional[str], memory: Optional[ReadOnlyConversationTokenDBBufferSharedMemory]) -> \ - Tuple[Union[str | List[BaseMessage]], List[str]]: + Tuple[Union[str | List[BaseMessage]], Optional[List[str]]]: # disable template string in query query_params = OutLinePromptTemplate.from_template(template=query).input_variables if query_params: @@ -165,9 +165,9 @@ And answer according to the language of the user's question. if isinstance(llm, BaseChatModel): # use chat llm as completion model - return [HumanMessage(content=prompt_content)], list() + return [HumanMessage(content=prompt_content)], None else: - return prompt_content, list() + return prompt_content, None else: messages: List[BaseMessage] = []