mirror of https://github.com/langgenius/dify.git
fix: change model mode. (#1520)
This commit is contained in:
parent
a80d8286c2
commit
8835435558
|
|
@ -277,7 +277,6 @@ class Completion:
|
|||
if app_model_config.prompt_type == 'simple':
|
||||
prompt_messages, stop_words = prompt_transform.get_prompt(
|
||||
app_mode=mode,
|
||||
app_model_config=app_model_config,
|
||||
pre_prompt=app_model_config.pre_prompt,
|
||||
inputs=inputs,
|
||||
query=query,
|
||||
|
|
@ -366,7 +365,6 @@ class Completion:
|
|||
if app_model_config.prompt_type == 'simple':
|
||||
prompt_messages, _ = prompt_transform.get_prompt(
|
||||
app_mode=mode,
|
||||
app_model_config=app_model_config,
|
||||
pre_prompt=app_model_config.pre_prompt,
|
||||
inputs=inputs,
|
||||
query=query,
|
||||
|
|
|
|||
|
|
@ -27,7 +27,6 @@ class AppMode(enum.Enum):
|
|||
class PromptTransform:
|
||||
def get_prompt(self,
|
||||
app_mode: str,
|
||||
app_model_config: AppModelConfig,
|
||||
pre_prompt: str,
|
||||
inputs: dict,
|
||||
query: str,
|
||||
|
|
@ -36,10 +35,9 @@ class PromptTransform:
|
|||
memory: Optional[BaseChatMemory],
|
||||
model_instance: BaseLLM) -> \
|
||||
Tuple[List[PromptMessage], Optional[List[str]]]:
|
||||
model_mode = app_model_config.model_dict['mode']
|
||||
|
||||
app_mode_enum = AppMode(app_mode)
|
||||
model_mode_enum = ModelMode(model_mode)
|
||||
model_mode_enum = model_instance.model_mode
|
||||
|
||||
prompt_rules = self._read_prompt_rules_from_file(self._prompt_file_name(app_mode, model_instance))
|
||||
|
||||
|
|
|
|||
Loading…
Reference in New Issue