From b3dafd913b352cc59257f1c95a7a2a29b0f51411 Mon Sep 17 00:00:00 2001 From: Yongtao Huang Date: Fri, 19 Sep 2025 10:36:16 +0800 Subject: [PATCH] Chore: correct inconsistent logging and typo (#25945) --- ...qlalchemy_workflow_node_execution_repository.py | 4 +--- .../graph_engine/layers/execution_limits.py | 2 +- api/services/rag_pipeline/rag_pipeline.py | 14 +++++++------- .../rag_pipeline/rag_pipeline_transform_service.py | 6 +++--- 4 files changed, 12 insertions(+), 14 deletions(-) diff --git a/api/core/repositories/sqlalchemy_workflow_node_execution_repository.py b/api/core/repositories/sqlalchemy_workflow_node_execution_repository.py index fc160cbbe4..4399ec01cc 100644 --- a/api/core/repositories/sqlalchemy_workflow_node_execution_repository.py +++ b/api/core/repositories/sqlalchemy_workflow_node_execution_repository.py @@ -417,12 +417,10 @@ class SQLAlchemyWorkflowNodeExecutionRepository(WorkflowNodeExecutionRepository) if db_model is not None: offload_data = db_model.offload_data - else: db_model = self._to_db_model(domain_model) - offload_data = [] + offload_data = db_model.offload_data - offload_data = db_model.offload_data if domain_model.inputs is not None: result = self._truncate_and_upload( domain_model.inputs, diff --git a/api/core/workflow/graph_engine/layers/execution_limits.py b/api/core/workflow/graph_engine/layers/execution_limits.py index d74dc9b082..e39af89837 100644 --- a/api/core/workflow/graph_engine/layers/execution_limits.py +++ b/api/core/workflow/graph_engine/layers/execution_limits.py @@ -147,4 +147,4 @@ class ExecutionLimitsLayer(GraphEngineLayer): self.logger.debug("Abort command sent to engine") except Exception: - self.logger.exception("Failed to send abort command: %s") + self.logger.exception("Failed to send abort command") diff --git a/api/services/rag_pipeline/rag_pipeline.py b/api/services/rag_pipeline/rag_pipeline.py index 9ceb60666f..1e0578d12f 100644 --- a/api/services/rag_pipeline/rag_pipeline.py +++ b/api/services/rag_pipeline/rag_pipeline.py @@ -1327,14 +1327,14 @@ class RagPipelineService: """ Retry error document """ - document_pipeline_excution_log = ( + document_pipeline_execution_log = ( db.session.query(DocumentPipelineExecutionLog) .where(DocumentPipelineExecutionLog.document_id == document.id) .first() ) - if not document_pipeline_excution_log: + if not document_pipeline_execution_log: raise ValueError("Document pipeline execution log not found") - pipeline = db.session.query(Pipeline).where(Pipeline.id == document_pipeline_excution_log.pipeline_id).first() + pipeline = db.session.query(Pipeline).where(Pipeline.id == document_pipeline_execution_log.pipeline_id).first() if not pipeline: raise ValueError("Pipeline not found") # convert to app config @@ -1346,10 +1346,10 @@ class RagPipelineService: workflow=workflow, user=user, args={ - "inputs": document_pipeline_excution_log.input_data, - "start_node_id": document_pipeline_excution_log.datasource_node_id, - "datasource_type": document_pipeline_excution_log.datasource_type, - "datasource_info_list": [json.loads(document_pipeline_excution_log.datasource_info)], + "inputs": document_pipeline_execution_log.input_data, + "start_node_id": document_pipeline_execution_log.datasource_node_id, + "datasource_type": document_pipeline_execution_log.datasource_type, + "datasource_info_list": [json.loads(document_pipeline_execution_log.datasource_info)], "original_document_id": document.id, }, invoke_from=InvokeFrom.PUBLISHED, diff --git a/api/services/rag_pipeline/rag_pipeline_transform_service.py b/api/services/rag_pipeline/rag_pipeline_transform_service.py index 8435030183..db9508824b 100644 --- a/api/services/rag_pipeline/rag_pipeline_transform_service.py +++ b/api/services/rag_pipeline/rag_pipeline_transform_service.py @@ -38,11 +38,11 @@ class RagPipelineTransformService: indexing_technique = dataset.indexing_technique if not datasource_type and not indexing_technique: - return self._transfrom_to_empty_pipeline(dataset) + return self._transform_to_empty_pipeline(dataset) doc_form = dataset.doc_form if not doc_form: - return self._transfrom_to_empty_pipeline(dataset) + return self._transform_to_empty_pipeline(dataset) retrieval_model = dataset.retrieval_model pipeline_yaml = self._get_transform_yaml(doc_form, datasource_type, indexing_technique) # deal dependencies @@ -263,7 +263,7 @@ class RagPipelineTransformService: logger.debug("Installing missing pipeline plugins %s", need_install_plugin_unique_identifiers) PluginService.install_from_marketplace_pkg(tenant_id, need_install_plugin_unique_identifiers) - def _transfrom_to_empty_pipeline(self, dataset: Dataset): + def _transform_to_empty_pipeline(self, dataset: Dataset): pipeline = Pipeline( tenant_id=dataset.tenant_id, name=dataset.name,