mirror of
https://github.com/langgenius/dify.git
synced 2026-03-14 13:51:33 +08:00
- Move enqueue_draft_node_execution_trace import inside call site in workflow_service.py - Make gateway.py enterprise type imports lazy (routing dicts built on first call) - Restore typed ModelConfig in llm_generator method signatures (revert dict regression) - Fix generate_structured_output using wrong key model_parameters -> completion_params - Replace unsafe cast(str, msg.content) with get_text_content() across llm_generator - Remove duplicated payload classes from generator.py, import from core.llm_generator.entities - Gate _lookup_app_and_workspace_names and credential lookups in ops_trace_manager behind is_enterprise_telemetry_enabled()
136 lines
5.1 KiB
Python
136 lines
5.1 KiB
Python
"""
|
|
Base parser interface and utilities for OpenTelemetry node parsers.
|
|
|
|
Content gating: ``should_include_content()`` controls whether content-bearing
|
|
span attributes (inputs, outputs, prompts, completions, documents) are written.
|
|
Gate is only active in EE (``ENTERPRISE_ENABLED=True``) when
|
|
``ENTERPRISE_INCLUDE_CONTENT=False``; CE behaviour is unchanged.
|
|
"""
|
|
|
|
import json
|
|
from typing import Any, Protocol
|
|
|
|
from opentelemetry.trace import Span
|
|
from opentelemetry.trace.status import Status, StatusCode
|
|
from pydantic import BaseModel
|
|
|
|
from configs import dify_config
|
|
from core.file.models import File
|
|
from core.variables import Segment
|
|
from core.workflow.enums import NodeType
|
|
from core.workflow.graph_events import GraphNodeEventBase
|
|
from core.workflow.nodes.base.node import Node
|
|
from extensions.otel.semconv.gen_ai import ChainAttributes, GenAIAttributes
|
|
|
|
|
|
def should_include_content() -> bool:
|
|
"""Return True if content should be written to spans.
|
|
|
|
CE (ENTERPRISE_ENABLED=False): always True — no behaviour change.
|
|
EE: follows ENTERPRISE_INCLUDE_CONTENT (default True).
|
|
"""
|
|
if not dify_config.ENTERPRISE_ENABLED:
|
|
return True
|
|
return dify_config.ENTERPRISE_INCLUDE_CONTENT
|
|
|
|
|
|
def safe_json_dumps(obj: Any, ensure_ascii: bool = False) -> str:
|
|
"""
|
|
Safely serialize objects to JSON, handling non-serializable types.
|
|
|
|
Handles:
|
|
- Segment types (ArrayFileSegment, FileSegment, etc.) - converts to their value
|
|
- File objects - converts to dict using to_dict()
|
|
- BaseModel objects - converts using model_dump()
|
|
- Other types - falls back to str() representation
|
|
|
|
Args:
|
|
obj: Object to serialize
|
|
ensure_ascii: Whether to ensure ASCII encoding
|
|
|
|
Returns:
|
|
JSON string representation of the object
|
|
"""
|
|
|
|
def _convert_value(value: Any) -> Any:
|
|
"""Recursively convert non-serializable values."""
|
|
if value is None:
|
|
return None
|
|
if isinstance(value, (bool, int, float, str)):
|
|
return value
|
|
if isinstance(value, Segment):
|
|
# Convert Segment to its underlying value
|
|
return _convert_value(value.value)
|
|
if isinstance(value, File):
|
|
# Convert File to dict
|
|
return value.to_dict()
|
|
if isinstance(value, BaseModel):
|
|
# Convert Pydantic model to dict
|
|
return _convert_value(value.model_dump(mode="json"))
|
|
if isinstance(value, dict):
|
|
return {k: _convert_value(v) for k, v in value.items()}
|
|
if isinstance(value, (list, tuple)):
|
|
return [_convert_value(item) for item in value]
|
|
# Fallback to string representation for unknown types
|
|
return str(value)
|
|
|
|
try:
|
|
converted = _convert_value(obj)
|
|
return json.dumps(converted, ensure_ascii=ensure_ascii)
|
|
except (TypeError, ValueError) as e:
|
|
# If conversion still fails, return error message as string
|
|
return json.dumps(
|
|
{"error": f"Failed to serialize: {type(obj).__name__}", "message": str(e)}, ensure_ascii=ensure_ascii
|
|
)
|
|
|
|
|
|
class NodeOTelParser(Protocol):
|
|
"""Parser interface for node-specific OpenTelemetry enrichment."""
|
|
|
|
def parse(
|
|
self, *, node: Node, span: "Span", error: Exception | None, result_event: GraphNodeEventBase | None = None
|
|
) -> None: ...
|
|
|
|
|
|
class DefaultNodeOTelParser:
|
|
"""Fallback parser used when no node-specific parser is registered."""
|
|
|
|
def parse(
|
|
self, *, node: Node, span: "Span", error: Exception | None, result_event: GraphNodeEventBase | None = None
|
|
) -> None:
|
|
span.set_attribute("node.id", node.id)
|
|
if node.execution_id:
|
|
span.set_attribute("node.execution_id", node.execution_id)
|
|
if hasattr(node, "node_type") and node.node_type:
|
|
span.set_attribute("node.type", node.node_type.value)
|
|
|
|
span.set_attribute(GenAIAttributes.FRAMEWORK, "dify")
|
|
|
|
node_type = getattr(node, "node_type", None)
|
|
if isinstance(node_type, NodeType):
|
|
if node_type == NodeType.LLM:
|
|
span.set_attribute(GenAIAttributes.SPAN_KIND, "LLM")
|
|
elif node_type == NodeType.KNOWLEDGE_RETRIEVAL:
|
|
span.set_attribute(GenAIAttributes.SPAN_KIND, "RETRIEVER")
|
|
elif node_type == NodeType.TOOL:
|
|
span.set_attribute(GenAIAttributes.SPAN_KIND, "TOOL")
|
|
else:
|
|
span.set_attribute(GenAIAttributes.SPAN_KIND, "TASK")
|
|
else:
|
|
span.set_attribute(GenAIAttributes.SPAN_KIND, "TASK")
|
|
|
|
# Extract inputs and outputs from result_event
|
|
if result_event and result_event.node_run_result:
|
|
node_run_result = result_event.node_run_result
|
|
if should_include_content():
|
|
if node_run_result.inputs:
|
|
span.set_attribute(ChainAttributes.INPUT_VALUE, safe_json_dumps(node_run_result.inputs))
|
|
if node_run_result.outputs:
|
|
span.set_attribute(ChainAttributes.OUTPUT_VALUE, safe_json_dumps(node_run_result.outputs))
|
|
|
|
if error:
|
|
span.record_exception(error)
|
|
span.set_status(Status(StatusCode.ERROR, str(error)))
|
|
else:
|
|
span.set_status(Status(StatusCode.OK))
|