mirror of
https://github.com/langgenius/dify.git
synced 2026-05-10 05:56:31 +08:00
Signed-off-by: majiayu000 <1835304752@qq.com> Signed-off-by: dependabot[bot] <support@github.com> Signed-off-by: NeatGuyCoding <15627489+NeatGuyCoding@users.noreply.github.com> Signed-off-by: -LAN- <laipz8200@outlook.com> Signed-off-by: yihong0618 <zouzou0208@gmail.com> Co-authored-by: QuantumGhost <obelisk.reg+git@gmail.com> Co-authored-by: 盐粒 Yanli <yanli@dify.ai> Co-authored-by: wangxiaolei <fatelei@gmail.com> Co-authored-by: Stephen Zhou <38493346+hyoban@users.noreply.github.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> Co-authored-by: Cursx <33718736+Cursx@users.noreply.github.com> Co-authored-by: autofix-ci[bot] <114827586+autofix-ci[bot]@users.noreply.github.com> Co-authored-by: lif <1835304752@qq.com> Co-authored-by: 非法操作 <hjlarry@163.com> Co-authored-by: Asuka Minato <i@asukaminato.eu.org> Co-authored-by: fenglin <790872612@qq.com> Co-authored-by: qiaofenglin <qiaofenglin@baidu.com> Co-authored-by: -LAN- <laipz8200@outlook.com> Co-authored-by: TomoOkuyama <49631611+TomoOkuyama@users.noreply.github.com> Co-authored-by: Tomo Okuyama <tomo.okuyama@intersystems.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: zyssyz123 <916125788@qq.com> Co-authored-by: hj24 <mambahj24@gmail.com> Co-authored-by: Coding On Star <447357187@qq.com> Co-authored-by: CodingOnStar <hanxujiang@dify.ai> Co-authored-by: yyh <92089059+lyzno1@users.noreply.github.com> Co-authored-by: Xiangxuan Qu <fghpdf@outlook.com> Co-authored-by: fghpdf <fghpdf@users.noreply.github.com> Co-authored-by: coopercoder <whitetiger0127@163.com> Co-authored-by: zhaiguangpeng <zhaiguangpeng@didiglobal.com> Co-authored-by: Junyan Qin (Chin) <rockchinq@gmail.com> Co-authored-by: E.G <146701565+GlobalStar117@users.noreply.github.com> Co-authored-by: GlobalStar117 <GlobalStar117@users.noreply.github.com> Co-authored-by: Claude Haiku 4.5 <noreply@anthropic.com> Co-authored-by: CodingOnStar <hanxujiang@dify.com> Co-authored-by: crazywoola <100913391+crazywoola@users.noreply.github.com> Co-authored-by: heyszt <270985384@qq.com> Co-authored-by: NeatGuyCoding <15627489+NeatGuyCoding@users.noreply.github.com> Co-authored-by: Yeuoly <45712896+Yeuoly@users.noreply.github.com> Co-authored-by: zxhlyh <jasonapring2015@outlook.com> Co-authored-by: moonpanda <chuanzegao@163.com> Co-authored-by: warlocgao <warlocgao@tencent.com> Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: claude[bot] <41898282+claude[bot]@users.noreply.github.com> Co-authored-by: KVOJJJin <jzongcode@gmail.com> Co-authored-by: eux <euxx@users.noreply.github.com> Co-authored-by: bangjiehan <bangjiehan@gmail.com> Co-authored-by: FFXN <31929997+FFXN@users.noreply.github.com> Co-authored-by: Jyong <76649700+JohnJyong@users.noreply.github.com> Co-authored-by: Nie Ronghua <nieronghua@sf-express.com> Co-authored-by: JQSevenMiao <141806521+JQSevenMiao@users.noreply.github.com> Co-authored-by: jiasiqi <jiasiqi3@tal.com> Co-authored-by: Seokrin Taron Sung <sungsjade@gmail.com> Co-authored-by: CrabSAMA <40541269+CrabSAMA@users.noreply.github.com> Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> Co-authored-by: yihong <zouzou0208@gmail.com> Co-authored-by: Joel <iamjoel007@gmail.com> Co-authored-by: Wu Tianwei <30284043+WTW0313@users.noreply.github.com> Co-authored-by: yessenia <yessenia.contact@gmail.com> Co-authored-by: Jax <anobaka@qq.com> Co-authored-by: niveshdandyan <155956228+niveshdandyan@users.noreply.github.com> Co-authored-by: OSS Contributor <oss-contributor@example.com> Co-authored-by: niveshdandyan <niveshdandyan@users.noreply.github.com> Co-authored-by: Sean Kenneth Doherty <Smaster7772@gmail.com>
156 lines
5.1 KiB
Python
156 lines
5.1 KiB
Python
"""
|
|
Parser for LLM nodes that captures LLM-specific metadata.
|
|
"""
|
|
|
|
import logging
|
|
from collections.abc import Mapping
|
|
from typing import Any
|
|
|
|
from opentelemetry.trace import Span
|
|
|
|
from core.workflow.graph_events import GraphNodeEventBase
|
|
from core.workflow.nodes.base.node import Node
|
|
from extensions.otel.parser.base import DefaultNodeOTelParser, safe_json_dumps
|
|
from extensions.otel.semconv.gen_ai import LLMAttributes
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
def _format_input_messages(process_data: Mapping[str, Any]) -> str:
|
|
"""
|
|
Format input messages from process_data for LLM spans.
|
|
|
|
Args:
|
|
process_data: Process data containing prompts
|
|
|
|
Returns:
|
|
JSON string of formatted input messages
|
|
"""
|
|
try:
|
|
if not isinstance(process_data, dict):
|
|
return safe_json_dumps([])
|
|
|
|
prompts = process_data.get("prompts", [])
|
|
if not prompts:
|
|
return safe_json_dumps([])
|
|
|
|
valid_roles = {"system", "user", "assistant", "tool"}
|
|
input_messages = []
|
|
for prompt in prompts:
|
|
if not isinstance(prompt, dict):
|
|
continue
|
|
|
|
role = prompt.get("role", "")
|
|
text = prompt.get("text", "")
|
|
|
|
if not role or role not in valid_roles:
|
|
continue
|
|
|
|
if text:
|
|
message = {"role": role, "parts": [{"type": "text", "content": text}]}
|
|
input_messages.append(message)
|
|
|
|
return safe_json_dumps(input_messages)
|
|
except Exception as e:
|
|
logger.warning("Failed to format input messages: %s", e, exc_info=True)
|
|
return safe_json_dumps([])
|
|
|
|
|
|
def _format_output_messages(outputs: Mapping[str, Any]) -> str:
|
|
"""
|
|
Format output messages from outputs for LLM spans.
|
|
|
|
Args:
|
|
outputs: Output data containing text and finish_reason
|
|
|
|
Returns:
|
|
JSON string of formatted output messages
|
|
"""
|
|
try:
|
|
if not isinstance(outputs, dict):
|
|
return safe_json_dumps([])
|
|
|
|
text = outputs.get("text", "")
|
|
finish_reason = outputs.get("finish_reason", "")
|
|
|
|
if not text:
|
|
return safe_json_dumps([])
|
|
|
|
valid_finish_reasons = {"stop", "length", "content_filter", "tool_call", "error"}
|
|
if finish_reason not in valid_finish_reasons:
|
|
finish_reason = "stop"
|
|
|
|
output_message = {
|
|
"role": "assistant",
|
|
"parts": [{"type": "text", "content": text}],
|
|
"finish_reason": finish_reason,
|
|
}
|
|
|
|
return safe_json_dumps([output_message])
|
|
except Exception as e:
|
|
logger.warning("Failed to format output messages: %s", e, exc_info=True)
|
|
return safe_json_dumps([])
|
|
|
|
|
|
class LLMNodeOTelParser:
|
|
"""Parser for LLM nodes that captures LLM-specific metadata."""
|
|
|
|
def __init__(self) -> None:
|
|
self._delegate = DefaultNodeOTelParser()
|
|
|
|
def parse(
|
|
self, *, node: Node, span: "Span", error: Exception | None, result_event: GraphNodeEventBase | None = None
|
|
) -> None:
|
|
self._delegate.parse(node=node, span=span, error=error, result_event=result_event)
|
|
|
|
if not result_event or not result_event.node_run_result:
|
|
return
|
|
|
|
node_run_result = result_event.node_run_result
|
|
process_data = node_run_result.process_data or {}
|
|
outputs = node_run_result.outputs or {}
|
|
|
|
# Extract usage data (from process_data or outputs)
|
|
usage_data = process_data.get("usage") or outputs.get("usage") or {}
|
|
|
|
# Model and provider information
|
|
model_name = process_data.get("model_name") or ""
|
|
model_provider = process_data.get("model_provider") or ""
|
|
|
|
if model_name:
|
|
span.set_attribute(LLMAttributes.REQUEST_MODEL, model_name)
|
|
if model_provider:
|
|
span.set_attribute(LLMAttributes.PROVIDER_NAME, model_provider)
|
|
|
|
# Token usage
|
|
if usage_data:
|
|
prompt_tokens = usage_data.get("prompt_tokens", 0)
|
|
completion_tokens = usage_data.get("completion_tokens", 0)
|
|
total_tokens = usage_data.get("total_tokens", 0)
|
|
|
|
span.set_attribute(LLMAttributes.USAGE_INPUT_TOKENS, prompt_tokens)
|
|
span.set_attribute(LLMAttributes.USAGE_OUTPUT_TOKENS, completion_tokens)
|
|
span.set_attribute(LLMAttributes.USAGE_TOTAL_TOKENS, total_tokens)
|
|
|
|
# Prompts and completion
|
|
prompts = process_data.get("prompts", [])
|
|
if prompts:
|
|
prompts_json = safe_json_dumps(prompts)
|
|
span.set_attribute(LLMAttributes.PROMPT, prompts_json)
|
|
|
|
text_output = str(outputs.get("text", ""))
|
|
if text_output:
|
|
span.set_attribute(LLMAttributes.COMPLETION, text_output)
|
|
|
|
# Finish reason
|
|
finish_reason = outputs.get("finish_reason") or ""
|
|
if finish_reason:
|
|
span.set_attribute(LLMAttributes.RESPONSE_FINISH_REASON, finish_reason)
|
|
|
|
# Structured input/output messages
|
|
gen_ai_input_message = _format_input_messages(process_data)
|
|
gen_ai_output_message = _format_output_messages(outputs)
|
|
|
|
span.set_attribute(LLMAttributes.INPUT_MESSAGE, gen_ai_input_message)
|
|
span.set_attribute(LLMAttributes.OUTPUT_MESSAGE, gen_ai_output_message)
|