test(trigger): add container integration tests for trigger (#29527)

Signed-off-by: Stream <Stream_2@qq.com>
Co-authored-by: autofix-ci[bot] <114827586+autofix-ci[bot]@users.noreply.github.com>
This commit is contained in:
Stream 2025-12-12 11:31:34 +08:00 committed by GitHub
parent 61ee1b9094
commit 8daf9ce98d
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 1094 additions and 0 deletions

View File

@ -0,0 +1,182 @@
"""
Fixtures for trigger integration tests.
This module provides fixtures for creating test data (tenant, account, app)
and mock objects used across trigger-related tests.
"""
from __future__ import annotations
from collections.abc import Generator
from typing import Any
import pytest
from sqlalchemy.orm import Session
from models.account import Account, Tenant, TenantAccountJoin, TenantAccountRole
from models.model import App
@pytest.fixture
def tenant_and_account(db_session_with_containers: Session) -> Generator[tuple[Tenant, Account], None, None]:
"""
Create a tenant and account for testing.
This fixture creates a tenant, account, and their association,
then cleans up after the test completes.
Yields:
tuple[Tenant, Account]: The created tenant and account
"""
tenant = Tenant(name="trigger-e2e")
account = Account(name="tester", email="tester@example.com", interface_language="en-US")
db_session_with_containers.add_all([tenant, account])
db_session_with_containers.commit()
join = TenantAccountJoin(tenant_id=tenant.id, account_id=account.id, role=TenantAccountRole.OWNER.value)
db_session_with_containers.add(join)
db_session_with_containers.commit()
yield tenant, account
# Cleanup
db_session_with_containers.query(TenantAccountJoin).filter_by(tenant_id=tenant.id).delete()
db_session_with_containers.query(Account).filter_by(id=account.id).delete()
db_session_with_containers.query(Tenant).filter_by(id=tenant.id).delete()
db_session_with_containers.commit()
@pytest.fixture
def app_model(
db_session_with_containers: Session, tenant_and_account: tuple[Tenant, Account]
) -> Generator[App, None, None]:
"""
Create an app for testing.
This fixture creates a workflow app associated with the tenant and account,
then cleans up after the test completes.
Yields:
App: The created app
"""
tenant, account = tenant_and_account
app = App(
tenant_id=tenant.id,
name="trigger-app",
description="trigger e2e",
mode="workflow",
icon_type="emoji",
icon="robot",
icon_background="#FFEAD5",
enable_site=True,
enable_api=True,
api_rpm=100,
api_rph=1000,
is_demo=False,
is_public=False,
is_universal=False,
created_by=account.id,
)
db_session_with_containers.add(app)
db_session_with_containers.commit()
yield app
# Cleanup - delete related records first
from models.trigger import (
AppTrigger,
TriggerSubscription,
WorkflowPluginTrigger,
WorkflowSchedulePlan,
WorkflowTriggerLog,
WorkflowWebhookTrigger,
)
from models.workflow import Workflow
db_session_with_containers.query(WorkflowTriggerLog).filter_by(app_id=app.id).delete()
db_session_with_containers.query(WorkflowSchedulePlan).filter_by(app_id=app.id).delete()
db_session_with_containers.query(WorkflowWebhookTrigger).filter_by(app_id=app.id).delete()
db_session_with_containers.query(WorkflowPluginTrigger).filter_by(app_id=app.id).delete()
db_session_with_containers.query(AppTrigger).filter_by(app_id=app.id).delete()
db_session_with_containers.query(TriggerSubscription).filter_by(tenant_id=tenant.id).delete()
db_session_with_containers.query(Workflow).filter_by(app_id=app.id).delete()
db_session_with_containers.query(App).filter_by(id=app.id).delete()
db_session_with_containers.commit()
class MockCeleryGroup:
"""Mock for celery group() function that collects dispatched tasks."""
def __init__(self) -> None:
self.collected: list[dict[str, Any]] = []
self._applied = False
def __call__(self, items: Any) -> MockCeleryGroup:
self.collected = list(items)
return self
def apply_async(self) -> None:
self._applied = True
@property
def applied(self) -> bool:
return self._applied
class MockCelerySignature:
"""Mock for celery task signature that returns task info dict."""
def s(self, schedule_id: str) -> dict[str, str]:
return {"schedule_id": schedule_id}
@pytest.fixture
def mock_celery_group() -> MockCeleryGroup:
"""
Provide a mock celery group for testing task dispatch.
Returns:
MockCeleryGroup: Mock group that collects dispatched tasks
"""
return MockCeleryGroup()
@pytest.fixture
def mock_celery_signature() -> MockCelerySignature:
"""
Provide a mock celery signature for testing task dispatch.
Returns:
MockCelerySignature: Mock signature generator
"""
return MockCelerySignature()
class MockPluginSubscription:
"""Mock plugin subscription for testing plugin triggers."""
def __init__(
self,
subscription_id: str = "sub-1",
tenant_id: str = "tenant-1",
provider_id: str = "provider-1",
) -> None:
self.id = subscription_id
self.tenant_id = tenant_id
self.provider_id = provider_id
self.credentials: dict[str, str] = {"token": "secret"}
self.credential_type = "api-key"
def to_entity(self) -> MockPluginSubscription:
return self
@pytest.fixture
def mock_plugin_subscription() -> MockPluginSubscription:
"""
Provide a mock plugin subscription for testing.
Returns:
MockPluginSubscription: Mock subscription instance
"""
return MockPluginSubscription()

View File

@ -0,0 +1,911 @@
from __future__ import annotations
import importlib
import json
import time
from datetime import timedelta
from types import SimpleNamespace
from typing import Any
import pytest
from flask import Flask, Response
from flask.testing import FlaskClient
from sqlalchemy.orm import Session
from configs import dify_config
from core.plugin.entities.request import TriggerInvokeEventResponse
from core.trigger.debug import event_selectors
from core.trigger.debug.event_bus import TriggerDebugEventBus
from core.trigger.debug.event_selectors import PluginTriggerDebugEventPoller, WebhookTriggerDebugEventPoller
from core.trigger.debug.events import PluginTriggerDebugEvent, build_plugin_pool_key
from core.workflow.enums import NodeType
from libs.datetime_utils import naive_utc_now
from models.account import Account, Tenant
from models.enums import AppTriggerStatus, AppTriggerType, CreatorUserRole, WorkflowTriggerStatus
from models.model import App
from models.trigger import (
AppTrigger,
TriggerSubscription,
WorkflowPluginTrigger,
WorkflowSchedulePlan,
WorkflowTriggerLog,
WorkflowWebhookTrigger,
)
from models.workflow import Workflow
from schedule import workflow_schedule_task
from schedule.workflow_schedule_task import poll_workflow_schedules
from services import feature_service as feature_service_module
from services.trigger import webhook_service
from services.trigger.schedule_service import ScheduleService
from services.workflow_service import WorkflowService
from tasks import trigger_processing_tasks
from .conftest import MockCeleryGroup, MockCelerySignature, MockPluginSubscription
# Test constants
WEBHOOK_ID_PRODUCTION = "wh1234567890123456789012"
WEBHOOK_ID_DEBUG = "whdebug1234567890123456"
TEST_TRIGGER_URL = "https://trigger.example.com/base"
def _build_workflow_graph(root_node_id: str, trigger_type: NodeType) -> str:
"""Build a minimal workflow graph JSON for testing."""
node_data: dict[str, Any] = {"type": trigger_type.value, "title": "trigger"}
if trigger_type == NodeType.TRIGGER_WEBHOOK:
node_data.update(
{
"method": "POST",
"content_type": "application/json",
"headers": [],
"params": [],
"body": [],
}
)
graph = {
"nodes": [
{"id": root_node_id, "data": node_data},
{"id": "answer-1", "data": {"type": NodeType.ANSWER.value, "title": "answer"}},
],
"edges": [{"source": root_node_id, "target": "answer-1", "sourceHandle": "success"}],
}
return json.dumps(graph)
def test_publish_blocks_start_and_trigger_coexistence(
db_session_with_containers: Session,
tenant_and_account: tuple[Tenant, Account],
app_model: App,
monkeypatch: pytest.MonkeyPatch,
) -> None:
"""Publishing should fail when both start and trigger nodes coexist."""
tenant, account = tenant_and_account
graph = {
"nodes": [
{"id": "start", "data": {"type": NodeType.START.value}},
{"id": "trig", "data": {"type": NodeType.TRIGGER_WEBHOOK.value}},
],
"edges": [],
}
draft_workflow = Workflow.new(
tenant_id=tenant.id,
app_id=app_model.id,
type="workflow",
version=Workflow.VERSION_DRAFT,
graph=json.dumps(graph),
features=json.dumps({}),
created_by=account.id,
environment_variables=[],
conversation_variables=[],
rag_pipeline_variables=[],
)
db_session_with_containers.add(draft_workflow)
db_session_with_containers.commit()
workflow_service = WorkflowService()
monkeypatch.setattr(
feature_service_module.FeatureService,
"get_system_features",
classmethod(lambda _cls: SimpleNamespace(plugin_manager=SimpleNamespace(enabled=False))),
)
monkeypatch.setattr("services.workflow_service.dify_config", SimpleNamespace(BILLING_ENABLED=False))
with pytest.raises(ValueError, match="Start node and trigger nodes cannot coexist"):
workflow_service.publish_workflow(session=db_session_with_containers, app_model=app_model, account=account)
def test_trigger_url_uses_config_base(monkeypatch: pytest.MonkeyPatch) -> None:
"""TRIGGER_URL config should be reflected in generated webhook and plugin endpoints."""
original_url = getattr(dify_config, "TRIGGER_URL", None)
try:
monkeypatch.setattr(dify_config, "TRIGGER_URL", TEST_TRIGGER_URL)
endpoint_module = importlib.reload(importlib.import_module("core.trigger.utils.endpoint"))
assert (
endpoint_module.generate_webhook_trigger_endpoint(WEBHOOK_ID_PRODUCTION)
== f"{TEST_TRIGGER_URL}/triggers/webhook/{WEBHOOK_ID_PRODUCTION}"
)
assert (
endpoint_module.generate_webhook_trigger_endpoint(WEBHOOK_ID_PRODUCTION, True)
== f"{TEST_TRIGGER_URL}/triggers/webhook-debug/{WEBHOOK_ID_PRODUCTION}"
)
assert (
endpoint_module.generate_plugin_trigger_endpoint_url("end-1") == f"{TEST_TRIGGER_URL}/triggers/plugin/end-1"
)
finally:
# Restore original config and reload module
if original_url is not None:
monkeypatch.setattr(dify_config, "TRIGGER_URL", original_url)
importlib.reload(importlib.import_module("core.trigger.utils.endpoint"))
def test_webhook_trigger_creates_trigger_log(
test_client_with_containers: FlaskClient,
db_session_with_containers: Session,
tenant_and_account: tuple[Tenant, Account],
app_model: App,
monkeypatch: pytest.MonkeyPatch,
) -> None:
"""Production webhook trigger should create a trigger log in the database."""
tenant, account = tenant_and_account
webhook_node_id = "webhook-node"
graph_json = _build_workflow_graph(webhook_node_id, NodeType.TRIGGER_WEBHOOK)
published_workflow = Workflow.new(
tenant_id=tenant.id,
app_id=app_model.id,
type="workflow",
version=Workflow.version_from_datetime(naive_utc_now()),
graph=graph_json,
features=json.dumps({}),
created_by=account.id,
environment_variables=[],
conversation_variables=[],
rag_pipeline_variables=[],
)
db_session_with_containers.add(published_workflow)
app_model.workflow_id = published_workflow.id
db_session_with_containers.commit()
webhook_trigger = WorkflowWebhookTrigger(
app_id=app_model.id,
node_id=webhook_node_id,
tenant_id=tenant.id,
webhook_id=WEBHOOK_ID_PRODUCTION,
created_by=account.id,
)
app_trigger = AppTrigger(
tenant_id=tenant.id,
app_id=app_model.id,
node_id=webhook_node_id,
trigger_type=AppTriggerType.TRIGGER_WEBHOOK,
status=AppTriggerStatus.ENABLED,
title="webhook",
)
db_session_with_containers.add_all([webhook_trigger, app_trigger])
db_session_with_containers.commit()
def _fake_trigger_workflow_async(session: Session, user: Any, trigger_data: Any) -> SimpleNamespace:
log = WorkflowTriggerLog(
tenant_id=trigger_data.tenant_id,
app_id=trigger_data.app_id,
workflow_id=trigger_data.workflow_id,
root_node_id=trigger_data.root_node_id,
trigger_metadata=trigger_data.trigger_metadata.model_dump_json() if trigger_data.trigger_metadata else "{}",
trigger_type=trigger_data.trigger_type,
workflow_run_id=None,
outputs=None,
trigger_data=trigger_data.model_dump_json(),
inputs=json.dumps(dict(trigger_data.inputs)),
status=WorkflowTriggerStatus.SUCCEEDED,
error="",
queue_name="triggered_workflow_dispatcher",
celery_task_id="celery-test",
created_by_role=CreatorUserRole.ACCOUNT,
created_by=account.id,
)
session.add(log)
session.commit()
return SimpleNamespace(workflow_trigger_log_id=log.id, task_id=None, status="queued", queue="test")
monkeypatch.setattr(
webhook_service.AsyncWorkflowService,
"trigger_workflow_async",
_fake_trigger_workflow_async,
)
response = test_client_with_containers.post(f"/triggers/webhook/{webhook_trigger.webhook_id}", json={"foo": "bar"})
assert response.status_code == 200
db_session_with_containers.expire_all()
logs = db_session_with_containers.query(WorkflowTriggerLog).filter_by(app_id=app_model.id).all()
assert logs, "Webhook trigger should create trigger log"
@pytest.mark.parametrize("schedule_type", ["visual", "cron"])
def test_schedule_poll_dispatches_due_plan(
db_session_with_containers: Session,
tenant_and_account: tuple[Tenant, Account],
app_model: App,
mock_celery_group: MockCeleryGroup,
mock_celery_signature: MockCelerySignature,
monkeypatch: pytest.MonkeyPatch,
schedule_type: str,
) -> None:
"""Schedule plans (both visual and cron) should be polled and dispatched when due."""
tenant, _ = tenant_and_account
app_trigger = AppTrigger(
tenant_id=tenant.id,
app_id=app_model.id,
node_id=f"schedule-{schedule_type}",
trigger_type=AppTriggerType.TRIGGER_SCHEDULE,
status=AppTriggerStatus.ENABLED,
title=f"schedule-{schedule_type}",
)
plan = WorkflowSchedulePlan(
app_id=app_model.id,
node_id=f"schedule-{schedule_type}",
tenant_id=tenant.id,
cron_expression="* * * * *",
timezone="UTC",
next_run_at=naive_utc_now() - timedelta(minutes=1),
)
db_session_with_containers.add_all([app_trigger, plan])
db_session_with_containers.commit()
next_time = naive_utc_now() + timedelta(hours=1)
monkeypatch.setattr(workflow_schedule_task, "calculate_next_run_at", lambda *_args, **_kwargs: next_time)
monkeypatch.setattr(workflow_schedule_task, "group", mock_celery_group)
monkeypatch.setattr(workflow_schedule_task, "run_schedule_trigger", mock_celery_signature)
poll_workflow_schedules()
assert mock_celery_group.collected, f"Should dispatch signatures for due {schedule_type} schedules"
scheduled_ids = {sig["schedule_id"] for sig in mock_celery_group.collected}
assert plan.id in scheduled_ids
def test_schedule_visual_debug_poll_generates_event(monkeypatch: pytest.MonkeyPatch) -> None:
"""Visual mode schedule node should generate event in single-step debug."""
base_now = naive_utc_now()
monkeypatch.setattr(event_selectors, "naive_utc_now", lambda: base_now)
monkeypatch.setattr(
event_selectors,
"calculate_next_run_at",
lambda *_args, **_kwargs: base_now - timedelta(minutes=1),
)
node_config = {
"id": "schedule-visual",
"data": {
"type": NodeType.TRIGGER_SCHEDULE.value,
"mode": "visual",
"frequency": "daily",
"visual_config": {"time": "3:00 PM"},
"timezone": "UTC",
},
}
poller = event_selectors.ScheduleTriggerDebugEventPoller(
tenant_id="tenant",
user_id="user",
app_id="app",
node_config=node_config,
node_id="schedule-visual",
)
event = poller.poll()
assert event is not None
assert event.workflow_args["inputs"] == {}
def test_plugin_trigger_dispatches_and_debug_events(
test_client_with_containers: FlaskClient,
mock_plugin_subscription: MockPluginSubscription,
monkeypatch: pytest.MonkeyPatch,
) -> None:
"""Plugin trigger endpoint should dispatch events and generate debug events."""
endpoint_id = "1cc7fa12-3f7b-4f6a-9c8d-1234567890ab"
debug_events: list[dict[str, Any]] = []
dispatched_payloads: list[dict[str, Any]] = []
def _fake_process_endpoint(_endpoint_id: str, _request: Any) -> Response:
dispatch_data = {
"user_id": "end-user",
"tenant_id": mock_plugin_subscription.tenant_id,
"endpoint_id": _endpoint_id,
"provider_id": mock_plugin_subscription.provider_id,
"subscription_id": mock_plugin_subscription.id,
"timestamp": int(time.time()),
"events": ["created", "updated"],
"request_id": f"req-{_endpoint_id}",
}
trigger_processing_tasks.dispatch_triggered_workflows_async.delay(dispatch_data)
return Response("ok", status=202)
monkeypatch.setattr(
"services.trigger.trigger_service.TriggerService.process_endpoint",
staticmethod(_fake_process_endpoint),
)
monkeypatch.setattr(
trigger_processing_tasks.TriggerDebugEventBus,
"dispatch",
staticmethod(lambda **kwargs: debug_events.append(kwargs) or 1),
)
def _fake_delay(dispatch_data: dict[str, Any]) -> None:
dispatched_payloads.append(dispatch_data)
trigger_processing_tasks.dispatch_trigger_debug_event(
events=dispatch_data["events"],
user_id=dispatch_data["user_id"],
timestamp=dispatch_data["timestamp"],
request_id=dispatch_data["request_id"],
subscription=mock_plugin_subscription,
)
monkeypatch.setattr(
trigger_processing_tasks.dispatch_triggered_workflows_async,
"delay",
staticmethod(_fake_delay),
)
response = test_client_with_containers.post(f"/triggers/plugin/{endpoint_id}", json={"hello": "world"})
assert response.status_code == 202
assert dispatched_payloads, "Plugin trigger should enqueue workflow dispatch payload"
assert debug_events, "Plugin trigger should dispatch debug events"
dispatched_event_names = {event["event"].name for event in debug_events}
assert dispatched_event_names == {"created", "updated"}
def test_webhook_debug_dispatches_event(
test_client_with_containers: FlaskClient,
db_session_with_containers: Session,
tenant_and_account: tuple[Tenant, Account],
app_model: App,
monkeypatch: pytest.MonkeyPatch,
) -> None:
"""Webhook single-step debug should dispatch debug event and be pollable."""
tenant, account = tenant_and_account
webhook_node_id = "webhook-debug-node"
graph_json = _build_workflow_graph(webhook_node_id, NodeType.TRIGGER_WEBHOOK)
draft_workflow = Workflow.new(
tenant_id=tenant.id,
app_id=app_model.id,
type="workflow",
version=Workflow.VERSION_DRAFT,
graph=graph_json,
features=json.dumps({}),
created_by=account.id,
environment_variables=[],
conversation_variables=[],
rag_pipeline_variables=[],
)
db_session_with_containers.add(draft_workflow)
db_session_with_containers.commit()
webhook_trigger = WorkflowWebhookTrigger(
app_id=app_model.id,
node_id=webhook_node_id,
tenant_id=tenant.id,
webhook_id=WEBHOOK_ID_DEBUG,
created_by=account.id,
)
db_session_with_containers.add(webhook_trigger)
db_session_with_containers.commit()
debug_events: list[dict[str, Any]] = []
original_dispatch = TriggerDebugEventBus.dispatch
monkeypatch.setattr(
"controllers.trigger.webhook.TriggerDebugEventBus.dispatch",
lambda **kwargs: (debug_events.append(kwargs), original_dispatch(**kwargs))[1],
)
# Listener polls first to enter waiting pool
poller = WebhookTriggerDebugEventPoller(
tenant_id=tenant.id,
user_id=account.id,
app_id=app_model.id,
node_config=draft_workflow.get_node_config_by_id(webhook_node_id),
node_id=webhook_node_id,
)
assert poller.poll() is None
response = test_client_with_containers.post(
f"/triggers/webhook-debug/{webhook_trigger.webhook_id}",
json={"foo": "bar"},
headers={"Content-Type": "application/json"},
)
assert response.status_code == 200
assert debug_events, "Debug event should be sent to event bus"
# Second poll should get the event
event = poller.poll()
assert event is not None
assert event.workflow_args["inputs"]["webhook_body"]["foo"] == "bar"
assert debug_events[0]["pool_key"].endswith(f":{app_model.id}:{webhook_node_id}")
def test_plugin_single_step_debug_flow(
flask_app_with_containers: Flask,
monkeypatch: pytest.MonkeyPatch,
) -> None:
"""Plugin single-step debug: listen -> dispatch event -> poller receives and returns variables."""
tenant_id = "tenant-1"
app_id = "app-1"
user_id = "user-1"
node_id = "plugin-node"
provider_id = "langgenius/provider-1/provider-1"
node_config = {
"id": node_id,
"data": {
"type": NodeType.TRIGGER_PLUGIN.value,
"title": "plugin",
"plugin_id": "plugin-1",
"plugin_unique_identifier": "plugin-1",
"provider_id": provider_id,
"event_name": "created",
"subscription_id": "sub-1",
"parameters": {},
},
}
# Start listening
poller = PluginTriggerDebugEventPoller(
tenant_id=tenant_id,
user_id=user_id,
app_id=app_id,
node_config=node_config,
node_id=node_id,
)
assert poller.poll() is None
from core.trigger.debug.events import build_plugin_pool_key
pool_key = build_plugin_pool_key(
tenant_id=tenant_id,
provider_id=provider_id,
subscription_id="sub-1",
name="created",
)
TriggerDebugEventBus.dispatch(
tenant_id=tenant_id,
event=PluginTriggerDebugEvent(
timestamp=int(time.time()),
user_id=user_id,
name="created",
request_id="req-1",
subscription_id="sub-1",
provider_id="provider-1",
),
pool_key=pool_key,
)
from core.plugin.entities.request import TriggerInvokeEventResponse
monkeypatch.setattr(
"services.trigger.trigger_service.TriggerService.invoke_trigger_event",
staticmethod(
lambda **_kwargs: TriggerInvokeEventResponse(
variables={"echo": "pong"},
cancelled=False,
)
),
)
event = poller.poll()
assert event is not None
assert event.workflow_args["inputs"]["echo"] == "pong"
def test_schedule_trigger_creates_trigger_log(
db_session_with_containers: Session,
tenant_and_account: tuple[Tenant, Account],
app_model: App,
monkeypatch: pytest.MonkeyPatch,
) -> None:
"""Schedule trigger execution should create WorkflowTriggerLog in database."""
from tasks import workflow_schedule_tasks
tenant, account = tenant_and_account
# Create published workflow with schedule trigger node
schedule_node_id = "schedule-node"
graph = {
"nodes": [
{
"id": schedule_node_id,
"data": {
"type": NodeType.TRIGGER_SCHEDULE.value,
"title": "schedule",
"mode": "cron",
"cron_expression": "0 9 * * *",
"timezone": "UTC",
},
},
{"id": "answer-1", "data": {"type": NodeType.ANSWER.value, "title": "answer"}},
],
"edges": [{"source": schedule_node_id, "target": "answer-1", "sourceHandle": "success"}],
}
published_workflow = Workflow.new(
tenant_id=tenant.id,
app_id=app_model.id,
type="workflow",
version=Workflow.version_from_datetime(naive_utc_now()),
graph=json.dumps(graph),
features=json.dumps({}),
created_by=account.id,
environment_variables=[],
conversation_variables=[],
rag_pipeline_variables=[],
)
db_session_with_containers.add(published_workflow)
app_model.workflow_id = published_workflow.id
db_session_with_containers.commit()
# Create schedule plan
plan = WorkflowSchedulePlan(
app_id=app_model.id,
node_id=schedule_node_id,
tenant_id=tenant.id,
cron_expression="0 9 * * *",
timezone="UTC",
next_run_at=naive_utc_now() - timedelta(minutes=1),
)
app_trigger = AppTrigger(
tenant_id=tenant.id,
app_id=app_model.id,
node_id=schedule_node_id,
trigger_type=AppTriggerType.TRIGGER_SCHEDULE,
status=AppTriggerStatus.ENABLED,
title="schedule",
)
db_session_with_containers.add_all([plan, app_trigger])
db_session_with_containers.commit()
# Mock AsyncWorkflowService to create WorkflowTriggerLog
def _fake_trigger_workflow_async(session: Session, user: Any, trigger_data: Any) -> SimpleNamespace:
log = WorkflowTriggerLog(
tenant_id=trigger_data.tenant_id,
app_id=trigger_data.app_id,
workflow_id=published_workflow.id,
root_node_id=trigger_data.root_node_id,
trigger_metadata="{}",
trigger_type=AppTriggerType.TRIGGER_SCHEDULE,
workflow_run_id=None,
outputs=None,
trigger_data=trigger_data.model_dump_json(),
inputs=json.dumps(dict(trigger_data.inputs)),
status=WorkflowTriggerStatus.SUCCEEDED,
error="",
queue_name="schedule_executor",
celery_task_id="celery-schedule-test",
created_by_role=CreatorUserRole.ACCOUNT,
created_by=account.id,
)
session.add(log)
session.commit()
return SimpleNamespace(workflow_trigger_log_id=log.id, task_id=None, status="queued", queue="test")
monkeypatch.setattr(
workflow_schedule_tasks.AsyncWorkflowService,
"trigger_workflow_async",
_fake_trigger_workflow_async,
)
# Mock quota to avoid rate limiting
from enums import quota_type
monkeypatch.setattr(quota_type.QuotaType.TRIGGER, "consume", lambda _tenant_id: quota_type.unlimited())
# Execute schedule trigger
workflow_schedule_tasks.run_schedule_trigger(plan.id)
# Verify WorkflowTriggerLog was created
db_session_with_containers.expire_all()
logs = db_session_with_containers.query(WorkflowTriggerLog).filter_by(app_id=app_model.id).all()
assert logs, "Schedule trigger should create WorkflowTriggerLog"
assert logs[0].trigger_type == AppTriggerType.TRIGGER_SCHEDULE
assert logs[0].root_node_id == schedule_node_id
@pytest.mark.parametrize(
("mode", "frequency", "visual_config", "cron_expression", "expected_cron"),
[
# Visual mode: hourly
("visual", "hourly", {"on_minute": 30}, None, "30 * * * *"),
# Visual mode: daily
("visual", "daily", {"time": "3:00 PM"}, None, "0 15 * * *"),
# Visual mode: weekly
("visual", "weekly", {"time": "9:00 AM", "weekdays": ["mon", "wed", "fri"]}, None, "0 9 * * 1,3,5"),
# Visual mode: monthly
("visual", "monthly", {"time": "10:30 AM", "monthly_days": [1, 15]}, None, "30 10 1,15 * *"),
# Cron mode: direct expression
("cron", None, None, "*/5 * * * *", "*/5 * * * *"),
],
)
def test_schedule_visual_cron_conversion(
mode: str,
frequency: str | None,
visual_config: dict[str, Any] | None,
cron_expression: str | None,
expected_cron: str,
) -> None:
"""Schedule visual config should correctly convert to cron expression."""
node_config: dict[str, Any] = {
"id": "schedule-node",
"data": {
"type": NodeType.TRIGGER_SCHEDULE.value,
"mode": mode,
"timezone": "UTC",
},
}
if mode == "visual":
node_config["data"]["frequency"] = frequency
node_config["data"]["visual_config"] = visual_config
else:
node_config["data"]["cron_expression"] = cron_expression
config = ScheduleService.to_schedule_config(node_config)
assert config.cron_expression == expected_cron, f"Expected {expected_cron}, got {config.cron_expression}"
assert config.timezone == "UTC"
assert config.node_id == "schedule-node"
def test_plugin_trigger_full_chain_with_db_verification(
test_client_with_containers: FlaskClient,
db_session_with_containers: Session,
tenant_and_account: tuple[Tenant, Account],
app_model: App,
monkeypatch: pytest.MonkeyPatch,
) -> None:
"""Plugin trigger should create WorkflowTriggerLog and WorkflowPluginTrigger records."""
tenant, account = tenant_and_account
# Create published workflow with plugin trigger node
plugin_node_id = "plugin-trigger-node"
provider_id = "langgenius/test-provider/test-provider"
subscription_id = "sub-plugin-test"
endpoint_id = "2cc7fa12-3f7b-4f6a-9c8d-1234567890ab"
graph = {
"nodes": [
{
"id": plugin_node_id,
"data": {
"type": NodeType.TRIGGER_PLUGIN.value,
"title": "plugin",
"plugin_id": "test-plugin",
"plugin_unique_identifier": "test-plugin",
"provider_id": provider_id,
"event_name": "test_event",
"subscription_id": subscription_id,
"parameters": {},
},
},
{"id": "answer-1", "data": {"type": NodeType.ANSWER.value, "title": "answer"}},
],
"edges": [{"source": plugin_node_id, "target": "answer-1", "sourceHandle": "success"}],
}
published_workflow = Workflow.new(
tenant_id=tenant.id,
app_id=app_model.id,
type="workflow",
version=Workflow.version_from_datetime(naive_utc_now()),
graph=json.dumps(graph),
features=json.dumps({}),
created_by=account.id,
environment_variables=[],
conversation_variables=[],
rag_pipeline_variables=[],
)
db_session_with_containers.add(published_workflow)
app_model.workflow_id = published_workflow.id
db_session_with_containers.commit()
# Create trigger subscription
subscription = TriggerSubscription(
name="test-subscription",
tenant_id=tenant.id,
user_id=account.id,
provider_id=provider_id,
endpoint_id=endpoint_id,
parameters={},
properties={},
credentials={"token": "test-secret"},
credential_type="api-key",
)
db_session_with_containers.add(subscription)
db_session_with_containers.commit()
# Update subscription_id to match the created subscription
graph["nodes"][0]["data"]["subscription_id"] = subscription.id
published_workflow.graph = json.dumps(graph)
db_session_with_containers.commit()
# Create WorkflowPluginTrigger
plugin_trigger = WorkflowPluginTrigger(
app_id=app_model.id,
tenant_id=tenant.id,
node_id=plugin_node_id,
provider_id=provider_id,
event_name="test_event",
subscription_id=subscription.id,
)
app_trigger = AppTrigger(
tenant_id=tenant.id,
app_id=app_model.id,
node_id=plugin_node_id,
trigger_type=AppTriggerType.TRIGGER_PLUGIN,
status=AppTriggerStatus.ENABLED,
title="plugin",
)
db_session_with_containers.add_all([plugin_trigger, app_trigger])
db_session_with_containers.commit()
# Track dispatched data
dispatched_data: list[dict[str, Any]] = []
def _fake_process_endpoint(_endpoint_id: str, _request: Any) -> Response:
dispatch_data = {
"user_id": "end-user",
"tenant_id": tenant.id,
"endpoint_id": _endpoint_id,
"provider_id": provider_id,
"subscription_id": subscription.id,
"timestamp": int(time.time()),
"events": ["test_event"],
"request_id": f"req-{_endpoint_id}",
}
dispatched_data.append(dispatch_data)
return Response("ok", status=202)
monkeypatch.setattr(
"services.trigger.trigger_service.TriggerService.process_endpoint",
staticmethod(_fake_process_endpoint),
)
response = test_client_with_containers.post(f"/triggers/plugin/{endpoint_id}", json={"test": "data"})
assert response.status_code == 202
assert dispatched_data, "Plugin trigger should dispatch event data"
assert dispatched_data[0]["subscription_id"] == subscription.id
assert dispatched_data[0]["events"] == ["test_event"]
# Verify database records exist
db_session_with_containers.expire_all()
plugin_triggers = (
db_session_with_containers.query(WorkflowPluginTrigger)
.filter_by(app_id=app_model.id, node_id=plugin_node_id)
.all()
)
assert plugin_triggers, "WorkflowPluginTrigger record should exist"
assert plugin_triggers[0].provider_id == provider_id
assert plugin_triggers[0].event_name == "test_event"
def test_plugin_debug_via_http_endpoint(
test_client_with_containers: FlaskClient,
db_session_with_containers: Session,
tenant_and_account: tuple[Tenant, Account],
app_model: App,
monkeypatch: pytest.MonkeyPatch,
) -> None:
"""Plugin single-step debug via HTTP endpoint should dispatch debug event and be pollable."""
tenant, account = tenant_and_account
provider_id = "langgenius/debug-provider/debug-provider"
endpoint_id = "3cc7fa12-3f7b-4f6a-9c8d-1234567890ab"
event_name = "debug_event"
# Create subscription
subscription = TriggerSubscription(
name="debug-subscription",
tenant_id=tenant.id,
user_id=account.id,
provider_id=provider_id,
endpoint_id=endpoint_id,
parameters={},
properties={},
credentials={"token": "debug-secret"},
credential_type="api-key",
)
db_session_with_containers.add(subscription)
db_session_with_containers.commit()
# Create plugin trigger node config
node_id = "plugin-debug-node"
node_config = {
"id": node_id,
"data": {
"type": NodeType.TRIGGER_PLUGIN.value,
"title": "plugin-debug",
"plugin_id": "debug-plugin",
"plugin_unique_identifier": "debug-plugin",
"provider_id": provider_id,
"event_name": event_name,
"subscription_id": subscription.id,
"parameters": {},
},
}
# Start listening with poller
poller = PluginTriggerDebugEventPoller(
tenant_id=tenant.id,
user_id=account.id,
app_id=app_model.id,
node_config=node_config,
node_id=node_id,
)
assert poller.poll() is None, "First poll should return None (waiting)"
# Track debug events dispatched
debug_events: list[dict[str, Any]] = []
original_dispatch = TriggerDebugEventBus.dispatch
def _tracking_dispatch(**kwargs: Any) -> int:
debug_events.append(kwargs)
return original_dispatch(**kwargs)
monkeypatch.setattr(TriggerDebugEventBus, "dispatch", staticmethod(_tracking_dispatch))
# Mock process_endpoint to trigger debug event dispatch
def _fake_process_endpoint(_endpoint_id: str, _request: Any) -> Response:
# Simulate what happens inside process_endpoint + dispatch_triggered_workflows_async
pool_key = build_plugin_pool_key(
tenant_id=tenant.id,
provider_id=provider_id,
subscription_id=subscription.id,
name=event_name,
)
TriggerDebugEventBus.dispatch(
tenant_id=tenant.id,
event=PluginTriggerDebugEvent(
timestamp=int(time.time()),
user_id="end-user",
name=event_name,
request_id=f"req-{_endpoint_id}",
subscription_id=subscription.id,
provider_id=provider_id,
),
pool_key=pool_key,
)
return Response("ok", status=202)
monkeypatch.setattr(
"services.trigger.trigger_service.TriggerService.process_endpoint",
staticmethod(_fake_process_endpoint),
)
# Call HTTP endpoint
response = test_client_with_containers.post(f"/triggers/plugin/{endpoint_id}", json={"debug": "payload"})
assert response.status_code == 202
assert debug_events, "Debug event should be dispatched via HTTP endpoint"
assert debug_events[0]["event"].name == event_name
# Mock invoke_trigger_event for poller
monkeypatch.setattr(
"services.trigger.trigger_service.TriggerService.invoke_trigger_event",
staticmethod(
lambda **_kwargs: TriggerInvokeEventResponse(
variables={"http_debug": "success"},
cancelled=False,
)
),
)
# Second poll should receive the event
event = poller.poll()
assert event is not None, "Poller should receive debug event after HTTP trigger"
assert event.workflow_args["inputs"]["http_debug"] == "success"