|
|
|
|
@ -15,6 +15,7 @@ import contexts
|
|
|
|
|
from configs import dify_config
|
|
|
|
|
from core.app.entities.app_invoke_entities import InvokeFrom
|
|
|
|
|
from core.datasource.entities.datasource_entities import (
|
|
|
|
|
DatasourceInvokeMessage,
|
|
|
|
|
DatasourceProviderType,
|
|
|
|
|
GetOnlineDocumentPagesResponse,
|
|
|
|
|
GetWebsiteCrawlResponse,
|
|
|
|
|
@ -31,7 +32,7 @@ from core.workflow.entities.workflow_node_execution import (
|
|
|
|
|
)
|
|
|
|
|
from core.workflow.enums import SystemVariableKey
|
|
|
|
|
from core.workflow.errors import WorkflowNodeRunFailedError
|
|
|
|
|
from core.workflow.graph_engine.entities.event import InNodeEvent
|
|
|
|
|
from core.workflow.graph_engine.entities.event import DatasourceRunEvent, InNodeEvent
|
|
|
|
|
from core.workflow.nodes.base.node import BaseNode
|
|
|
|
|
from core.workflow.nodes.enums import ErrorStrategy, NodeType
|
|
|
|
|
from core.workflow.nodes.event.event import RunCompletedEvent
|
|
|
|
|
@ -423,69 +424,11 @@ class RagPipelineService:
|
|
|
|
|
|
|
|
|
|
return workflow_node_execution
|
|
|
|
|
|
|
|
|
|
def run_datasource_workflow_node_status(
|
|
|
|
|
self, pipeline: Pipeline, node_id: str, job_id: str, account: Account, datasource_type: str, is_published: bool
|
|
|
|
|
) -> dict:
|
|
|
|
|
"""
|
|
|
|
|
Run published workflow datasource
|
|
|
|
|
"""
|
|
|
|
|
if is_published:
|
|
|
|
|
# fetch published workflow by app_model
|
|
|
|
|
workflow = self.get_published_workflow(pipeline=pipeline)
|
|
|
|
|
else:
|
|
|
|
|
workflow = self.get_draft_workflow(pipeline=pipeline)
|
|
|
|
|
if not workflow:
|
|
|
|
|
raise ValueError("Workflow not initialized")
|
|
|
|
|
|
|
|
|
|
# run draft workflow node
|
|
|
|
|
datasource_node_data = None
|
|
|
|
|
start_at = time.perf_counter()
|
|
|
|
|
datasource_nodes = workflow.graph_dict.get("nodes", [])
|
|
|
|
|
for datasource_node in datasource_nodes:
|
|
|
|
|
if datasource_node.get("id") == node_id:
|
|
|
|
|
datasource_node_data = datasource_node.get("data", {})
|
|
|
|
|
break
|
|
|
|
|
if not datasource_node_data:
|
|
|
|
|
raise ValueError("Datasource node data not found")
|
|
|
|
|
|
|
|
|
|
from core.datasource.datasource_manager import DatasourceManager
|
|
|
|
|
|
|
|
|
|
datasource_runtime = DatasourceManager.get_datasource_runtime(
|
|
|
|
|
provider_id=f"{datasource_node_data.get('plugin_id')}/{datasource_node_data.get('provider_name')}",
|
|
|
|
|
datasource_name=datasource_node_data.get("datasource_name"),
|
|
|
|
|
tenant_id=pipeline.tenant_id,
|
|
|
|
|
datasource_type=DatasourceProviderType(datasource_type),
|
|
|
|
|
)
|
|
|
|
|
datasource_provider_service = DatasourceProviderService()
|
|
|
|
|
credentials = datasource_provider_service.get_real_datasource_credentials(
|
|
|
|
|
tenant_id=pipeline.tenant_id,
|
|
|
|
|
provider=datasource_node_data.get('provider_name'),
|
|
|
|
|
plugin_id=datasource_node_data.get('plugin_id'),
|
|
|
|
|
)
|
|
|
|
|
if credentials:
|
|
|
|
|
datasource_runtime.runtime.credentials = credentials[0].get("credentials")
|
|
|
|
|
match datasource_type:
|
|
|
|
|
|
|
|
|
|
case DatasourceProviderType.WEBSITE_CRAWL:
|
|
|
|
|
datasource_runtime = cast(WebsiteCrawlDatasourcePlugin, datasource_runtime)
|
|
|
|
|
website_crawl_result: GetWebsiteCrawlResponse = datasource_runtime._get_website_crawl(
|
|
|
|
|
user_id=account.id,
|
|
|
|
|
datasource_parameters={"job_id": job_id},
|
|
|
|
|
provider_type=datasource_runtime.datasource_provider_type(),
|
|
|
|
|
)
|
|
|
|
|
return {
|
|
|
|
|
"result": [result for result in website_crawl_result.result],
|
|
|
|
|
"job_id": website_crawl_result.result.job_id,
|
|
|
|
|
"status": website_crawl_result.result.status,
|
|
|
|
|
"provider_type": datasource_node_data.get("provider_type"),
|
|
|
|
|
}
|
|
|
|
|
case _:
|
|
|
|
|
raise ValueError(f"Unsupported datasource provider: {datasource_runtime.datasource_provider_type}")
|
|
|
|
|
|
|
|
|
|
def run_datasource_workflow_node(
|
|
|
|
|
self, pipeline: Pipeline, node_id: str, user_inputs: dict, account: Account, datasource_type: str,
|
|
|
|
|
is_published: bool
|
|
|
|
|
) -> dict:
|
|
|
|
|
) -> Generator[DatasourceRunEvent, None, None]:
|
|
|
|
|
"""
|
|
|
|
|
Run published workflow datasource
|
|
|
|
|
"""
|
|
|
|
|
@ -532,29 +475,25 @@ class RagPipelineService:
|
|
|
|
|
match datasource_type:
|
|
|
|
|
case DatasourceProviderType.ONLINE_DOCUMENT:
|
|
|
|
|
datasource_runtime = cast(OnlineDocumentDatasourcePlugin, datasource_runtime)
|
|
|
|
|
online_document_result: GetOnlineDocumentPagesResponse = datasource_runtime._get_online_document_pages(
|
|
|
|
|
online_document_result: Generator[DatasourceInvokeMessage, None, None] = datasource_runtime._get_online_document_pages(
|
|
|
|
|
user_id=account.id,
|
|
|
|
|
datasource_parameters=user_inputs,
|
|
|
|
|
provider_type=datasource_runtime.datasource_provider_type(),
|
|
|
|
|
)
|
|
|
|
|
return {
|
|
|
|
|
"result": [page.model_dump() for page in online_document_result.result],
|
|
|
|
|
"provider_type": datasource_node_data.get("provider_type"),
|
|
|
|
|
}
|
|
|
|
|
for message in online_document_result:
|
|
|
|
|
yield DatasourceRunEvent(
|
|
|
|
|
status="success",
|
|
|
|
|
result=message.model_dump(),
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
case DatasourceProviderType.WEBSITE_CRAWL:
|
|
|
|
|
datasource_runtime = cast(WebsiteCrawlDatasourcePlugin, datasource_runtime)
|
|
|
|
|
website_crawl_result: GetWebsiteCrawlResponse = datasource_runtime._get_website_crawl(
|
|
|
|
|
website_crawl_result: Generator[DatasourceInvokeMessage, None, None] = datasource_runtime._get_website_crawl(
|
|
|
|
|
user_id=account.id,
|
|
|
|
|
datasource_parameters=user_inputs,
|
|
|
|
|
provider_type=datasource_runtime.datasource_provider_type(),
|
|
|
|
|
)
|
|
|
|
|
return {
|
|
|
|
|
"result": [result.model_dump() for result in website_crawl_result.result.web_info_list] if website_crawl_result.result.web_info_list else [],
|
|
|
|
|
"job_id": website_crawl_result.result.job_id,
|
|
|
|
|
"status": website_crawl_result.result.status,
|
|
|
|
|
"provider_type": datasource_node_data.get("provider_type"),
|
|
|
|
|
}
|
|
|
|
|
yield from website_crawl_result
|
|
|
|
|
case _:
|
|
|
|
|
raise ValueError(f"Unsupported datasource provider: {datasource_runtime.datasource_provider_type}")
|
|
|
|
|
|
|
|
|
|
|