mirror of
https://github.com/langgenius/dify.git
synced 2026-05-04 08:26:27 +08:00
feat: enhance pipeline settings with execution log and processing capabilities
This commit is contained in:
parent
55516c4e57
commit
335e1e3602
@ -147,9 +147,6 @@ const ChunkPreview = ({
|
|||||||
{!isPending && currentDocForm === ChunkingMode.parentChild && estimateData?.preview && (
|
{!isPending && currentDocForm === ChunkingMode.parentChild && estimateData?.preview && (
|
||||||
estimateData?.preview?.map((item, index) => {
|
estimateData?.preview?.map((item, index) => {
|
||||||
const indexForLabel = index + 1
|
const indexForLabel = index + 1
|
||||||
// const childChunks = parentChildConfig.chunkForContext === 'full-doc'
|
|
||||||
// ? item.child_chunks.slice(0, FULL_DOC_PREVIEW_LENGTH)
|
|
||||||
// : item.child_chunks
|
|
||||||
return (
|
return (
|
||||||
<ChunkContainer
|
<ChunkContainer
|
||||||
key={item.content}
|
key={item.content}
|
||||||
|
|||||||
@ -1,15 +1,19 @@
|
|||||||
import { useCallback, useRef, useState } from 'react'
|
import { useCallback, useMemo, useRef, useState } from 'react'
|
||||||
import type { CrawlResultItem, DocumentItem, FileIndexingEstimateResponse } from '@/models/datasets'
|
import type { CrawlResultItem, CustomFile, FileIndexingEstimateResponse } from '@/models/datasets'
|
||||||
import type { NotionPage } from '@/models/common'
|
import type { NotionPage } from '@/models/common'
|
||||||
import { useTranslation } from 'react-i18next'
|
import { useTranslation } from 'react-i18next'
|
||||||
import { useDatasetDetailContextWithSelector } from '@/context/dataset-detail'
|
|
||||||
import { useDocumentDetail } from '@/service/knowledge/use-document'
|
|
||||||
import AppUnavailable from '@/app/components/base/app-unavailable'
|
import AppUnavailable from '@/app/components/base/app-unavailable'
|
||||||
import ChunkPreview from '../../../create-from-pipeline/preview/chunk-preview'
|
import ChunkPreview from '../../../create-from-pipeline/preview/chunk-preview'
|
||||||
import Loading from '@/app/components/base/loading'
|
import Loading from '@/app/components/base/loading'
|
||||||
import type { DatasourceType } from '@/models/pipeline'
|
|
||||||
import ProcessDocuments from './process-documents'
|
import ProcessDocuments from './process-documents'
|
||||||
import LeftHeader from './left-header'
|
import LeftHeader from './left-header'
|
||||||
|
import { usePipelineExecutionLog, useRunPublishedPipeline } from '@/service/use-pipeline'
|
||||||
|
import type { PublishedPipelineRunPreviewResponse } from '@/models/pipeline'
|
||||||
|
import { DatasourceType } from '@/models/pipeline'
|
||||||
|
import { noop } from 'lodash-es'
|
||||||
|
import { useDatasetDetailContextWithSelector } from '@/context/dataset-detail'
|
||||||
|
import { useRouter } from 'next/navigation'
|
||||||
|
import { useInvalidDocumentList } from '@/service/knowledge/use-document'
|
||||||
|
|
||||||
type PipelineSettingsProps = {
|
type PipelineSettingsProps = {
|
||||||
datasetId: string
|
datasetId: string
|
||||||
@ -21,25 +25,100 @@ const PipelineSettings = ({
|
|||||||
documentId,
|
documentId,
|
||||||
}: PipelineSettingsProps) => {
|
}: PipelineSettingsProps) => {
|
||||||
const { t } = useTranslation()
|
const { t } = useTranslation()
|
||||||
const pipelineId = useDatasetDetailContextWithSelector(s => s.dataset?.pipeline_id)
|
const { push } = useRouter()
|
||||||
const [estimateData, setEstimateData] = useState<FileIndexingEstimateResponse | undefined>(undefined)
|
const [estimateData, setEstimateData] = useState<FileIndexingEstimateResponse | undefined>(undefined)
|
||||||
|
const pipelineId = useDatasetDetailContextWithSelector(state => state.dataset?.pipeline_id)
|
||||||
|
|
||||||
const isPreview = useRef(false)
|
const isPreview = useRef(false)
|
||||||
const formRef = useRef<any>(null)
|
const formRef = useRef<any>(null)
|
||||||
|
|
||||||
const { data: documentDetail, error, isFetching: isFetchingDocumentDetail } = useDocumentDetail({
|
const { data: lastRunData, isFetching: isFetchingLastRunData, isError } = usePipelineExecutionLog({
|
||||||
datasetId,
|
dataset_id: datasetId,
|
||||||
documentId,
|
document_id: documentId,
|
||||||
params: { metadata: 'without' },
|
|
||||||
})
|
})
|
||||||
|
|
||||||
const handlePreviewChunks = useCallback(async (data: Record<string, any>) => {
|
const files = useMemo(() => {
|
||||||
// todo: Preview
|
const files: CustomFile[] = []
|
||||||
}, [])
|
if (lastRunData?.datasource_type === DatasourceType.localFile) {
|
||||||
|
const { related_id, name, extension } = lastRunData.datasource_info
|
||||||
|
files.push({
|
||||||
|
id: related_id,
|
||||||
|
name,
|
||||||
|
extension,
|
||||||
|
} as CustomFile)
|
||||||
|
}
|
||||||
|
return files
|
||||||
|
}, [lastRunData])
|
||||||
|
|
||||||
|
const websitePages = useMemo(() => {
|
||||||
|
const websitePages: CrawlResultItem[] = []
|
||||||
|
if (lastRunData?.datasource_type === DatasourceType.websiteCrawl) {
|
||||||
|
const { content, description, source_url, title } = lastRunData.datasource_info
|
||||||
|
websitePages.push({
|
||||||
|
markdown: content,
|
||||||
|
description,
|
||||||
|
source_url,
|
||||||
|
title,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
return websitePages
|
||||||
|
}, [lastRunData])
|
||||||
|
|
||||||
|
const onlineDocuments = useMemo(() => {
|
||||||
|
const onlineDocuments: NotionPage[] = []
|
||||||
|
if (lastRunData?.datasource_type === DatasourceType.onlineDocument) {
|
||||||
|
const { workspace_id, page } = lastRunData.datasource_info
|
||||||
|
onlineDocuments.push({
|
||||||
|
workspace_id,
|
||||||
|
...page,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
return onlineDocuments
|
||||||
|
}, [lastRunData])
|
||||||
|
|
||||||
|
const { mutateAsync: runPublishedPipeline, isIdle, isPending } = useRunPublishedPipeline()
|
||||||
|
|
||||||
|
const handlePreviewChunks = useCallback(async (data: Record<string, any>) => {
|
||||||
|
if (!lastRunData)
|
||||||
|
return
|
||||||
|
const datasourceInfoList: Record<string, any>[] = []
|
||||||
|
const documentInfo = lastRunData.datasource_info
|
||||||
|
datasourceInfoList.push(documentInfo)
|
||||||
|
await runPublishedPipeline({
|
||||||
|
pipeline_id: pipelineId!,
|
||||||
|
inputs: data,
|
||||||
|
start_node_id: lastRunData.datasource_node_id,
|
||||||
|
datasource_type: lastRunData.datasource_type,
|
||||||
|
datasource_info_list: datasourceInfoList,
|
||||||
|
is_preview: true,
|
||||||
|
}, {
|
||||||
|
onSuccess: (res) => {
|
||||||
|
setEstimateData((res as PublishedPipelineRunPreviewResponse).data.outputs)
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}, [lastRunData, pipelineId, runPublishedPipeline])
|
||||||
|
|
||||||
|
const invalidDocumentList = useInvalidDocumentList(datasetId)
|
||||||
const handleProcess = useCallback(async (data: Record<string, any>) => {
|
const handleProcess = useCallback(async (data: Record<string, any>) => {
|
||||||
// todo: Process
|
if (!lastRunData)
|
||||||
}, [])
|
return
|
||||||
|
const datasourceInfoList: Record<string, any>[] = []
|
||||||
|
const documentInfo = lastRunData.datasource_info
|
||||||
|
datasourceInfoList.push(documentInfo)
|
||||||
|
await runPublishedPipeline({
|
||||||
|
pipeline_id: pipelineId!,
|
||||||
|
inputs: data,
|
||||||
|
start_node_id: lastRunData.datasource_node_id,
|
||||||
|
datasource_type: lastRunData.datasource_type,
|
||||||
|
datasource_info_list: datasourceInfoList,
|
||||||
|
is_preview: false,
|
||||||
|
}, {
|
||||||
|
onSuccess: () => {
|
||||||
|
invalidDocumentList()
|
||||||
|
push(`/datasets/${datasetId}/documents/${documentId}`)
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}, [datasetId, documentId, invalidDocumentList, lastRunData, pipelineId, push, runPublishedPipeline])
|
||||||
|
|
||||||
const onClickProcess = useCallback(() => {
|
const onClickProcess = useCallback(() => {
|
||||||
isPreview.current = false
|
isPreview.current = false
|
||||||
@ -55,25 +134,13 @@ const PipelineSettings = ({
|
|||||||
isPreview.current ? handlePreviewChunks(data) : handleProcess(data)
|
isPreview.current ? handlePreviewChunks(data) : handleProcess(data)
|
||||||
}, [handlePreviewChunks, handleProcess])
|
}, [handlePreviewChunks, handleProcess])
|
||||||
|
|
||||||
const handlePreviewFileChange = useCallback((file: DocumentItem) => {
|
if (isFetchingLastRunData) {
|
||||||
onClickPreview()
|
|
||||||
}, [onClickPreview])
|
|
||||||
|
|
||||||
const handlePreviewOnlineDocumentChange = useCallback((page: NotionPage) => {
|
|
||||||
onClickPreview()
|
|
||||||
}, [onClickPreview])
|
|
||||||
|
|
||||||
const handlePreviewWebsiteChange = useCallback((website: CrawlResultItem) => {
|
|
||||||
onClickPreview()
|
|
||||||
}, [onClickPreview])
|
|
||||||
|
|
||||||
if (isFetchingDocumentDetail) {
|
|
||||||
return (
|
return (
|
||||||
<Loading type='app' />
|
<Loading type='app' />
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (error)
|
if (isError)
|
||||||
return <AppUnavailable code={500} unknownReason={t('datasetCreation.error.unavailable') as string} />
|
return <AppUnavailable code={500} unknownReason={t('datasetCreation.error.unavailable') as string} />
|
||||||
|
|
||||||
return (
|
return (
|
||||||
@ -85,7 +152,8 @@ const PipelineSettings = ({
|
|||||||
<div className='grow overflow-y-auto'>
|
<div className='grow overflow-y-auto'>
|
||||||
<ProcessDocuments
|
<ProcessDocuments
|
||||||
ref={formRef}
|
ref={formRef}
|
||||||
documentId={documentId}
|
lastRunInputData={lastRunData!.input_data}
|
||||||
|
datasourceNodeId={lastRunData!.datasource_node_id}
|
||||||
onProcess={onClickProcess}
|
onProcess={onClickProcess}
|
||||||
onPreview={onClickPreview}
|
onPreview={onClickPreview}
|
||||||
onSubmit={handleSubmit}
|
onSubmit={handleSubmit}
|
||||||
@ -95,22 +163,17 @@ const PipelineSettings = ({
|
|||||||
{/* Preview */}
|
{/* Preview */}
|
||||||
<div className='flex h-full flex-1 pl-2 pt-2'>
|
<div className='flex h-full flex-1 pl-2 pt-2'>
|
||||||
<ChunkPreview
|
<ChunkPreview
|
||||||
dataSourceType={documentDetail!.data_source_type as DatasourceType}
|
dataSourceType={lastRunData!.datasource_type}
|
||||||
// @ts-expect-error mock data // todo: remove mock data
|
files={files}
|
||||||
files={[{
|
onlineDocuments={onlineDocuments}
|
||||||
id: '12345678',
|
websitePages={websitePages}
|
||||||
name: 'test-file',
|
isIdle={isIdle}
|
||||||
extension: 'txt',
|
isPending={isPending && isPreview.current}
|
||||||
}]}
|
|
||||||
onlineDocuments={[]}
|
|
||||||
websitePages={[]}
|
|
||||||
isIdle={true}
|
|
||||||
isPending={true}
|
|
||||||
estimateData={estimateData}
|
estimateData={estimateData}
|
||||||
onPreview={onClickPreview}
|
onPreview={onClickPreview}
|
||||||
handlePreviewFileChange={handlePreviewFileChange}
|
handlePreviewFileChange={noop}
|
||||||
handlePreviewOnlineDocumentChange={handlePreviewOnlineDocumentChange}
|
handlePreviewOnlineDocumentChange={noop}
|
||||||
handlePreviewWebsitePageChange={handlePreviewWebsiteChange}
|
handlePreviewWebsitePageChange={noop}
|
||||||
/>
|
/>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@ -3,11 +3,13 @@ import Button from '@/app/components/base/button'
|
|||||||
import { useTranslation } from 'react-i18next'
|
import { useTranslation } from 'react-i18next'
|
||||||
|
|
||||||
type ActionsProps = {
|
type ActionsProps = {
|
||||||
|
runDisabled?: boolean
|
||||||
onProcess: () => void
|
onProcess: () => void
|
||||||
}
|
}
|
||||||
|
|
||||||
const Actions = ({
|
const Actions = ({
|
||||||
onProcess,
|
onProcess,
|
||||||
|
runDisabled,
|
||||||
}: ActionsProps) => {
|
}: ActionsProps) => {
|
||||||
const { t } = useTranslation()
|
const { t } = useTranslation()
|
||||||
|
|
||||||
@ -16,6 +18,7 @@ const Actions = ({
|
|||||||
<Button
|
<Button
|
||||||
variant='primary'
|
variant='primary'
|
||||||
onClick={onProcess}
|
onClick={onProcess}
|
||||||
|
disabled={runDisabled}
|
||||||
>
|
>
|
||||||
{t('datasetPipeline.operations.saveAndProcess')}
|
{t('datasetPipeline.operations.saveAndProcess')}
|
||||||
</Button>
|
</Button>
|
||||||
|
|||||||
@ -1,10 +1,58 @@
|
|||||||
import type { BaseConfiguration } from '@/app/components/base/form/form-scenarios/base/types'
|
import { useMemo } from 'react'
|
||||||
|
import { useDatasetDetailContextWithSelector } from '@/context/dataset-detail'
|
||||||
|
import { usePublishedPipelineProcessingParams } from '@/service/use-pipeline'
|
||||||
|
import { VAR_TYPE_MAP } from '@/models/pipeline'
|
||||||
|
import { BaseFieldType } from '@/app/components/base/form/form-scenarios/base/types'
|
||||||
|
|
||||||
export const useConfigurations = (documentdId: string) => {
|
export const useConfigurations = (lastRunInputData: Record<string, any>, datasourceNodeId: string) => {
|
||||||
const initialData: Record<string, any> = {}
|
const pipelineId = useDatasetDetailContextWithSelector(state => state.dataset?.pipeline_id)
|
||||||
const configurations: BaseConfiguration[] = []
|
const { data: paramsConfig, isFetching: isFetchingParams } = usePublishedPipelineProcessingParams({
|
||||||
|
pipeline_id: pipelineId!,
|
||||||
|
node_id: datasourceNodeId,
|
||||||
|
})
|
||||||
|
|
||||||
|
const initialData = useMemo(() => {
|
||||||
|
const variables = paramsConfig?.variables || []
|
||||||
|
return variables.reduce((acc, item) => {
|
||||||
|
const type = VAR_TYPE_MAP[item.type]
|
||||||
|
const variableName = item.variable
|
||||||
|
if ([BaseFieldType.textInput, BaseFieldType.paragraph, BaseFieldType.select].includes(type))
|
||||||
|
acc[item.variable] = lastRunInputData[variableName] ?? ''
|
||||||
|
if (type === BaseFieldType.numberInput)
|
||||||
|
acc[item.variable] = lastRunInputData[variableName] ?? 0
|
||||||
|
if (type === BaseFieldType.checkbox)
|
||||||
|
acc[item.variable] = lastRunInputData[variableName]
|
||||||
|
if ([BaseFieldType.file, BaseFieldType.fileList].includes(type))
|
||||||
|
acc[item.variable] = lastRunInputData[variableName]
|
||||||
|
return acc
|
||||||
|
}, {} as Record<string, any>)
|
||||||
|
}, [lastRunInputData, paramsConfig?.variables])
|
||||||
|
|
||||||
|
const configurations = useMemo(() => {
|
||||||
|
const variables = paramsConfig?.variables || []
|
||||||
|
const configs = variables.map(item => ({
|
||||||
|
type: VAR_TYPE_MAP[item.type],
|
||||||
|
variable: item.variable,
|
||||||
|
label: item.label,
|
||||||
|
required: item.required,
|
||||||
|
maxLength: item.max_length,
|
||||||
|
options: item.options?.map(option => ({
|
||||||
|
label: option,
|
||||||
|
value: option,
|
||||||
|
})),
|
||||||
|
showConditions: [],
|
||||||
|
placeholder: item.placeholder,
|
||||||
|
tooltip: item.tooltips,
|
||||||
|
unit: item.unit,
|
||||||
|
allowedFileTypes: item.allowed_file_types,
|
||||||
|
allowedFileExtensions: item.allowed_file_extensions,
|
||||||
|
allowedFileUploadMethods: item.allowed_file_upload_methods,
|
||||||
|
}))
|
||||||
|
return configs
|
||||||
|
}, [paramsConfig])
|
||||||
|
|
||||||
return {
|
return {
|
||||||
|
isFetchingParams,
|
||||||
initialData,
|
initialData,
|
||||||
configurations,
|
configurations,
|
||||||
}
|
}
|
||||||
|
|||||||
@ -4,7 +4,8 @@ import Actions from './actions'
|
|||||||
import Form from '../../../../create-from-pipeline/process-documents/form'
|
import Form from '../../../../create-from-pipeline/process-documents/form'
|
||||||
|
|
||||||
type ProcessDocumentsProps = {
|
type ProcessDocumentsProps = {
|
||||||
documentId: string
|
datasourceNodeId: string
|
||||||
|
lastRunInputData: Record<string, any>
|
||||||
ref: React.RefObject<any>
|
ref: React.RefObject<any>
|
||||||
onProcess: () => void
|
onProcess: () => void
|
||||||
onPreview: () => void
|
onPreview: () => void
|
||||||
@ -12,13 +13,14 @@ type ProcessDocumentsProps = {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const ProcessDocuments = ({
|
const ProcessDocuments = ({
|
||||||
documentId,
|
datasourceNodeId,
|
||||||
|
lastRunInputData,
|
||||||
onProcess,
|
onProcess,
|
||||||
onPreview,
|
onPreview,
|
||||||
onSubmit,
|
onSubmit,
|
||||||
ref,
|
ref,
|
||||||
}: ProcessDocumentsProps) => {
|
}: ProcessDocumentsProps) => {
|
||||||
const { initialData, configurations } = useConfigurations(documentId)
|
const { isFetchingParams, initialData, configurations } = useConfigurations(lastRunInputData, datasourceNodeId)
|
||||||
const schema = generateZodSchema(configurations)
|
const schema = generateZodSchema(configurations)
|
||||||
|
|
||||||
return (
|
return (
|
||||||
@ -31,7 +33,7 @@ const ProcessDocuments = ({
|
|||||||
onSubmit={onSubmit}
|
onSubmit={onSubmit}
|
||||||
onPreview={onPreview}
|
onPreview={onPreview}
|
||||||
/>
|
/>
|
||||||
<Actions onProcess={onProcess} />
|
<Actions runDisabled={isFetchingParams} onProcess={onProcess} />
|
||||||
</div>
|
</div>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|||||||
@ -242,3 +242,15 @@ export type InitialDocumentDetail = {
|
|||||||
name: string
|
name: string
|
||||||
position: number
|
position: number
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export type PipelineExecutionLogRequest = {
|
||||||
|
dataset_id: string
|
||||||
|
document_id: string
|
||||||
|
}
|
||||||
|
|
||||||
|
export type PipelineExecutionLogResponse = {
|
||||||
|
datasource_info: Record<string, any>
|
||||||
|
datasource_type: DatasourceType
|
||||||
|
input_data: Record<string, any>
|
||||||
|
datasource_node_id: string
|
||||||
|
}
|
||||||
|
|||||||
@ -8,6 +8,8 @@ import type {
|
|||||||
ImportPipelineDSLRequest,
|
ImportPipelineDSLRequest,
|
||||||
ImportPipelineDSLResponse,
|
ImportPipelineDSLResponse,
|
||||||
PipelineCheckDependenciesResponse,
|
PipelineCheckDependenciesResponse,
|
||||||
|
PipelineExecutionLogRequest,
|
||||||
|
PipelineExecutionLogResponse,
|
||||||
PipelinePreProcessingParamsRequest,
|
PipelinePreProcessingParamsRequest,
|
||||||
PipelinePreProcessingParamsResponse,
|
PipelinePreProcessingParamsResponse,
|
||||||
PipelineProcessingParamsRequest,
|
PipelineProcessingParamsRequest,
|
||||||
@ -311,3 +313,14 @@ export const usePublishAsCustomizedPipeline = () => {
|
|||||||
},
|
},
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export const usePipelineExecutionLog = (params: PipelineExecutionLogRequest) => {
|
||||||
|
const { dataset_id, document_id } = params
|
||||||
|
return useQuery<PipelineExecutionLogResponse>({
|
||||||
|
queryKey: [NAME_SPACE, 'pipeline-execution-log', dataset_id, document_id],
|
||||||
|
queryFn: () => {
|
||||||
|
return get<PipelineExecutionLogResponse>(`/datasets/${dataset_id}/documents/${document_id}/pipeline-execution-log`)
|
||||||
|
},
|
||||||
|
staleTime: 0,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user