mirror of
https://github.com/simstudioai/sim.git
synced 2026-02-12 07:24:55 -05:00
Compare commits
2 Commits
sim-609-to
...
fix/confl
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
2ba8bd37bd | ||
|
|
102d24eb14 |
@@ -29,7 +29,7 @@ const patchBodySchema = z
|
|||||||
description: z
|
description: z
|
||||||
.string()
|
.string()
|
||||||
.trim()
|
.trim()
|
||||||
.max(2000, 'Description must be 2000 characters or less')
|
.max(500, 'Description must be 500 characters or less')
|
||||||
.nullable()
|
.nullable()
|
||||||
.optional(),
|
.optional(),
|
||||||
isActive: z.literal(true).optional(), // Set to true to activate this version
|
isActive: z.literal(true).optional(), // Set to true to activate this version
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ import {
|
|||||||
import { generateRequestId } from '@/lib/core/utils/request'
|
import { generateRequestId } from '@/lib/core/utils/request'
|
||||||
import { SSE_HEADERS } from '@/lib/core/utils/sse'
|
import { SSE_HEADERS } from '@/lib/core/utils/sse'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { createExecutionEventWriter, setExecutionMeta } from '@/lib/execution/event-buffer'
|
import { markExecutionCancelled } from '@/lib/execution/cancellation'
|
||||||
import { processInputFileFields } from '@/lib/execution/files'
|
import { processInputFileFields } from '@/lib/execution/files'
|
||||||
import { preprocessExecution } from '@/lib/execution/preprocessing'
|
import { preprocessExecution } from '@/lib/execution/preprocessing'
|
||||||
import { LoggingSession } from '@/lib/logs/execution/logging-session'
|
import { LoggingSession } from '@/lib/logs/execution/logging-session'
|
||||||
@@ -700,27 +700,15 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
const timeoutController = createTimeoutAbortController(preprocessResult.executionTimeout?.sync)
|
const timeoutController = createTimeoutAbortController(preprocessResult.executionTimeout?.sync)
|
||||||
let isStreamClosed = false
|
let isStreamClosed = false
|
||||||
|
|
||||||
const eventWriter = createExecutionEventWriter(executionId)
|
|
||||||
setExecutionMeta(executionId, {
|
|
||||||
status: 'active',
|
|
||||||
userId: actorUserId,
|
|
||||||
workflowId,
|
|
||||||
}).catch(() => {})
|
|
||||||
|
|
||||||
const stream = new ReadableStream<Uint8Array>({
|
const stream = new ReadableStream<Uint8Array>({
|
||||||
async start(controller) {
|
async start(controller) {
|
||||||
let finalMetaStatus: 'complete' | 'error' | 'cancelled' | null = null
|
|
||||||
|
|
||||||
const sendEvent = (event: ExecutionEvent) => {
|
const sendEvent = (event: ExecutionEvent) => {
|
||||||
if (!isStreamClosed) {
|
if (isStreamClosed) return
|
||||||
try {
|
|
||||||
controller.enqueue(encodeSSEEvent(event))
|
try {
|
||||||
} catch {
|
controller.enqueue(encodeSSEEvent(event))
|
||||||
isStreamClosed = true
|
} catch {
|
||||||
}
|
isStreamClosed = true
|
||||||
}
|
|
||||||
if (event.type !== 'stream:chunk' && event.type !== 'stream:done') {
|
|
||||||
eventWriter.write(event).catch(() => {})
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -841,12 +829,14 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
|
|
||||||
const reader = streamingExec.stream.getReader()
|
const reader = streamingExec.stream.getReader()
|
||||||
const decoder = new TextDecoder()
|
const decoder = new TextDecoder()
|
||||||
|
let chunkCount = 0
|
||||||
|
|
||||||
try {
|
try {
|
||||||
while (true) {
|
while (true) {
|
||||||
const { done, value } = await reader.read()
|
const { done, value } = await reader.read()
|
||||||
if (done) break
|
if (done) break
|
||||||
|
|
||||||
|
chunkCount++
|
||||||
const chunk = decoder.decode(value, { stream: true })
|
const chunk = decoder.decode(value, { stream: true })
|
||||||
sendEvent({
|
sendEvent({
|
||||||
type: 'stream:chunk',
|
type: 'stream:chunk',
|
||||||
@@ -961,7 +951,6 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
duration: result.metadata?.duration || 0,
|
duration: result.metadata?.duration || 0,
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
finalMetaStatus = 'error'
|
|
||||||
} else {
|
} else {
|
||||||
logger.info(`[${requestId}] Workflow execution was cancelled`)
|
logger.info(`[${requestId}] Workflow execution was cancelled`)
|
||||||
|
|
||||||
@@ -974,7 +963,6 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
duration: result.metadata?.duration || 0,
|
duration: result.metadata?.duration || 0,
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
finalMetaStatus = 'cancelled'
|
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
@@ -998,7 +986,6 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
endTime: result.metadata?.endTime || new Date().toISOString(),
|
endTime: result.metadata?.endTime || new Date().toISOString(),
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
finalMetaStatus = 'complete'
|
|
||||||
} catch (error: unknown) {
|
} catch (error: unknown) {
|
||||||
const isTimeout = isTimeoutError(error) || timeoutController.isTimedOut()
|
const isTimeout = isTimeoutError(error) || timeoutController.isTimedOut()
|
||||||
const errorMessage = isTimeout
|
const errorMessage = isTimeout
|
||||||
@@ -1030,18 +1017,7 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
duration: executionResult?.metadata?.duration || 0,
|
duration: executionResult?.metadata?.duration || 0,
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
finalMetaStatus = 'error'
|
|
||||||
} finally {
|
} finally {
|
||||||
try {
|
|
||||||
await eventWriter.close()
|
|
||||||
} catch (closeError) {
|
|
||||||
logger.warn(`[${requestId}] Failed to close event writer`, {
|
|
||||||
error: closeError instanceof Error ? closeError.message : String(closeError),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
if (finalMetaStatus) {
|
|
||||||
setExecutionMeta(executionId, { status: finalMetaStatus }).catch(() => {})
|
|
||||||
}
|
|
||||||
timeoutController.cleanup()
|
timeoutController.cleanup()
|
||||||
if (executionId) {
|
if (executionId) {
|
||||||
await cleanupExecutionBase64Cache(executionId)
|
await cleanupExecutionBase64Cache(executionId)
|
||||||
@@ -1056,7 +1032,10 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
},
|
},
|
||||||
cancel() {
|
cancel() {
|
||||||
isStreamClosed = true
|
isStreamClosed = true
|
||||||
logger.info(`[${requestId}] Client disconnected from SSE stream`)
|
timeoutController.cleanup()
|
||||||
|
logger.info(`[${requestId}] Client aborted SSE stream, signalling cancellation`)
|
||||||
|
timeoutController.abort()
|
||||||
|
markExecutionCancelled(executionId).catch(() => {})
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|||||||
@@ -1,170 +0,0 @@
|
|||||||
import { createLogger } from '@sim/logger'
|
|
||||||
import { type NextRequest, NextResponse } from 'next/server'
|
|
||||||
import { checkHybridAuth } from '@/lib/auth/hybrid'
|
|
||||||
import { SSE_HEADERS } from '@/lib/core/utils/sse'
|
|
||||||
import {
|
|
||||||
type ExecutionStreamStatus,
|
|
||||||
getExecutionMeta,
|
|
||||||
readExecutionEvents,
|
|
||||||
} from '@/lib/execution/event-buffer'
|
|
||||||
import { formatSSEEvent } from '@/lib/workflows/executor/execution-events'
|
|
||||||
import { authorizeWorkflowByWorkspacePermission } from '@/lib/workflows/utils'
|
|
||||||
|
|
||||||
const logger = createLogger('ExecutionStreamReconnectAPI')
|
|
||||||
|
|
||||||
const POLL_INTERVAL_MS = 500
|
|
||||||
const MAX_POLL_DURATION_MS = 10 * 60 * 1000 // 10 minutes
|
|
||||||
|
|
||||||
function isTerminalStatus(status: ExecutionStreamStatus): boolean {
|
|
||||||
return status === 'complete' || status === 'error' || status === 'cancelled'
|
|
||||||
}
|
|
||||||
|
|
||||||
export const runtime = 'nodejs'
|
|
||||||
export const dynamic = 'force-dynamic'
|
|
||||||
|
|
||||||
export async function GET(
|
|
||||||
req: NextRequest,
|
|
||||||
{ params }: { params: Promise<{ id: string; executionId: string }> }
|
|
||||||
) {
|
|
||||||
const { id: workflowId, executionId } = await params
|
|
||||||
|
|
||||||
try {
|
|
||||||
const auth = await checkHybridAuth(req, { requireWorkflowId: false })
|
|
||||||
if (!auth.success || !auth.userId) {
|
|
||||||
return NextResponse.json({ error: auth.error || 'Unauthorized' }, { status: 401 })
|
|
||||||
}
|
|
||||||
|
|
||||||
const workflowAuthorization = await authorizeWorkflowByWorkspacePermission({
|
|
||||||
workflowId,
|
|
||||||
userId: auth.userId,
|
|
||||||
action: 'read',
|
|
||||||
})
|
|
||||||
if (!workflowAuthorization.allowed) {
|
|
||||||
return NextResponse.json(
|
|
||||||
{ error: workflowAuthorization.message || 'Access denied' },
|
|
||||||
{ status: workflowAuthorization.status }
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
const meta = await getExecutionMeta(executionId)
|
|
||||||
if (!meta) {
|
|
||||||
return NextResponse.json({ error: 'Execution buffer not found or expired' }, { status: 404 })
|
|
||||||
}
|
|
||||||
|
|
||||||
if (meta.workflowId && meta.workflowId !== workflowId) {
|
|
||||||
return NextResponse.json(
|
|
||||||
{ error: 'Execution does not belong to this workflow' },
|
|
||||||
{ status: 403 }
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
const fromParam = req.nextUrl.searchParams.get('from')
|
|
||||||
const parsed = fromParam ? Number.parseInt(fromParam, 10) : 0
|
|
||||||
const fromEventId = Number.isFinite(parsed) && parsed >= 0 ? parsed : 0
|
|
||||||
|
|
||||||
logger.info('Reconnection stream requested', {
|
|
||||||
workflowId,
|
|
||||||
executionId,
|
|
||||||
fromEventId,
|
|
||||||
metaStatus: meta.status,
|
|
||||||
})
|
|
||||||
|
|
||||||
const encoder = new TextEncoder()
|
|
||||||
|
|
||||||
let closed = false
|
|
||||||
|
|
||||||
const stream = new ReadableStream<Uint8Array>({
|
|
||||||
async start(controller) {
|
|
||||||
let lastEventId = fromEventId
|
|
||||||
const pollDeadline = Date.now() + MAX_POLL_DURATION_MS
|
|
||||||
|
|
||||||
const enqueue = (text: string) => {
|
|
||||||
if (closed) return
|
|
||||||
try {
|
|
||||||
controller.enqueue(encoder.encode(text))
|
|
||||||
} catch {
|
|
||||||
closed = true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
try {
|
|
||||||
const events = await readExecutionEvents(executionId, lastEventId)
|
|
||||||
for (const entry of events) {
|
|
||||||
if (closed) return
|
|
||||||
enqueue(formatSSEEvent(entry.event))
|
|
||||||
lastEventId = entry.eventId
|
|
||||||
}
|
|
||||||
|
|
||||||
const currentMeta = await getExecutionMeta(executionId)
|
|
||||||
if (!currentMeta || isTerminalStatus(currentMeta.status)) {
|
|
||||||
enqueue('data: [DONE]\n\n')
|
|
||||||
if (!closed) controller.close()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
while (!closed && Date.now() < pollDeadline) {
|
|
||||||
await new Promise((resolve) => setTimeout(resolve, POLL_INTERVAL_MS))
|
|
||||||
if (closed) return
|
|
||||||
|
|
||||||
const newEvents = await readExecutionEvents(executionId, lastEventId)
|
|
||||||
for (const entry of newEvents) {
|
|
||||||
if (closed) return
|
|
||||||
enqueue(formatSSEEvent(entry.event))
|
|
||||||
lastEventId = entry.eventId
|
|
||||||
}
|
|
||||||
|
|
||||||
const polledMeta = await getExecutionMeta(executionId)
|
|
||||||
if (!polledMeta || isTerminalStatus(polledMeta.status)) {
|
|
||||||
const finalEvents = await readExecutionEvents(executionId, lastEventId)
|
|
||||||
for (const entry of finalEvents) {
|
|
||||||
if (closed) return
|
|
||||||
enqueue(formatSSEEvent(entry.event))
|
|
||||||
lastEventId = entry.eventId
|
|
||||||
}
|
|
||||||
enqueue('data: [DONE]\n\n')
|
|
||||||
if (!closed) controller.close()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!closed) {
|
|
||||||
logger.warn('Reconnection stream poll deadline reached', { executionId })
|
|
||||||
enqueue('data: [DONE]\n\n')
|
|
||||||
controller.close()
|
|
||||||
}
|
|
||||||
} catch (error) {
|
|
||||||
logger.error('Error in reconnection stream', {
|
|
||||||
executionId,
|
|
||||||
error: error instanceof Error ? error.message : String(error),
|
|
||||||
})
|
|
||||||
if (!closed) {
|
|
||||||
try {
|
|
||||||
controller.close()
|
|
||||||
} catch {}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
cancel() {
|
|
||||||
closed = true
|
|
||||||
logger.info('Client disconnected from reconnection stream', { executionId })
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
return new NextResponse(stream, {
|
|
||||||
headers: {
|
|
||||||
...SSE_HEADERS,
|
|
||||||
'X-Execution-Id': executionId,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
} catch (error: any) {
|
|
||||||
logger.error('Failed to start reconnection stream', {
|
|
||||||
workflowId,
|
|
||||||
executionId,
|
|
||||||
error: error.message,
|
|
||||||
})
|
|
||||||
return NextResponse.json(
|
|
||||||
{ error: error.message || 'Failed to start reconnection stream' },
|
|
||||||
{ status: 500 }
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -113,7 +113,7 @@ export function VersionDescriptionModal({
|
|||||||
className='min-h-[120px] resize-none'
|
className='min-h-[120px] resize-none'
|
||||||
value={description}
|
value={description}
|
||||||
onChange={(e) => setDescription(e.target.value)}
|
onChange={(e) => setDescription(e.target.value)}
|
||||||
maxLength={2000}
|
maxLength={500}
|
||||||
disabled={isGenerating}
|
disabled={isGenerating}
|
||||||
/>
|
/>
|
||||||
<div className='flex items-center justify-between'>
|
<div className='flex items-center justify-between'>
|
||||||
@@ -123,7 +123,7 @@ export function VersionDescriptionModal({
|
|||||||
</p>
|
</p>
|
||||||
)}
|
)}
|
||||||
{!updateMutation.error && !generateMutation.error && <div />}
|
{!updateMutation.error && !generateMutation.error && <div />}
|
||||||
<p className='text-[11px] text-[var(--text-tertiary)]'>{description.length}/2000</p>
|
<p className='text-[11px] text-[var(--text-tertiary)]'>{description.length}/500</p>
|
||||||
</div>
|
</div>
|
||||||
</ModalBody>
|
</ModalBody>
|
||||||
<ModalFooter>
|
<ModalFooter>
|
||||||
|
|||||||
@@ -57,21 +57,6 @@ export function useChangeDetection({
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (block.triggerMode) {
|
|
||||||
const triggerConfigValue = blockSubValues?.triggerConfig
|
|
||||||
if (
|
|
||||||
triggerConfigValue &&
|
|
||||||
typeof triggerConfigValue === 'object' &&
|
|
||||||
!subBlocks.triggerConfig
|
|
||||||
) {
|
|
||||||
subBlocks.triggerConfig = {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: triggerConfigValue,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
blocksWithSubBlocks[blockId] = {
|
blocksWithSubBlocks[blockId] = {
|
||||||
...block,
|
...block,
|
||||||
subBlocks,
|
subBlocks,
|
||||||
|
|||||||
@@ -1,189 +0,0 @@
|
|||||||
'use client'
|
|
||||||
|
|
||||||
import type React from 'react'
|
|
||||||
import { useRef, useState } from 'react'
|
|
||||||
import { ArrowLeftRight, ArrowUp } from 'lucide-react'
|
|
||||||
import { Button, Input, Label, Tooltip } from '@/components/emcn'
|
|
||||||
import { cn } from '@/lib/core/utils/cn'
|
|
||||||
import type { WandControlHandlers } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/sub-block'
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Props for a generic parameter with label component
|
|
||||||
*/
|
|
||||||
export interface ParameterWithLabelProps {
|
|
||||||
paramId: string
|
|
||||||
title: string
|
|
||||||
isRequired: boolean
|
|
||||||
visibility: string
|
|
||||||
wandConfig?: {
|
|
||||||
enabled: boolean
|
|
||||||
prompt?: string
|
|
||||||
placeholder?: string
|
|
||||||
}
|
|
||||||
canonicalToggle?: {
|
|
||||||
mode: 'basic' | 'advanced'
|
|
||||||
disabled?: boolean
|
|
||||||
onToggle?: () => void
|
|
||||||
}
|
|
||||||
disabled: boolean
|
|
||||||
isPreview: boolean
|
|
||||||
children: (wandControlRef: React.MutableRefObject<WandControlHandlers | null>) => React.ReactNode
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Generic wrapper component for parameters that manages wand state and renders label + input
|
|
||||||
*/
|
|
||||||
export function ParameterWithLabel({
|
|
||||||
paramId,
|
|
||||||
title,
|
|
||||||
isRequired,
|
|
||||||
visibility,
|
|
||||||
wandConfig,
|
|
||||||
canonicalToggle,
|
|
||||||
disabled,
|
|
||||||
isPreview,
|
|
||||||
children,
|
|
||||||
}: ParameterWithLabelProps) {
|
|
||||||
const [isSearchActive, setIsSearchActive] = useState(false)
|
|
||||||
const [searchQuery, setSearchQuery] = useState('')
|
|
||||||
const searchInputRef = useRef<HTMLInputElement>(null)
|
|
||||||
const wandControlRef = useRef<WandControlHandlers | null>(null)
|
|
||||||
|
|
||||||
const isWandEnabled = wandConfig?.enabled ?? false
|
|
||||||
const showWand = isWandEnabled && !isPreview && !disabled
|
|
||||||
|
|
||||||
const handleSearchClick = (): void => {
|
|
||||||
setIsSearchActive(true)
|
|
||||||
setTimeout(() => {
|
|
||||||
searchInputRef.current?.focus()
|
|
||||||
}, 0)
|
|
||||||
}
|
|
||||||
|
|
||||||
const handleSearchBlur = (): void => {
|
|
||||||
if (!searchQuery.trim() && !wandControlRef.current?.isWandStreaming) {
|
|
||||||
setIsSearchActive(false)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const handleSearchChange = (value: string): void => {
|
|
||||||
setSearchQuery(value)
|
|
||||||
}
|
|
||||||
|
|
||||||
const handleSearchSubmit = (): void => {
|
|
||||||
if (searchQuery.trim() && wandControlRef.current) {
|
|
||||||
wandControlRef.current.onWandTrigger(searchQuery)
|
|
||||||
setSearchQuery('')
|
|
||||||
setIsSearchActive(false)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const handleSearchCancel = (): void => {
|
|
||||||
setSearchQuery('')
|
|
||||||
setIsSearchActive(false)
|
|
||||||
}
|
|
||||||
|
|
||||||
const isStreaming = wandControlRef.current?.isWandStreaming ?? false
|
|
||||||
|
|
||||||
return (
|
|
||||||
<div key={paramId} className='relative min-w-0 space-y-[6px]'>
|
|
||||||
<div className='flex items-center justify-between gap-[6px] pl-[2px]'>
|
|
||||||
<Label className='flex items-center gap-[6px] whitespace-nowrap font-medium text-[13px] text-[var(--text-primary)]'>
|
|
||||||
{title}
|
|
||||||
{isRequired && visibility === 'user-only' && <span className='ml-0.5'>*</span>}
|
|
||||||
{visibility !== 'user-only' && (
|
|
||||||
<span className='ml-[6px] text-[12px] text-[var(--text-tertiary)]'>(optional)</span>
|
|
||||||
)}
|
|
||||||
</Label>
|
|
||||||
<div className='flex min-w-0 flex-1 items-center justify-end gap-[6px]'>
|
|
||||||
{showWand &&
|
|
||||||
(!isSearchActive ? (
|
|
||||||
<Button
|
|
||||||
variant='active'
|
|
||||||
className='-my-1 h-5 px-2 py-0 text-[11px]'
|
|
||||||
onClick={handleSearchClick}
|
|
||||||
>
|
|
||||||
Generate
|
|
||||||
</Button>
|
|
||||||
) : (
|
|
||||||
<div className='-my-1 flex min-w-[120px] max-w-[280px] flex-1 items-center gap-[4px]'>
|
|
||||||
<Input
|
|
||||||
ref={searchInputRef}
|
|
||||||
value={isStreaming ? 'Generating...' : searchQuery}
|
|
||||||
onChange={(e: React.ChangeEvent<HTMLInputElement>) =>
|
|
||||||
handleSearchChange(e.target.value)
|
|
||||||
}
|
|
||||||
onBlur={(e: React.FocusEvent<HTMLInputElement>) => {
|
|
||||||
const relatedTarget = e.relatedTarget as HTMLElement | null
|
|
||||||
if (relatedTarget?.closest('button')) return
|
|
||||||
handleSearchBlur()
|
|
||||||
}}
|
|
||||||
onKeyDown={(e: React.KeyboardEvent<HTMLInputElement>) => {
|
|
||||||
if (e.key === 'Enter' && searchQuery.trim() && !isStreaming) {
|
|
||||||
handleSearchSubmit()
|
|
||||||
} else if (e.key === 'Escape') {
|
|
||||||
handleSearchCancel()
|
|
||||||
}
|
|
||||||
}}
|
|
||||||
disabled={isStreaming}
|
|
||||||
className={cn(
|
|
||||||
'h-5 min-w-[80px] flex-1 text-[11px]',
|
|
||||||
isStreaming && 'text-muted-foreground'
|
|
||||||
)}
|
|
||||||
placeholder='Generate with AI...'
|
|
||||||
/>
|
|
||||||
<Button
|
|
||||||
variant='tertiary'
|
|
||||||
disabled={!searchQuery.trim() || isStreaming}
|
|
||||||
onMouseDown={(e: React.MouseEvent) => {
|
|
||||||
e.preventDefault()
|
|
||||||
e.stopPropagation()
|
|
||||||
}}
|
|
||||||
onClick={(e: React.MouseEvent) => {
|
|
||||||
e.stopPropagation()
|
|
||||||
handleSearchSubmit()
|
|
||||||
}}
|
|
||||||
className='h-[20px] w-[20px] flex-shrink-0 p-0'
|
|
||||||
>
|
|
||||||
<ArrowUp className='h-[12px] w-[12px]' />
|
|
||||||
</Button>
|
|
||||||
</div>
|
|
||||||
))}
|
|
||||||
{canonicalToggle && !isPreview && (
|
|
||||||
<Tooltip.Root>
|
|
||||||
<Tooltip.Trigger asChild>
|
|
||||||
<button
|
|
||||||
type='button'
|
|
||||||
className='flex h-[12px] w-[12px] flex-shrink-0 items-center justify-center bg-transparent p-0 disabled:cursor-not-allowed disabled:opacity-50'
|
|
||||||
onClick={canonicalToggle.onToggle}
|
|
||||||
disabled={canonicalToggle.disabled || disabled}
|
|
||||||
aria-label={
|
|
||||||
canonicalToggle.mode === 'advanced'
|
|
||||||
? 'Switch to selector'
|
|
||||||
: 'Switch to manual ID'
|
|
||||||
}
|
|
||||||
>
|
|
||||||
<ArrowLeftRight
|
|
||||||
className={cn(
|
|
||||||
'!h-[12px] !w-[12px]',
|
|
||||||
canonicalToggle.mode === 'advanced'
|
|
||||||
? 'text-[var(--text-primary)]'
|
|
||||||
: 'text-[var(--text-secondary)]'
|
|
||||||
)}
|
|
||||||
/>
|
|
||||||
</button>
|
|
||||||
</Tooltip.Trigger>
|
|
||||||
<Tooltip.Content side='top'>
|
|
||||||
<p>
|
|
||||||
{canonicalToggle.mode === 'advanced'
|
|
||||||
? 'Switch to selector'
|
|
||||||
: 'Switch to manual ID'}
|
|
||||||
</p>
|
|
||||||
</Tooltip.Content>
|
|
||||||
</Tooltip.Root>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
<div className='relative w-full min-w-0'>{children(wandControlRef)}</div>
|
|
||||||
</div>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
@@ -1,109 +0,0 @@
|
|||||||
'use client'
|
|
||||||
|
|
||||||
import { useEffect, useMemo, useRef } from 'react'
|
|
||||||
import { useSubBlockValue } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/hooks/use-sub-block-value'
|
|
||||||
import { SubBlock } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/sub-block'
|
|
||||||
import type { SubBlockConfig as BlockSubBlockConfig } from '@/blocks/types'
|
|
||||||
|
|
||||||
interface ToolSubBlockRendererProps {
|
|
||||||
blockId: string
|
|
||||||
subBlockId: string
|
|
||||||
toolIndex: number
|
|
||||||
subBlock: BlockSubBlockConfig
|
|
||||||
effectiveParamId: string
|
|
||||||
toolParams: Record<string, string> | undefined
|
|
||||||
onParamChange: (toolIndex: number, paramId: string, value: string) => void
|
|
||||||
disabled: boolean
|
|
||||||
canonicalToggle?: {
|
|
||||||
mode: 'basic' | 'advanced'
|
|
||||||
disabled?: boolean
|
|
||||||
onToggle?: () => void
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Bridges the subblock store with StoredTool.params via a synthetic store key,
|
|
||||||
* then delegates all rendering to SubBlock for full parity.
|
|
||||||
*
|
|
||||||
* Two effects handle bidirectional sync:
|
|
||||||
* - tool.params → store (external changes)
|
|
||||||
* - store → tool.params (user interaction)
|
|
||||||
*/
|
|
||||||
export function ToolSubBlockRenderer({
|
|
||||||
blockId,
|
|
||||||
subBlockId,
|
|
||||||
toolIndex,
|
|
||||||
subBlock,
|
|
||||||
effectiveParamId,
|
|
||||||
toolParams,
|
|
||||||
onParamChange,
|
|
||||||
disabled,
|
|
||||||
canonicalToggle,
|
|
||||||
}: ToolSubBlockRendererProps) {
|
|
||||||
const syntheticId = `${subBlockId}-tool-${toolIndex}-${effectiveParamId}`
|
|
||||||
const [storeValue, setStoreValue] = useSubBlockValue(blockId, syntheticId)
|
|
||||||
|
|
||||||
const toolParamValue = toolParams?.[effectiveParamId] ?? ''
|
|
||||||
|
|
||||||
/** Tracks the last value we pushed to the store from tool.params to avoid echo loops */
|
|
||||||
const lastPushedToStoreRef = useRef<string | null>(null)
|
|
||||||
/** Tracks the last value we synced back to tool.params from the store */
|
|
||||||
const lastPushedToParamsRef = useRef<string | null>(null)
|
|
||||||
|
|
||||||
// Sync tool.params → store: push when the prop value changes (including first mount)
|
|
||||||
useEffect(() => {
|
|
||||||
if (!toolParamValue && lastPushedToStoreRef.current === null) {
|
|
||||||
// Skip initializing the store with an empty value on first mount —
|
|
||||||
// let the SubBlock component use its own default.
|
|
||||||
lastPushedToStoreRef.current = toolParamValue
|
|
||||||
lastPushedToParamsRef.current = toolParamValue
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if (toolParamValue !== lastPushedToStoreRef.current) {
|
|
||||||
lastPushedToStoreRef.current = toolParamValue
|
|
||||||
lastPushedToParamsRef.current = toolParamValue
|
|
||||||
setStoreValue(toolParamValue)
|
|
||||||
}
|
|
||||||
}, [toolParamValue, setStoreValue])
|
|
||||||
|
|
||||||
// Sync store → tool.params: push when the user changes the value via SubBlock
|
|
||||||
useEffect(() => {
|
|
||||||
if (storeValue == null) return
|
|
||||||
const stringValue = typeof storeValue === 'string' ? storeValue : JSON.stringify(storeValue)
|
|
||||||
if (stringValue !== lastPushedToParamsRef.current) {
|
|
||||||
lastPushedToParamsRef.current = stringValue
|
|
||||||
lastPushedToStoreRef.current = stringValue
|
|
||||||
onParamChange(toolIndex, effectiveParamId, stringValue)
|
|
||||||
}
|
|
||||||
}, [storeValue, toolIndex, effectiveParamId, onParamChange])
|
|
||||||
|
|
||||||
// Determine if the parameter is optional for the user (LLM can fill it)
|
|
||||||
const visibility = subBlock.paramVisibility ?? 'user-or-llm'
|
|
||||||
const isOptionalForUser = visibility !== 'user-only'
|
|
||||||
|
|
||||||
const labelSuffix = useMemo(
|
|
||||||
() =>
|
|
||||||
isOptionalForUser ? (
|
|
||||||
<span className='ml-[6px] text-[12px] text-[var(--text-tertiary)]'>(optional)</span>
|
|
||||||
) : null,
|
|
||||||
[isOptionalForUser]
|
|
||||||
)
|
|
||||||
|
|
||||||
// Suppress SubBlock's "*" required indicator for optional-for-user params
|
|
||||||
const config = {
|
|
||||||
...subBlock,
|
|
||||||
id: syntheticId,
|
|
||||||
...(isOptionalForUser && { required: false }),
|
|
||||||
}
|
|
||||||
|
|
||||||
return (
|
|
||||||
<SubBlock
|
|
||||||
blockId={blockId}
|
|
||||||
config={config}
|
|
||||||
isPreview={false}
|
|
||||||
disabled={disabled}
|
|
||||||
canonicalToggle={canonicalToggle}
|
|
||||||
labelSuffix={labelSuffix}
|
|
||||||
/>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
@@ -2,12 +2,37 @@
|
|||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
import { describe, expect, it } from 'vitest'
|
import { describe, expect, it } from 'vitest'
|
||||||
import type { StoredTool } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/components/tool-input/types'
|
|
||||||
import {
|
interface StoredTool {
|
||||||
isCustomToolAlreadySelected,
|
type: string
|
||||||
isMcpToolAlreadySelected,
|
title?: string
|
||||||
isWorkflowAlreadySelected,
|
toolId?: string
|
||||||
} from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/components/tool-input/utils'
|
params?: Record<string, string>
|
||||||
|
customToolId?: string
|
||||||
|
schema?: any
|
||||||
|
code?: string
|
||||||
|
operation?: string
|
||||||
|
usageControl?: 'auto' | 'force' | 'none'
|
||||||
|
}
|
||||||
|
|
||||||
|
const isMcpToolAlreadySelected = (selectedTools: StoredTool[], mcpToolId: string): boolean => {
|
||||||
|
return selectedTools.some((tool) => tool.type === 'mcp' && tool.toolId === mcpToolId)
|
||||||
|
}
|
||||||
|
|
||||||
|
const isCustomToolAlreadySelected = (
|
||||||
|
selectedTools: StoredTool[],
|
||||||
|
customToolId: string
|
||||||
|
): boolean => {
|
||||||
|
return selectedTools.some(
|
||||||
|
(tool) => tool.type === 'custom-tool' && tool.customToolId === customToolId
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
const isWorkflowAlreadySelected = (selectedTools: StoredTool[], workflowId: string): boolean => {
|
||||||
|
return selectedTools.some(
|
||||||
|
(tool) => tool.type === 'workflow_input' && tool.params?.workflowId === workflowId
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
describe('isMcpToolAlreadySelected', () => {
|
describe('isMcpToolAlreadySelected', () => {
|
||||||
describe('basic functionality', () => {
|
describe('basic functionality', () => {
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -1,31 +0,0 @@
|
|||||||
/**
|
|
||||||
* Represents a tool selected and configured in the workflow
|
|
||||||
*
|
|
||||||
* @remarks
|
|
||||||
* For custom tools (new format), we only store: type, customToolId, usageControl, isExpanded.
|
|
||||||
* Everything else (title, schema, code) is loaded dynamically from the database.
|
|
||||||
* Legacy custom tools with inline schema/code are still supported for backwards compatibility.
|
|
||||||
*/
|
|
||||||
export interface StoredTool {
|
|
||||||
/** Block type identifier */
|
|
||||||
type: string
|
|
||||||
/** Display title for the tool (optional for new custom tool format) */
|
|
||||||
title?: string
|
|
||||||
/** Direct tool ID for execution (optional for new custom tool format) */
|
|
||||||
toolId?: string
|
|
||||||
/** Parameter values configured by the user (optional for new custom tool format) */
|
|
||||||
params?: Record<string, string>
|
|
||||||
/** Whether the tool details are expanded in UI */
|
|
||||||
isExpanded?: boolean
|
|
||||||
/** Database ID for custom tools (new format - reference only) */
|
|
||||||
customToolId?: string
|
|
||||||
/** Tool schema for custom tools (legacy format - inline JSON schema) */
|
|
||||||
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
||||||
schema?: Record<string, any>
|
|
||||||
/** Implementation code for custom tools (legacy format - inline) */
|
|
||||||
code?: string
|
|
||||||
/** Selected operation for multi-operation tools */
|
|
||||||
operation?: string
|
|
||||||
/** Tool usage control mode for LLM */
|
|
||||||
usageControl?: 'auto' | 'force' | 'none'
|
|
||||||
}
|
|
||||||
@@ -1,32 +0,0 @@
|
|||||||
import type { StoredTool } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/components/tool-input/types'
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Checks if an MCP tool is already selected.
|
|
||||||
*/
|
|
||||||
export function isMcpToolAlreadySelected(selectedTools: StoredTool[], mcpToolId: string): boolean {
|
|
||||||
return selectedTools.some((tool) => tool.type === 'mcp' && tool.toolId === mcpToolId)
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Checks if a custom tool is already selected.
|
|
||||||
*/
|
|
||||||
export function isCustomToolAlreadySelected(
|
|
||||||
selectedTools: StoredTool[],
|
|
||||||
customToolId: string
|
|
||||||
): boolean {
|
|
||||||
return selectedTools.some(
|
|
||||||
(tool) => tool.type === 'custom-tool' && tool.customToolId === customToolId
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Checks if a workflow is already selected.
|
|
||||||
*/
|
|
||||||
export function isWorkflowAlreadySelected(
|
|
||||||
selectedTools: StoredTool[],
|
|
||||||
workflowId: string
|
|
||||||
): boolean {
|
|
||||||
return selectedTools.some(
|
|
||||||
(tool) => tool.type === 'workflow_input' && tool.params?.workflowId === workflowId
|
|
||||||
)
|
|
||||||
}
|
|
||||||
@@ -76,7 +76,6 @@ interface SubBlockProps {
|
|||||||
disabled?: boolean
|
disabled?: boolean
|
||||||
onToggle?: () => void
|
onToggle?: () => void
|
||||||
}
|
}
|
||||||
labelSuffix?: React.ReactNode
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -203,8 +202,7 @@ const renderLabel = (
|
|||||||
showCopyButton: boolean
|
showCopyButton: boolean
|
||||||
copied: boolean
|
copied: boolean
|
||||||
onCopy: () => void
|
onCopy: () => void
|
||||||
},
|
}
|
||||||
labelSuffix?: React.ReactNode
|
|
||||||
): JSX.Element | null => {
|
): JSX.Element | null => {
|
||||||
if (config.type === 'switch') return null
|
if (config.type === 'switch') return null
|
||||||
if (!config.title) return null
|
if (!config.title) return null
|
||||||
@@ -220,7 +218,6 @@ const renderLabel = (
|
|||||||
<Label className='flex items-center gap-[6px] whitespace-nowrap'>
|
<Label className='flex items-center gap-[6px] whitespace-nowrap'>
|
||||||
{config.title}
|
{config.title}
|
||||||
{required && <span className='ml-0.5'>*</span>}
|
{required && <span className='ml-0.5'>*</span>}
|
||||||
{labelSuffix}
|
|
||||||
{config.type === 'code' &&
|
{config.type === 'code' &&
|
||||||
config.language === 'json' &&
|
config.language === 'json' &&
|
||||||
!isValidJson &&
|
!isValidJson &&
|
||||||
@@ -388,8 +385,7 @@ const arePropsEqual = (prevProps: SubBlockProps, nextProps: SubBlockProps): bool
|
|||||||
prevProps.disabled === nextProps.disabled &&
|
prevProps.disabled === nextProps.disabled &&
|
||||||
prevProps.fieldDiffStatus === nextProps.fieldDiffStatus &&
|
prevProps.fieldDiffStatus === nextProps.fieldDiffStatus &&
|
||||||
prevProps.allowExpandInPreview === nextProps.allowExpandInPreview &&
|
prevProps.allowExpandInPreview === nextProps.allowExpandInPreview &&
|
||||||
canonicalToggleEqual &&
|
canonicalToggleEqual
|
||||||
prevProps.labelSuffix === nextProps.labelSuffix
|
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -419,7 +415,6 @@ function SubBlockComponent({
|
|||||||
fieldDiffStatus,
|
fieldDiffStatus,
|
||||||
allowExpandInPreview,
|
allowExpandInPreview,
|
||||||
canonicalToggle,
|
canonicalToggle,
|
||||||
labelSuffix,
|
|
||||||
}: SubBlockProps): JSX.Element {
|
}: SubBlockProps): JSX.Element {
|
||||||
const [isValidJson, setIsValidJson] = useState(true)
|
const [isValidJson, setIsValidJson] = useState(true)
|
||||||
const [isSearchActive, setIsSearchActive] = useState(false)
|
const [isSearchActive, setIsSearchActive] = useState(false)
|
||||||
@@ -1064,8 +1059,7 @@ function SubBlockComponent({
|
|||||||
showCopyButton: Boolean(config.showCopyButton && config.useWebhookUrl),
|
showCopyButton: Boolean(config.showCopyButton && config.useWebhookUrl),
|
||||||
copied,
|
copied,
|
||||||
onCopy: handleCopy,
|
onCopy: handleCopy,
|
||||||
},
|
}
|
||||||
labelSuffix
|
|
||||||
)}
|
)}
|
||||||
{renderInput()}
|
{renderInput()}
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { useCallback, useEffect, useRef, useState } from 'react'
|
import { useCallback, useRef, useState } from 'react'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { useQueryClient } from '@tanstack/react-query'
|
import { useQueryClient } from '@tanstack/react-query'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
@@ -46,13 +46,7 @@ import { useWorkflowStore } from '@/stores/workflows/workflow/store'
|
|||||||
|
|
||||||
const logger = createLogger('useWorkflowExecution')
|
const logger = createLogger('useWorkflowExecution')
|
||||||
|
|
||||||
/**
|
// Debug state validation result
|
||||||
* Module-level Set tracking which workflows have an active reconnection effect.
|
|
||||||
* Prevents multiple hook instances (from different components) from starting
|
|
||||||
* concurrent reconnection streams for the same workflow during the same mount cycle.
|
|
||||||
*/
|
|
||||||
const activeReconnections = new Set<string>()
|
|
||||||
|
|
||||||
interface DebugValidationResult {
|
interface DebugValidationResult {
|
||||||
isValid: boolean
|
isValid: boolean
|
||||||
error?: string
|
error?: string
|
||||||
@@ -60,7 +54,7 @@ interface DebugValidationResult {
|
|||||||
|
|
||||||
interface BlockEventHandlerConfig {
|
interface BlockEventHandlerConfig {
|
||||||
workflowId?: string
|
workflowId?: string
|
||||||
executionIdRef: { current: string }
|
executionId?: string
|
||||||
workflowEdges: Array<{ id: string; target: string; sourceHandle?: string | null }>
|
workflowEdges: Array<{ id: string; target: string; sourceHandle?: string | null }>
|
||||||
activeBlocksSet: Set<string>
|
activeBlocksSet: Set<string>
|
||||||
accumulatedBlockLogs: BlockLog[]
|
accumulatedBlockLogs: BlockLog[]
|
||||||
@@ -114,15 +108,12 @@ export function useWorkflowExecution() {
|
|||||||
const queryClient = useQueryClient()
|
const queryClient = useQueryClient()
|
||||||
const currentWorkflow = useCurrentWorkflow()
|
const currentWorkflow = useCurrentWorkflow()
|
||||||
const { activeWorkflowId, workflows } = useWorkflowRegistry()
|
const { activeWorkflowId, workflows } = useWorkflowRegistry()
|
||||||
const { toggleConsole, addConsole, updateConsole, cancelRunningEntries, clearExecutionEntries } =
|
const { toggleConsole, addConsole, updateConsole, cancelRunningEntries } =
|
||||||
useTerminalConsoleStore()
|
useTerminalConsoleStore()
|
||||||
const hasHydrated = useTerminalConsoleStore((s) => s._hasHydrated)
|
|
||||||
const { getAllVariables } = useEnvironmentStore()
|
const { getAllVariables } = useEnvironmentStore()
|
||||||
const { getVariablesByWorkflowId, variables } = useVariablesStore()
|
const { getVariablesByWorkflowId, variables } = useVariablesStore()
|
||||||
const { isExecuting, isDebugging, pendingBlocks, executor, debugContext } =
|
const { isExecuting, isDebugging, pendingBlocks, executor, debugContext } =
|
||||||
useCurrentWorkflowExecution()
|
useCurrentWorkflowExecution()
|
||||||
const setCurrentExecutionId = useExecutionStore((s) => s.setCurrentExecutionId)
|
|
||||||
const getCurrentExecutionId = useExecutionStore((s) => s.getCurrentExecutionId)
|
|
||||||
const setIsExecuting = useExecutionStore((s) => s.setIsExecuting)
|
const setIsExecuting = useExecutionStore((s) => s.setIsExecuting)
|
||||||
const setIsDebugging = useExecutionStore((s) => s.setIsDebugging)
|
const setIsDebugging = useExecutionStore((s) => s.setIsDebugging)
|
||||||
const setPendingBlocks = useExecutionStore((s) => s.setPendingBlocks)
|
const setPendingBlocks = useExecutionStore((s) => s.setPendingBlocks)
|
||||||
@@ -306,7 +297,7 @@ export function useWorkflowExecution() {
|
|||||||
(config: BlockEventHandlerConfig) => {
|
(config: BlockEventHandlerConfig) => {
|
||||||
const {
|
const {
|
||||||
workflowId,
|
workflowId,
|
||||||
executionIdRef,
|
executionId,
|
||||||
workflowEdges,
|
workflowEdges,
|
||||||
activeBlocksSet,
|
activeBlocksSet,
|
||||||
accumulatedBlockLogs,
|
accumulatedBlockLogs,
|
||||||
@@ -317,14 +308,6 @@ export function useWorkflowExecution() {
|
|||||||
onBlockCompleteCallback,
|
onBlockCompleteCallback,
|
||||||
} = config
|
} = config
|
||||||
|
|
||||||
/** Returns true if this execution was cancelled or superseded by another run. */
|
|
||||||
const isStaleExecution = () =>
|
|
||||||
!!(
|
|
||||||
workflowId &&
|
|
||||||
executionIdRef.current &&
|
|
||||||
useExecutionStore.getState().getCurrentExecutionId(workflowId) !== executionIdRef.current
|
|
||||||
)
|
|
||||||
|
|
||||||
const updateActiveBlocks = (blockId: string, isActive: boolean) => {
|
const updateActiveBlocks = (blockId: string, isActive: boolean) => {
|
||||||
if (!workflowId) return
|
if (!workflowId) return
|
||||||
if (isActive) {
|
if (isActive) {
|
||||||
@@ -377,7 +360,7 @@ export function useWorkflowExecution() {
|
|||||||
endedAt: data.endedAt,
|
endedAt: data.endedAt,
|
||||||
workflowId,
|
workflowId,
|
||||||
blockId: data.blockId,
|
blockId: data.blockId,
|
||||||
executionId: executionIdRef.current,
|
executionId,
|
||||||
blockName: data.blockName || 'Unknown Block',
|
blockName: data.blockName || 'Unknown Block',
|
||||||
blockType: data.blockType || 'unknown',
|
blockType: data.blockType || 'unknown',
|
||||||
iterationCurrent: data.iterationCurrent,
|
iterationCurrent: data.iterationCurrent,
|
||||||
@@ -400,7 +383,7 @@ export function useWorkflowExecution() {
|
|||||||
endedAt: data.endedAt,
|
endedAt: data.endedAt,
|
||||||
workflowId,
|
workflowId,
|
||||||
blockId: data.blockId,
|
blockId: data.blockId,
|
||||||
executionId: executionIdRef.current,
|
executionId,
|
||||||
blockName: data.blockName || 'Unknown Block',
|
blockName: data.blockName || 'Unknown Block',
|
||||||
blockType: data.blockType || 'unknown',
|
blockType: data.blockType || 'unknown',
|
||||||
iterationCurrent: data.iterationCurrent,
|
iterationCurrent: data.iterationCurrent,
|
||||||
@@ -427,7 +410,7 @@ export function useWorkflowExecution() {
|
|||||||
iterationType: data.iterationType,
|
iterationType: data.iterationType,
|
||||||
iterationContainerId: data.iterationContainerId,
|
iterationContainerId: data.iterationContainerId,
|
||||||
},
|
},
|
||||||
executionIdRef.current
|
executionId
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -449,12 +432,11 @@ export function useWorkflowExecution() {
|
|||||||
iterationType: data.iterationType,
|
iterationType: data.iterationType,
|
||||||
iterationContainerId: data.iterationContainerId,
|
iterationContainerId: data.iterationContainerId,
|
||||||
},
|
},
|
||||||
executionIdRef.current
|
executionId
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
const onBlockStarted = (data: BlockStartedData) => {
|
const onBlockStarted = (data: BlockStartedData) => {
|
||||||
if (isStaleExecution()) return
|
|
||||||
updateActiveBlocks(data.blockId, true)
|
updateActiveBlocks(data.blockId, true)
|
||||||
markIncomingEdges(data.blockId)
|
markIncomingEdges(data.blockId)
|
||||||
|
|
||||||
@@ -471,7 +453,7 @@ export function useWorkflowExecution() {
|
|||||||
endedAt: undefined,
|
endedAt: undefined,
|
||||||
workflowId,
|
workflowId,
|
||||||
blockId: data.blockId,
|
blockId: data.blockId,
|
||||||
executionId: executionIdRef.current,
|
executionId,
|
||||||
blockName: data.blockName || 'Unknown Block',
|
blockName: data.blockName || 'Unknown Block',
|
||||||
blockType: data.blockType || 'unknown',
|
blockType: data.blockType || 'unknown',
|
||||||
isRunning: true,
|
isRunning: true,
|
||||||
@@ -483,7 +465,6 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const onBlockCompleted = (data: BlockCompletedData) => {
|
const onBlockCompleted = (data: BlockCompletedData) => {
|
||||||
if (isStaleExecution()) return
|
|
||||||
updateActiveBlocks(data.blockId, false)
|
updateActiveBlocks(data.blockId, false)
|
||||||
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'success')
|
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'success')
|
||||||
|
|
||||||
@@ -514,7 +495,6 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const onBlockError = (data: BlockErrorData) => {
|
const onBlockError = (data: BlockErrorData) => {
|
||||||
if (isStaleExecution()) return
|
|
||||||
updateActiveBlocks(data.blockId, false)
|
updateActiveBlocks(data.blockId, false)
|
||||||
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'error')
|
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'error')
|
||||||
|
|
||||||
@@ -922,6 +902,10 @@ export function useWorkflowExecution() {
|
|||||||
|
|
||||||
// Update block logs with actual stream completion times
|
// Update block logs with actual stream completion times
|
||||||
if (result.logs && streamCompletionTimes.size > 0) {
|
if (result.logs && streamCompletionTimes.size > 0) {
|
||||||
|
const streamCompletionEndTime = new Date(
|
||||||
|
Math.max(...Array.from(streamCompletionTimes.values()))
|
||||||
|
).toISOString()
|
||||||
|
|
||||||
result.logs.forEach((log: BlockLog) => {
|
result.logs.forEach((log: BlockLog) => {
|
||||||
if (streamCompletionTimes.has(log.blockId)) {
|
if (streamCompletionTimes.has(log.blockId)) {
|
||||||
const completionTime = streamCompletionTimes.get(log.blockId)!
|
const completionTime = streamCompletionTimes.get(log.blockId)!
|
||||||
@@ -1003,6 +987,7 @@ export function useWorkflowExecution() {
|
|||||||
return { success: true, stream }
|
return { success: true, stream }
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// For manual (non-chat) execution
|
||||||
const manualExecutionId = uuidv4()
|
const manualExecutionId = uuidv4()
|
||||||
try {
|
try {
|
||||||
const result = await executeWorkflow(
|
const result = await executeWorkflow(
|
||||||
@@ -1017,10 +1002,29 @@ export function useWorkflowExecution() {
|
|||||||
if (result.metadata.pendingBlocks) {
|
if (result.metadata.pendingBlocks) {
|
||||||
setPendingBlocks(activeWorkflowId, result.metadata.pendingBlocks)
|
setPendingBlocks(activeWorkflowId, result.metadata.pendingBlocks)
|
||||||
}
|
}
|
||||||
|
} else if (result && 'success' in result) {
|
||||||
|
setExecutionResult(result)
|
||||||
|
// Reset execution state after successful non-debug execution
|
||||||
|
setIsExecuting(activeWorkflowId, false)
|
||||||
|
setIsDebugging(activeWorkflowId, false)
|
||||||
|
setActiveBlocks(activeWorkflowId, new Set())
|
||||||
|
|
||||||
|
if (isChatExecution) {
|
||||||
|
if (!result.metadata) {
|
||||||
|
result.metadata = { duration: 0, startTime: new Date().toISOString() }
|
||||||
|
}
|
||||||
|
;(result.metadata as any).source = 'chat'
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalidate subscription queries to update usage
|
||||||
|
setTimeout(() => {
|
||||||
|
queryClient.invalidateQueries({ queryKey: subscriptionKeys.all })
|
||||||
|
}, 1000)
|
||||||
}
|
}
|
||||||
return result
|
return result
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
const errorResult = handleExecutionError(error, { executionId: manualExecutionId })
|
const errorResult = handleExecutionError(error, { executionId: manualExecutionId })
|
||||||
|
// Note: Error logs are already persisted server-side via execution-core.ts
|
||||||
return errorResult
|
return errorResult
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@@ -1271,7 +1275,7 @@ export function useWorkflowExecution() {
|
|||||||
if (activeWorkflowId) {
|
if (activeWorkflowId) {
|
||||||
logger.info('Using server-side executor')
|
logger.info('Using server-side executor')
|
||||||
|
|
||||||
const executionIdRef = { current: '' }
|
const executionId = uuidv4()
|
||||||
|
|
||||||
let executionResult: ExecutionResult = {
|
let executionResult: ExecutionResult = {
|
||||||
success: false,
|
success: false,
|
||||||
@@ -1289,7 +1293,7 @@ export function useWorkflowExecution() {
|
|||||||
try {
|
try {
|
||||||
const blockHandlers = buildBlockEventHandlers({
|
const blockHandlers = buildBlockEventHandlers({
|
||||||
workflowId: activeWorkflowId,
|
workflowId: activeWorkflowId,
|
||||||
executionIdRef,
|
executionId,
|
||||||
workflowEdges,
|
workflowEdges,
|
||||||
activeBlocksSet,
|
activeBlocksSet,
|
||||||
accumulatedBlockLogs,
|
accumulatedBlockLogs,
|
||||||
@@ -1322,10 +1326,6 @@ export function useWorkflowExecution() {
|
|||||||
loops: clientWorkflowState.loops,
|
loops: clientWorkflowState.loops,
|
||||||
parallels: clientWorkflowState.parallels,
|
parallels: clientWorkflowState.parallels,
|
||||||
},
|
},
|
||||||
onExecutionId: (id) => {
|
|
||||||
executionIdRef.current = id
|
|
||||||
setCurrentExecutionId(activeWorkflowId, id)
|
|
||||||
},
|
|
||||||
callbacks: {
|
callbacks: {
|
||||||
onExecutionStarted: (data) => {
|
onExecutionStarted: (data) => {
|
||||||
logger.info('Server execution started:', data)
|
logger.info('Server execution started:', data)
|
||||||
@@ -1368,18 +1368,6 @@ export function useWorkflowExecution() {
|
|||||||
},
|
},
|
||||||
|
|
||||||
onExecutionCompleted: (data) => {
|
onExecutionCompleted: (data) => {
|
||||||
if (
|
|
||||||
activeWorkflowId &&
|
|
||||||
executionIdRef.current &&
|
|
||||||
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
|
||||||
executionIdRef.current
|
|
||||||
)
|
|
||||||
return
|
|
||||||
|
|
||||||
if (activeWorkflowId) {
|
|
||||||
setCurrentExecutionId(activeWorkflowId, null)
|
|
||||||
}
|
|
||||||
|
|
||||||
executionResult = {
|
executionResult = {
|
||||||
success: data.success,
|
success: data.success,
|
||||||
output: data.output,
|
output: data.output,
|
||||||
@@ -1437,33 +1425,9 @@ export function useWorkflowExecution() {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const workflowExecState = activeWorkflowId
|
|
||||||
? useExecutionStore.getState().getWorkflowExecution(activeWorkflowId)
|
|
||||||
: null
|
|
||||||
if (activeWorkflowId && !workflowExecState?.isDebugging) {
|
|
||||||
setExecutionResult(executionResult)
|
|
||||||
setIsExecuting(activeWorkflowId, false)
|
|
||||||
setActiveBlocks(activeWorkflowId, new Set())
|
|
||||||
setTimeout(() => {
|
|
||||||
queryClient.invalidateQueries({ queryKey: subscriptionKeys.all })
|
|
||||||
}, 1000)
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
|
|
||||||
onExecutionError: (data) => {
|
onExecutionError: (data) => {
|
||||||
if (
|
|
||||||
activeWorkflowId &&
|
|
||||||
executionIdRef.current &&
|
|
||||||
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
|
||||||
executionIdRef.current
|
|
||||||
)
|
|
||||||
return
|
|
||||||
|
|
||||||
if (activeWorkflowId) {
|
|
||||||
setCurrentExecutionId(activeWorkflowId, null)
|
|
||||||
}
|
|
||||||
|
|
||||||
executionResult = {
|
executionResult = {
|
||||||
success: false,
|
success: false,
|
||||||
output: {},
|
output: {},
|
||||||
@@ -1477,53 +1441,43 @@ export function useWorkflowExecution() {
|
|||||||
const isPreExecutionError = accumulatedBlockLogs.length === 0
|
const isPreExecutionError = accumulatedBlockLogs.length === 0
|
||||||
handleExecutionErrorConsole({
|
handleExecutionErrorConsole({
|
||||||
workflowId: activeWorkflowId,
|
workflowId: activeWorkflowId,
|
||||||
executionId: executionIdRef.current,
|
executionId,
|
||||||
error: data.error,
|
error: data.error,
|
||||||
durationMs: data.duration,
|
durationMs: data.duration,
|
||||||
blockLogs: accumulatedBlockLogs,
|
blockLogs: accumulatedBlockLogs,
|
||||||
isPreExecutionError,
|
isPreExecutionError,
|
||||||
})
|
})
|
||||||
|
|
||||||
if (activeWorkflowId) {
|
|
||||||
setIsExecuting(activeWorkflowId, false)
|
|
||||||
setIsDebugging(activeWorkflowId, false)
|
|
||||||
setActiveBlocks(activeWorkflowId, new Set())
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
|
|
||||||
onExecutionCancelled: (data) => {
|
onExecutionCancelled: (data) => {
|
||||||
if (
|
|
||||||
activeWorkflowId &&
|
|
||||||
executionIdRef.current &&
|
|
||||||
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
|
||||||
executionIdRef.current
|
|
||||||
)
|
|
||||||
return
|
|
||||||
|
|
||||||
if (activeWorkflowId) {
|
|
||||||
setCurrentExecutionId(activeWorkflowId, null)
|
|
||||||
}
|
|
||||||
|
|
||||||
handleExecutionCancelledConsole({
|
handleExecutionCancelledConsole({
|
||||||
workflowId: activeWorkflowId,
|
workflowId: activeWorkflowId,
|
||||||
executionId: executionIdRef.current,
|
executionId,
|
||||||
durationMs: data?.duration,
|
durationMs: data?.duration,
|
||||||
})
|
})
|
||||||
|
|
||||||
if (activeWorkflowId) {
|
|
||||||
setIsExecuting(activeWorkflowId, false)
|
|
||||||
setIsDebugging(activeWorkflowId, false)
|
|
||||||
setActiveBlocks(activeWorkflowId, new Set())
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
return executionResult
|
return executionResult
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
|
// Don't log abort errors - they're intentional user actions
|
||||||
if (error.name === 'AbortError' || error.message?.includes('aborted')) {
|
if (error.name === 'AbortError' || error.message?.includes('aborted')) {
|
||||||
logger.info('Execution aborted by user')
|
logger.info('Execution aborted by user')
|
||||||
return executionResult
|
|
||||||
|
// Reset execution state
|
||||||
|
if (activeWorkflowId) {
|
||||||
|
setIsExecuting(activeWorkflowId, false)
|
||||||
|
setActiveBlocks(activeWorkflowId, new Set())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return gracefully without error
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
output: {},
|
||||||
|
metadata: { duration: 0 },
|
||||||
|
logs: [],
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.error('Server-side execution failed:', error)
|
logger.error('Server-side execution failed:', error)
|
||||||
@@ -1531,6 +1485,7 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Fallback: should never reach here
|
||||||
throw new Error('Server-side execution is required')
|
throw new Error('Server-side execution is required')
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1762,28 +1717,25 @@ export function useWorkflowExecution() {
|
|||||||
* Handles cancelling the current workflow execution
|
* Handles cancelling the current workflow execution
|
||||||
*/
|
*/
|
||||||
const handleCancelExecution = useCallback(() => {
|
const handleCancelExecution = useCallback(() => {
|
||||||
if (!activeWorkflowId) return
|
|
||||||
logger.info('Workflow execution cancellation requested')
|
logger.info('Workflow execution cancellation requested')
|
||||||
|
|
||||||
const storedExecutionId = getCurrentExecutionId(activeWorkflowId)
|
// Cancel the execution stream for this workflow (server-side)
|
||||||
|
executionStream.cancel(activeWorkflowId ?? undefined)
|
||||||
|
|
||||||
if (storedExecutionId) {
|
// Mark current chat execution as superseded so its cleanup won't affect new executions
|
||||||
setCurrentExecutionId(activeWorkflowId, null)
|
currentChatExecutionIdRef.current = null
|
||||||
fetch(`/api/workflows/${activeWorkflowId}/executions/${storedExecutionId}/cancel`, {
|
|
||||||
method: 'POST',
|
// Mark all running entries as canceled in the terminal
|
||||||
}).catch(() => {})
|
if (activeWorkflowId) {
|
||||||
handleExecutionCancelledConsole({
|
cancelRunningEntries(activeWorkflowId)
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
executionId: storedExecutionId,
|
// Reset execution state - this triggers chat stream cleanup via useEffect in chat.tsx
|
||||||
})
|
setIsExecuting(activeWorkflowId, false)
|
||||||
|
setIsDebugging(activeWorkflowId, false)
|
||||||
|
setActiveBlocks(activeWorkflowId, new Set())
|
||||||
}
|
}
|
||||||
|
|
||||||
executionStream.cancel(activeWorkflowId)
|
// If in debug mode, also reset debug state
|
||||||
currentChatExecutionIdRef.current = null
|
|
||||||
setIsExecuting(activeWorkflowId, false)
|
|
||||||
setIsDebugging(activeWorkflowId, false)
|
|
||||||
setActiveBlocks(activeWorkflowId, new Set())
|
|
||||||
|
|
||||||
if (isDebugging) {
|
if (isDebugging) {
|
||||||
resetDebugState()
|
resetDebugState()
|
||||||
}
|
}
|
||||||
@@ -1795,9 +1747,7 @@ export function useWorkflowExecution() {
|
|||||||
setIsDebugging,
|
setIsDebugging,
|
||||||
setActiveBlocks,
|
setActiveBlocks,
|
||||||
activeWorkflowId,
|
activeWorkflowId,
|
||||||
getCurrentExecutionId,
|
cancelRunningEntries,
|
||||||
setCurrentExecutionId,
|
|
||||||
handleExecutionCancelledConsole,
|
|
||||||
])
|
])
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -1897,7 +1847,7 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
setIsExecuting(workflowId, true)
|
setIsExecuting(workflowId, true)
|
||||||
const executionIdRef = { current: '' }
|
const executionId = uuidv4()
|
||||||
const accumulatedBlockLogs: BlockLog[] = []
|
const accumulatedBlockLogs: BlockLog[] = []
|
||||||
const accumulatedBlockStates = new Map<string, BlockState>()
|
const accumulatedBlockStates = new Map<string, BlockState>()
|
||||||
const executedBlockIds = new Set<string>()
|
const executedBlockIds = new Set<string>()
|
||||||
@@ -1906,7 +1856,7 @@ export function useWorkflowExecution() {
|
|||||||
try {
|
try {
|
||||||
const blockHandlers = buildBlockEventHandlers({
|
const blockHandlers = buildBlockEventHandlers({
|
||||||
workflowId,
|
workflowId,
|
||||||
executionIdRef,
|
executionId,
|
||||||
workflowEdges,
|
workflowEdges,
|
||||||
activeBlocksSet,
|
activeBlocksSet,
|
||||||
accumulatedBlockLogs,
|
accumulatedBlockLogs,
|
||||||
@@ -1921,10 +1871,6 @@ export function useWorkflowExecution() {
|
|||||||
startBlockId: blockId,
|
startBlockId: blockId,
|
||||||
sourceSnapshot: effectiveSnapshot,
|
sourceSnapshot: effectiveSnapshot,
|
||||||
input: workflowInput,
|
input: workflowInput,
|
||||||
onExecutionId: (id) => {
|
|
||||||
executionIdRef.current = id
|
|
||||||
setCurrentExecutionId(workflowId, id)
|
|
||||||
},
|
|
||||||
callbacks: {
|
callbacks: {
|
||||||
onBlockStarted: blockHandlers.onBlockStarted,
|
onBlockStarted: blockHandlers.onBlockStarted,
|
||||||
onBlockCompleted: blockHandlers.onBlockCompleted,
|
onBlockCompleted: blockHandlers.onBlockCompleted,
|
||||||
@@ -1932,6 +1878,7 @@ export function useWorkflowExecution() {
|
|||||||
|
|
||||||
onExecutionCompleted: (data) => {
|
onExecutionCompleted: (data) => {
|
||||||
if (data.success) {
|
if (data.success) {
|
||||||
|
// Add the start block (trigger) to executed blocks
|
||||||
executedBlockIds.add(blockId)
|
executedBlockIds.add(blockId)
|
||||||
|
|
||||||
const mergedBlockStates: Record<string, BlockState> = {
|
const mergedBlockStates: Record<string, BlockState> = {
|
||||||
@@ -1955,10 +1902,6 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
setLastExecutionSnapshot(workflowId, updatedSnapshot)
|
setLastExecutionSnapshot(workflowId, updatedSnapshot)
|
||||||
}
|
}
|
||||||
|
|
||||||
setCurrentExecutionId(workflowId, null)
|
|
||||||
setIsExecuting(workflowId, false)
|
|
||||||
setActiveBlocks(workflowId, new Set())
|
|
||||||
},
|
},
|
||||||
|
|
||||||
onExecutionError: (data) => {
|
onExecutionError: (data) => {
|
||||||
@@ -1978,27 +1921,19 @@ export function useWorkflowExecution() {
|
|||||||
|
|
||||||
handleExecutionErrorConsole({
|
handleExecutionErrorConsole({
|
||||||
workflowId,
|
workflowId,
|
||||||
executionId: executionIdRef.current,
|
executionId,
|
||||||
error: data.error,
|
error: data.error,
|
||||||
durationMs: data.duration,
|
durationMs: data.duration,
|
||||||
blockLogs: accumulatedBlockLogs,
|
blockLogs: accumulatedBlockLogs,
|
||||||
})
|
})
|
||||||
|
|
||||||
setCurrentExecutionId(workflowId, null)
|
|
||||||
setIsExecuting(workflowId, false)
|
|
||||||
setActiveBlocks(workflowId, new Set())
|
|
||||||
},
|
},
|
||||||
|
|
||||||
onExecutionCancelled: (data) => {
|
onExecutionCancelled: (data) => {
|
||||||
handleExecutionCancelledConsole({
|
handleExecutionCancelledConsole({
|
||||||
workflowId,
|
workflowId,
|
||||||
executionId: executionIdRef.current,
|
executionId,
|
||||||
durationMs: data?.duration,
|
durationMs: data?.duration,
|
||||||
})
|
})
|
||||||
|
|
||||||
setCurrentExecutionId(workflowId, null)
|
|
||||||
setIsExecuting(workflowId, false)
|
|
||||||
setActiveBlocks(workflowId, new Set())
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
@@ -2007,20 +1942,14 @@ export function useWorkflowExecution() {
|
|||||||
logger.error('Run-from-block failed:', error)
|
logger.error('Run-from-block failed:', error)
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
const currentId = getCurrentExecutionId(workflowId)
|
setIsExecuting(workflowId, false)
|
||||||
if (currentId === null || currentId === executionIdRef.current) {
|
setActiveBlocks(workflowId, new Set())
|
||||||
setCurrentExecutionId(workflowId, null)
|
|
||||||
setIsExecuting(workflowId, false)
|
|
||||||
setActiveBlocks(workflowId, new Set())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
[
|
[
|
||||||
getLastExecutionSnapshot,
|
getLastExecutionSnapshot,
|
||||||
setLastExecutionSnapshot,
|
setLastExecutionSnapshot,
|
||||||
clearLastExecutionSnapshot,
|
clearLastExecutionSnapshot,
|
||||||
getCurrentExecutionId,
|
|
||||||
setCurrentExecutionId,
|
|
||||||
setIsExecuting,
|
setIsExecuting,
|
||||||
setActiveBlocks,
|
setActiveBlocks,
|
||||||
setBlockRunStatus,
|
setBlockRunStatus,
|
||||||
@@ -2050,213 +1979,29 @@ export function useWorkflowExecution() {
|
|||||||
|
|
||||||
const executionId = uuidv4()
|
const executionId = uuidv4()
|
||||||
try {
|
try {
|
||||||
await executeWorkflow(undefined, undefined, executionId, undefined, 'manual', blockId)
|
const result = await executeWorkflow(
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
executionId,
|
||||||
|
undefined,
|
||||||
|
'manual',
|
||||||
|
blockId
|
||||||
|
)
|
||||||
|
if (result && 'success' in result) {
|
||||||
|
setExecutionResult(result)
|
||||||
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
const errorResult = handleExecutionError(error, { executionId })
|
const errorResult = handleExecutionError(error, { executionId })
|
||||||
return errorResult
|
return errorResult
|
||||||
} finally {
|
} finally {
|
||||||
setCurrentExecutionId(workflowId, null)
|
|
||||||
setIsExecuting(workflowId, false)
|
setIsExecuting(workflowId, false)
|
||||||
setIsDebugging(workflowId, false)
|
setIsDebugging(workflowId, false)
|
||||||
setActiveBlocks(workflowId, new Set())
|
setActiveBlocks(workflowId, new Set())
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
[
|
[activeWorkflowId, setExecutionResult, setIsExecuting, setIsDebugging, setActiveBlocks]
|
||||||
activeWorkflowId,
|
|
||||||
setCurrentExecutionId,
|
|
||||||
setExecutionResult,
|
|
||||||
setIsExecuting,
|
|
||||||
setIsDebugging,
|
|
||||||
setActiveBlocks,
|
|
||||||
]
|
|
||||||
)
|
)
|
||||||
|
|
||||||
useEffect(() => {
|
|
||||||
if (!activeWorkflowId || !hasHydrated) return
|
|
||||||
|
|
||||||
const entries = useTerminalConsoleStore.getState().entries
|
|
||||||
const runningEntries = entries.filter(
|
|
||||||
(e) => e.isRunning && e.workflowId === activeWorkflowId && e.executionId
|
|
||||||
)
|
|
||||||
if (runningEntries.length === 0) return
|
|
||||||
|
|
||||||
if (activeReconnections.has(activeWorkflowId)) return
|
|
||||||
activeReconnections.add(activeWorkflowId)
|
|
||||||
|
|
||||||
executionStream.cancel(activeWorkflowId)
|
|
||||||
|
|
||||||
const sorted = [...runningEntries].sort((a, b) => {
|
|
||||||
const aTime = a.startedAt ? new Date(a.startedAt).getTime() : 0
|
|
||||||
const bTime = b.startedAt ? new Date(b.startedAt).getTime() : 0
|
|
||||||
return bTime - aTime
|
|
||||||
})
|
|
||||||
const executionId = sorted[0].executionId!
|
|
||||||
|
|
||||||
const otherExecutionIds = new Set(
|
|
||||||
sorted.filter((e) => e.executionId !== executionId).map((e) => e.executionId!)
|
|
||||||
)
|
|
||||||
if (otherExecutionIds.size > 0) {
|
|
||||||
cancelRunningEntries(activeWorkflowId)
|
|
||||||
}
|
|
||||||
|
|
||||||
setCurrentExecutionId(activeWorkflowId, executionId)
|
|
||||||
setIsExecuting(activeWorkflowId, true)
|
|
||||||
|
|
||||||
const workflowEdges = useWorkflowStore.getState().edges
|
|
||||||
const activeBlocksSet = new Set<string>()
|
|
||||||
const accumulatedBlockLogs: BlockLog[] = []
|
|
||||||
const accumulatedBlockStates = new Map<string, BlockState>()
|
|
||||||
const executedBlockIds = new Set<string>()
|
|
||||||
|
|
||||||
const executionIdRef = { current: executionId }
|
|
||||||
|
|
||||||
const handlers = buildBlockEventHandlers({
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
executionIdRef,
|
|
||||||
workflowEdges,
|
|
||||||
activeBlocksSet,
|
|
||||||
accumulatedBlockLogs,
|
|
||||||
accumulatedBlockStates,
|
|
||||||
executedBlockIds,
|
|
||||||
consoleMode: 'update',
|
|
||||||
includeStartConsoleEntry: true,
|
|
||||||
})
|
|
||||||
|
|
||||||
const originalEntries = entries
|
|
||||||
.filter((e) => e.executionId === executionId)
|
|
||||||
.map((e) => ({ ...e }))
|
|
||||||
|
|
||||||
let cleared = false
|
|
||||||
let reconnectionComplete = false
|
|
||||||
let cleanupRan = false
|
|
||||||
const clearOnce = () => {
|
|
||||||
if (!cleared) {
|
|
||||||
cleared = true
|
|
||||||
clearExecutionEntries(executionId)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const reconnectWorkflowId = activeWorkflowId
|
|
||||||
|
|
||||||
executionStream
|
|
||||||
.reconnect({
|
|
||||||
workflowId: reconnectWorkflowId,
|
|
||||||
executionId,
|
|
||||||
callbacks: {
|
|
||||||
onBlockStarted: (data) => {
|
|
||||||
clearOnce()
|
|
||||||
handlers.onBlockStarted(data)
|
|
||||||
},
|
|
||||||
onBlockCompleted: (data) => {
|
|
||||||
clearOnce()
|
|
||||||
handlers.onBlockCompleted(data)
|
|
||||||
},
|
|
||||||
onBlockError: (data) => {
|
|
||||||
clearOnce()
|
|
||||||
handlers.onBlockError(data)
|
|
||||||
},
|
|
||||||
onExecutionCompleted: () => {
|
|
||||||
const currentId = useExecutionStore
|
|
||||||
.getState()
|
|
||||||
.getCurrentExecutionId(reconnectWorkflowId)
|
|
||||||
if (currentId !== executionId) {
|
|
||||||
reconnectionComplete = true
|
|
||||||
activeReconnections.delete(reconnectWorkflowId)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
clearOnce()
|
|
||||||
reconnectionComplete = true
|
|
||||||
activeReconnections.delete(reconnectWorkflowId)
|
|
||||||
setCurrentExecutionId(reconnectWorkflowId, null)
|
|
||||||
setIsExecuting(reconnectWorkflowId, false)
|
|
||||||
setActiveBlocks(reconnectWorkflowId, new Set())
|
|
||||||
},
|
|
||||||
onExecutionError: (data) => {
|
|
||||||
const currentId = useExecutionStore
|
|
||||||
.getState()
|
|
||||||
.getCurrentExecutionId(reconnectWorkflowId)
|
|
||||||
if (currentId !== executionId) {
|
|
||||||
reconnectionComplete = true
|
|
||||||
activeReconnections.delete(reconnectWorkflowId)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
clearOnce()
|
|
||||||
reconnectionComplete = true
|
|
||||||
activeReconnections.delete(reconnectWorkflowId)
|
|
||||||
setCurrentExecutionId(reconnectWorkflowId, null)
|
|
||||||
setIsExecuting(reconnectWorkflowId, false)
|
|
||||||
setActiveBlocks(reconnectWorkflowId, new Set())
|
|
||||||
handleExecutionErrorConsole({
|
|
||||||
workflowId: reconnectWorkflowId,
|
|
||||||
executionId,
|
|
||||||
error: data.error,
|
|
||||||
blockLogs: accumulatedBlockLogs,
|
|
||||||
})
|
|
||||||
},
|
|
||||||
onExecutionCancelled: () => {
|
|
||||||
const currentId = useExecutionStore
|
|
||||||
.getState()
|
|
||||||
.getCurrentExecutionId(reconnectWorkflowId)
|
|
||||||
if (currentId !== executionId) {
|
|
||||||
reconnectionComplete = true
|
|
||||||
activeReconnections.delete(reconnectWorkflowId)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
clearOnce()
|
|
||||||
reconnectionComplete = true
|
|
||||||
activeReconnections.delete(reconnectWorkflowId)
|
|
||||||
setCurrentExecutionId(reconnectWorkflowId, null)
|
|
||||||
setIsExecuting(reconnectWorkflowId, false)
|
|
||||||
setActiveBlocks(reconnectWorkflowId, new Set())
|
|
||||||
handleExecutionCancelledConsole({
|
|
||||||
workflowId: reconnectWorkflowId,
|
|
||||||
executionId,
|
|
||||||
})
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
.catch((error) => {
|
|
||||||
logger.warn('Execution reconnection failed', { executionId, error })
|
|
||||||
})
|
|
||||||
.finally(() => {
|
|
||||||
if (reconnectionComplete || cleanupRan) return
|
|
||||||
const currentId = useExecutionStore.getState().getCurrentExecutionId(reconnectWorkflowId)
|
|
||||||
if (currentId !== executionId) return
|
|
||||||
reconnectionComplete = true
|
|
||||||
activeReconnections.delete(reconnectWorkflowId)
|
|
||||||
clearExecutionEntries(executionId)
|
|
||||||
for (const entry of originalEntries) {
|
|
||||||
addConsole({
|
|
||||||
workflowId: entry.workflowId,
|
|
||||||
blockId: entry.blockId,
|
|
||||||
blockName: entry.blockName,
|
|
||||||
blockType: entry.blockType,
|
|
||||||
executionId: entry.executionId,
|
|
||||||
executionOrder: entry.executionOrder,
|
|
||||||
isRunning: false,
|
|
||||||
warning: 'Execution result unavailable — check the logs page',
|
|
||||||
})
|
|
||||||
}
|
|
||||||
setCurrentExecutionId(reconnectWorkflowId, null)
|
|
||||||
setIsExecuting(reconnectWorkflowId, false)
|
|
||||||
setActiveBlocks(reconnectWorkflowId, new Set())
|
|
||||||
})
|
|
||||||
|
|
||||||
return () => {
|
|
||||||
cleanupRan = true
|
|
||||||
executionStream.cancel(reconnectWorkflowId)
|
|
||||||
activeReconnections.delete(reconnectWorkflowId)
|
|
||||||
|
|
||||||
if (cleared && !reconnectionComplete) {
|
|
||||||
clearExecutionEntries(executionId)
|
|
||||||
for (const entry of originalEntries) {
|
|
||||||
addConsole(entry)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// eslint-disable-next-line react-hooks/exhaustive-deps
|
|
||||||
}, [activeWorkflowId, hasHydrated])
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
isExecuting,
|
isExecuting,
|
||||||
isDebugging,
|
isDebugging,
|
||||||
|
|||||||
@@ -196,8 +196,6 @@ export interface SubBlockConfig {
|
|||||||
type: SubBlockType
|
type: SubBlockType
|
||||||
mode?: 'basic' | 'advanced' | 'both' | 'trigger' // Default is 'both' if not specified. 'trigger' means only shown in trigger mode
|
mode?: 'basic' | 'advanced' | 'both' | 'trigger' // Default is 'both' if not specified. 'trigger' means only shown in trigger mode
|
||||||
canonicalParamId?: string
|
canonicalParamId?: string
|
||||||
/** Controls parameter visibility in agent/tool-input context */
|
|
||||||
paramVisibility?: 'user-or-llm' | 'user-only' | 'llm-only' | 'hidden'
|
|
||||||
required?:
|
required?:
|
||||||
| boolean
|
| boolean
|
||||||
| {
|
| {
|
||||||
|
|||||||
@@ -423,7 +423,7 @@ interface GenerateVersionDescriptionVariables {
|
|||||||
|
|
||||||
const VERSION_DESCRIPTION_SYSTEM_PROMPT = `You are writing deployment version descriptions for a workflow automation platform.
|
const VERSION_DESCRIPTION_SYSTEM_PROMPT = `You are writing deployment version descriptions for a workflow automation platform.
|
||||||
|
|
||||||
Write a brief, factual description (1-3 sentences, under 2000 characters) that states what changed between versions.
|
Write a brief, factual description (1-3 sentences, under 400 characters) that states what changed between versions.
|
||||||
|
|
||||||
Guidelines:
|
Guidelines:
|
||||||
- Use the specific values provided (credential names, channel names, model names)
|
- Use the specific values provided (credential names, channel names, model names)
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { useCallback } from 'react'
|
import { useCallback, useRef } from 'react'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import type {
|
import type {
|
||||||
BlockCompletedData,
|
BlockCompletedData,
|
||||||
@@ -16,18 +16,6 @@ import type { SerializableExecutionState } from '@/executor/execution/types'
|
|||||||
|
|
||||||
const logger = createLogger('useExecutionStream')
|
const logger = createLogger('useExecutionStream')
|
||||||
|
|
||||||
/**
|
|
||||||
* Detects errors caused by the browser killing a fetch (page refresh, navigation, tab close).
|
|
||||||
* These should be treated as clean disconnects, not execution errors.
|
|
||||||
*/
|
|
||||||
function isClientDisconnectError(error: any): boolean {
|
|
||||||
if (error.name === 'AbortError') return true
|
|
||||||
const msg = (error.message ?? '').toLowerCase()
|
|
||||||
return (
|
|
||||||
msg.includes('network error') || msg.includes('failed to fetch') || msg.includes('load failed')
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Processes SSE events from a response body and invokes appropriate callbacks.
|
* Processes SSE events from a response body and invokes appropriate callbacks.
|
||||||
*/
|
*/
|
||||||
@@ -133,7 +121,6 @@ export interface ExecuteStreamOptions {
|
|||||||
parallels?: Record<string, any>
|
parallels?: Record<string, any>
|
||||||
}
|
}
|
||||||
stopAfterBlockId?: string
|
stopAfterBlockId?: string
|
||||||
onExecutionId?: (executionId: string) => void
|
|
||||||
callbacks?: ExecutionStreamCallbacks
|
callbacks?: ExecutionStreamCallbacks
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -142,40 +129,30 @@ export interface ExecuteFromBlockOptions {
|
|||||||
startBlockId: string
|
startBlockId: string
|
||||||
sourceSnapshot: SerializableExecutionState
|
sourceSnapshot: SerializableExecutionState
|
||||||
input?: any
|
input?: any
|
||||||
onExecutionId?: (executionId: string) => void
|
|
||||||
callbacks?: ExecutionStreamCallbacks
|
callbacks?: ExecutionStreamCallbacks
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface ReconnectStreamOptions {
|
|
||||||
workflowId: string
|
|
||||||
executionId: string
|
|
||||||
fromEventId?: number
|
|
||||||
callbacks?: ExecutionStreamCallbacks
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Module-level map shared across all hook instances.
|
|
||||||
* Ensures ANY instance can cancel streams started by ANY other instance,
|
|
||||||
* which is critical for SPA navigation where the original hook instance unmounts
|
|
||||||
* but the SSE stream must be cancellable from the new instance.
|
|
||||||
*/
|
|
||||||
const sharedAbortControllers = new Map<string, AbortController>()
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Hook for executing workflows via server-side SSE streaming.
|
* Hook for executing workflows via server-side SSE streaming.
|
||||||
* Supports concurrent executions via per-workflow AbortController maps.
|
* Supports concurrent executions via per-workflow AbortController maps.
|
||||||
*/
|
*/
|
||||||
export function useExecutionStream() {
|
export function useExecutionStream() {
|
||||||
const execute = useCallback(async (options: ExecuteStreamOptions) => {
|
const abortControllersRef = useRef<Map<string, AbortController>>(new Map())
|
||||||
const { workflowId, callbacks = {}, onExecutionId, ...payload } = options
|
const currentExecutionsRef = useRef<Map<string, { workflowId: string; executionId: string }>>(
|
||||||
|
new Map()
|
||||||
|
)
|
||||||
|
|
||||||
const existing = sharedAbortControllers.get(workflowId)
|
const execute = useCallback(async (options: ExecuteStreamOptions) => {
|
||||||
|
const { workflowId, callbacks = {}, ...payload } = options
|
||||||
|
|
||||||
|
const existing = abortControllersRef.current.get(workflowId)
|
||||||
if (existing) {
|
if (existing) {
|
||||||
existing.abort()
|
existing.abort()
|
||||||
}
|
}
|
||||||
|
|
||||||
const abortController = new AbortController()
|
const abortController = new AbortController()
|
||||||
sharedAbortControllers.set(workflowId, abortController)
|
abortControllersRef.current.set(workflowId, abortController)
|
||||||
|
currentExecutionsRef.current.delete(workflowId)
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
||||||
@@ -200,48 +177,42 @@ export function useExecutionStream() {
|
|||||||
throw new Error('No response body')
|
throw new Error('No response body')
|
||||||
}
|
}
|
||||||
|
|
||||||
const serverExecutionId = response.headers.get('X-Execution-Id')
|
const executionId = response.headers.get('X-Execution-Id')
|
||||||
if (serverExecutionId) {
|
if (executionId) {
|
||||||
onExecutionId?.(serverExecutionId)
|
currentExecutionsRef.current.set(workflowId, { workflowId, executionId })
|
||||||
}
|
}
|
||||||
|
|
||||||
const reader = response.body.getReader()
|
const reader = response.body.getReader()
|
||||||
await processSSEStream(reader, callbacks, 'Execution')
|
await processSSEStream(reader, callbacks, 'Execution')
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
if (isClientDisconnectError(error)) {
|
if (error.name === 'AbortError') {
|
||||||
logger.info('Execution stream disconnected (page unload or abort)')
|
logger.info('Execution stream cancelled')
|
||||||
return
|
callbacks.onExecutionCancelled?.({ duration: 0 })
|
||||||
|
} else {
|
||||||
|
logger.error('Execution stream error:', error)
|
||||||
|
callbacks.onExecutionError?.({
|
||||||
|
error: error.message || 'Unknown error',
|
||||||
|
duration: 0,
|
||||||
|
})
|
||||||
}
|
}
|
||||||
logger.error('Execution stream error:', error)
|
|
||||||
callbacks.onExecutionError?.({
|
|
||||||
error: error.message || 'Unknown error',
|
|
||||||
duration: 0,
|
|
||||||
})
|
|
||||||
throw error
|
throw error
|
||||||
} finally {
|
} finally {
|
||||||
if (sharedAbortControllers.get(workflowId) === abortController) {
|
abortControllersRef.current.delete(workflowId)
|
||||||
sharedAbortControllers.delete(workflowId)
|
currentExecutionsRef.current.delete(workflowId)
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}, [])
|
}, [])
|
||||||
|
|
||||||
const executeFromBlock = useCallback(async (options: ExecuteFromBlockOptions) => {
|
const executeFromBlock = useCallback(async (options: ExecuteFromBlockOptions) => {
|
||||||
const {
|
const { workflowId, startBlockId, sourceSnapshot, input, callbacks = {} } = options
|
||||||
workflowId,
|
|
||||||
startBlockId,
|
|
||||||
sourceSnapshot,
|
|
||||||
input,
|
|
||||||
onExecutionId,
|
|
||||||
callbacks = {},
|
|
||||||
} = options
|
|
||||||
|
|
||||||
const existing = sharedAbortControllers.get(workflowId)
|
const existing = abortControllersRef.current.get(workflowId)
|
||||||
if (existing) {
|
if (existing) {
|
||||||
existing.abort()
|
existing.abort()
|
||||||
}
|
}
|
||||||
|
|
||||||
const abortController = new AbortController()
|
const abortController = new AbortController()
|
||||||
sharedAbortControllers.set(workflowId, abortController)
|
abortControllersRef.current.set(workflowId, abortController)
|
||||||
|
currentExecutionsRef.current.delete(workflowId)
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
||||||
@@ -275,80 +246,64 @@ export function useExecutionStream() {
|
|||||||
throw new Error('No response body')
|
throw new Error('No response body')
|
||||||
}
|
}
|
||||||
|
|
||||||
const serverExecutionId = response.headers.get('X-Execution-Id')
|
const executionId = response.headers.get('X-Execution-Id')
|
||||||
if (serverExecutionId) {
|
if (executionId) {
|
||||||
onExecutionId?.(serverExecutionId)
|
currentExecutionsRef.current.set(workflowId, { workflowId, executionId })
|
||||||
}
|
}
|
||||||
|
|
||||||
const reader = response.body.getReader()
|
const reader = response.body.getReader()
|
||||||
await processSSEStream(reader, callbacks, 'Run-from-block')
|
await processSSEStream(reader, callbacks, 'Run-from-block')
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
if (isClientDisconnectError(error)) {
|
if (error.name === 'AbortError') {
|
||||||
logger.info('Run-from-block stream disconnected (page unload or abort)')
|
logger.info('Run-from-block execution cancelled')
|
||||||
return
|
callbacks.onExecutionCancelled?.({ duration: 0 })
|
||||||
|
} else {
|
||||||
|
logger.error('Run-from-block execution error:', error)
|
||||||
|
callbacks.onExecutionError?.({
|
||||||
|
error: error.message || 'Unknown error',
|
||||||
|
duration: 0,
|
||||||
|
})
|
||||||
}
|
}
|
||||||
logger.error('Run-from-block execution error:', error)
|
|
||||||
callbacks.onExecutionError?.({
|
|
||||||
error: error.message || 'Unknown error',
|
|
||||||
duration: 0,
|
|
||||||
})
|
|
||||||
throw error
|
throw error
|
||||||
} finally {
|
} finally {
|
||||||
if (sharedAbortControllers.get(workflowId) === abortController) {
|
abortControllersRef.current.delete(workflowId)
|
||||||
sharedAbortControllers.delete(workflowId)
|
currentExecutionsRef.current.delete(workflowId)
|
||||||
}
|
|
||||||
}
|
|
||||||
}, [])
|
|
||||||
|
|
||||||
const reconnect = useCallback(async (options: ReconnectStreamOptions) => {
|
|
||||||
const { workflowId, executionId, fromEventId = 0, callbacks = {} } = options
|
|
||||||
|
|
||||||
const existing = sharedAbortControllers.get(workflowId)
|
|
||||||
if (existing) {
|
|
||||||
existing.abort()
|
|
||||||
}
|
|
||||||
|
|
||||||
const abortController = new AbortController()
|
|
||||||
sharedAbortControllers.set(workflowId, abortController)
|
|
||||||
try {
|
|
||||||
const response = await fetch(
|
|
||||||
`/api/workflows/${workflowId}/executions/${executionId}/stream?from=${fromEventId}`,
|
|
||||||
{ signal: abortController.signal }
|
|
||||||
)
|
|
||||||
if (!response.ok) throw new Error(`Reconnect failed (${response.status})`)
|
|
||||||
if (!response.body) throw new Error('No response body')
|
|
||||||
|
|
||||||
await processSSEStream(response.body.getReader(), callbacks, 'Reconnect')
|
|
||||||
} catch (error: any) {
|
|
||||||
if (isClientDisconnectError(error)) return
|
|
||||||
logger.error('Reconnection stream error:', error)
|
|
||||||
throw error
|
|
||||||
} finally {
|
|
||||||
if (sharedAbortControllers.get(workflowId) === abortController) {
|
|
||||||
sharedAbortControllers.delete(workflowId)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}, [])
|
}, [])
|
||||||
|
|
||||||
const cancel = useCallback((workflowId?: string) => {
|
const cancel = useCallback((workflowId?: string) => {
|
||||||
if (workflowId) {
|
if (workflowId) {
|
||||||
const controller = sharedAbortControllers.get(workflowId)
|
const execution = currentExecutionsRef.current.get(workflowId)
|
||||||
|
if (execution) {
|
||||||
|
fetch(`/api/workflows/${execution.workflowId}/executions/${execution.executionId}/cancel`, {
|
||||||
|
method: 'POST',
|
||||||
|
}).catch(() => {})
|
||||||
|
}
|
||||||
|
|
||||||
|
const controller = abortControllersRef.current.get(workflowId)
|
||||||
if (controller) {
|
if (controller) {
|
||||||
controller.abort()
|
controller.abort()
|
||||||
sharedAbortControllers.delete(workflowId)
|
abortControllersRef.current.delete(workflowId)
|
||||||
}
|
}
|
||||||
|
currentExecutionsRef.current.delete(workflowId)
|
||||||
} else {
|
} else {
|
||||||
for (const [, controller] of sharedAbortControllers) {
|
for (const [, execution] of currentExecutionsRef.current) {
|
||||||
|
fetch(`/api/workflows/${execution.workflowId}/executions/${execution.executionId}/cancel`, {
|
||||||
|
method: 'POST',
|
||||||
|
}).catch(() => {})
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const [, controller] of abortControllersRef.current) {
|
||||||
controller.abort()
|
controller.abort()
|
||||||
}
|
}
|
||||||
sharedAbortControllers.clear()
|
abortControllersRef.current.clear()
|
||||||
|
currentExecutionsRef.current.clear()
|
||||||
}
|
}
|
||||||
}, [])
|
}, [])
|
||||||
|
|
||||||
return {
|
return {
|
||||||
execute,
|
execute,
|
||||||
executeFromBlock,
|
executeFromBlock,
|
||||||
reconnect,
|
|
||||||
cancel,
|
cancel,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -20,8 +20,6 @@ export interface BuildPayloadParams {
|
|||||||
fileAttachments?: Array<{ id: string; key: string; size: number; [key: string]: unknown }>
|
fileAttachments?: Array<{ id: string; key: string; size: number; [key: string]: unknown }>
|
||||||
commands?: string[]
|
commands?: string[]
|
||||||
chatId?: string
|
chatId?: string
|
||||||
conversationId?: string
|
|
||||||
prefetch?: boolean
|
|
||||||
implicitFeedback?: string
|
implicitFeedback?: string
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -66,10 +64,6 @@ export async function buildCopilotRequestPayload(
|
|||||||
fileAttachments,
|
fileAttachments,
|
||||||
commands,
|
commands,
|
||||||
chatId,
|
chatId,
|
||||||
conversationId,
|
|
||||||
prefetch,
|
|
||||||
conversationHistory,
|
|
||||||
implicitFeedback,
|
|
||||||
} = params
|
} = params
|
||||||
|
|
||||||
const selectedModel = options.selectedModel
|
const selectedModel = options.selectedModel
|
||||||
@@ -160,12 +154,6 @@ export async function buildCopilotRequestPayload(
|
|||||||
version: SIM_AGENT_VERSION,
|
version: SIM_AGENT_VERSION,
|
||||||
...(contexts && contexts.length > 0 ? { context: contexts } : {}),
|
...(contexts && contexts.length > 0 ? { context: contexts } : {}),
|
||||||
...(chatId ? { chatId } : {}),
|
...(chatId ? { chatId } : {}),
|
||||||
...(conversationId ? { conversationId } : {}),
|
|
||||||
...(Array.isArray(conversationHistory) && conversationHistory.length > 0
|
|
||||||
? { conversationHistory }
|
|
||||||
: {}),
|
|
||||||
...(typeof prefetch === 'boolean' ? { prefetch } : {}),
|
|
||||||
...(implicitFeedback ? { implicitFeedback } : {}),
|
|
||||||
...(processedFileContents.length > 0 ? { fileAttachments: processedFileContents } : {}),
|
...(processedFileContents.length > 0 ? { fileAttachments: processedFileContents } : {}),
|
||||||
...(integrationTools.length > 0 ? { integrationTools } : {}),
|
...(integrationTools.length > 0 ? { integrationTools } : {}),
|
||||||
...(credentials ? { credentials } : {}),
|
...(credentials ? { credentials } : {}),
|
||||||
|
|||||||
@@ -1,246 +0,0 @@
|
|||||||
import { createLogger } from '@sim/logger'
|
|
||||||
import { getRedisClient } from '@/lib/core/config/redis'
|
|
||||||
import type { ExecutionEvent } from '@/lib/workflows/executor/execution-events'
|
|
||||||
|
|
||||||
const logger = createLogger('ExecutionEventBuffer')
|
|
||||||
|
|
||||||
const REDIS_PREFIX = 'execution:stream:'
|
|
||||||
const TTL_SECONDS = 60 * 60 // 1 hour
|
|
||||||
const EVENT_LIMIT = 1000
|
|
||||||
const RESERVE_BATCH = 100
|
|
||||||
const FLUSH_INTERVAL_MS = 15
|
|
||||||
const FLUSH_MAX_BATCH = 200
|
|
||||||
|
|
||||||
function getEventsKey(executionId: string) {
|
|
||||||
return `${REDIS_PREFIX}${executionId}:events`
|
|
||||||
}
|
|
||||||
|
|
||||||
function getSeqKey(executionId: string) {
|
|
||||||
return `${REDIS_PREFIX}${executionId}:seq`
|
|
||||||
}
|
|
||||||
|
|
||||||
function getMetaKey(executionId: string) {
|
|
||||||
return `${REDIS_PREFIX}${executionId}:meta`
|
|
||||||
}
|
|
||||||
|
|
||||||
export type ExecutionStreamStatus = 'active' | 'complete' | 'error' | 'cancelled'
|
|
||||||
|
|
||||||
export interface ExecutionStreamMeta {
|
|
||||||
status: ExecutionStreamStatus
|
|
||||||
userId?: string
|
|
||||||
workflowId?: string
|
|
||||||
updatedAt?: string
|
|
||||||
}
|
|
||||||
|
|
||||||
export interface ExecutionEventEntry {
|
|
||||||
eventId: number
|
|
||||||
executionId: string
|
|
||||||
event: ExecutionEvent
|
|
||||||
}
|
|
||||||
|
|
||||||
export interface ExecutionEventWriter {
|
|
||||||
write: (event: ExecutionEvent) => Promise<ExecutionEventEntry>
|
|
||||||
flush: () => Promise<void>
|
|
||||||
close: () => Promise<void>
|
|
||||||
}
|
|
||||||
|
|
||||||
export async function setExecutionMeta(
|
|
||||||
executionId: string,
|
|
||||||
meta: Partial<ExecutionStreamMeta>
|
|
||||||
): Promise<void> {
|
|
||||||
const redis = getRedisClient()
|
|
||||||
if (!redis) {
|
|
||||||
logger.warn('setExecutionMeta: Redis client unavailable', { executionId })
|
|
||||||
return
|
|
||||||
}
|
|
||||||
try {
|
|
||||||
const key = getMetaKey(executionId)
|
|
||||||
const payload: Record<string, string> = {
|
|
||||||
updatedAt: new Date().toISOString(),
|
|
||||||
}
|
|
||||||
if (meta.status) payload.status = meta.status
|
|
||||||
if (meta.userId) payload.userId = meta.userId
|
|
||||||
if (meta.workflowId) payload.workflowId = meta.workflowId
|
|
||||||
await redis.hset(key, payload)
|
|
||||||
await redis.expire(key, TTL_SECONDS)
|
|
||||||
} catch (error) {
|
|
||||||
logger.warn('Failed to update execution meta', {
|
|
||||||
executionId,
|
|
||||||
error: error instanceof Error ? error.message : String(error),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
export async function getExecutionMeta(executionId: string): Promise<ExecutionStreamMeta | null> {
|
|
||||||
const redis = getRedisClient()
|
|
||||||
if (!redis) {
|
|
||||||
logger.warn('getExecutionMeta: Redis client unavailable', { executionId })
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
try {
|
|
||||||
const key = getMetaKey(executionId)
|
|
||||||
const meta = await redis.hgetall(key)
|
|
||||||
if (!meta || Object.keys(meta).length === 0) return null
|
|
||||||
return meta as unknown as ExecutionStreamMeta
|
|
||||||
} catch (error) {
|
|
||||||
logger.warn('Failed to read execution meta', {
|
|
||||||
executionId,
|
|
||||||
error: error instanceof Error ? error.message : String(error),
|
|
||||||
})
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
export async function readExecutionEvents(
|
|
||||||
executionId: string,
|
|
||||||
afterEventId: number
|
|
||||||
): Promise<ExecutionEventEntry[]> {
|
|
||||||
const redis = getRedisClient()
|
|
||||||
if (!redis) return []
|
|
||||||
try {
|
|
||||||
const raw = await redis.zrangebyscore(getEventsKey(executionId), afterEventId + 1, '+inf')
|
|
||||||
return raw
|
|
||||||
.map((entry) => {
|
|
||||||
try {
|
|
||||||
return JSON.parse(entry) as ExecutionEventEntry
|
|
||||||
} catch {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
})
|
|
||||||
.filter((entry): entry is ExecutionEventEntry => Boolean(entry))
|
|
||||||
} catch (error) {
|
|
||||||
logger.warn('Failed to read execution events', {
|
|
||||||
executionId,
|
|
||||||
error: error instanceof Error ? error.message : String(error),
|
|
||||||
})
|
|
||||||
return []
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
export function createExecutionEventWriter(executionId: string): ExecutionEventWriter {
|
|
||||||
const redis = getRedisClient()
|
|
||||||
if (!redis) {
|
|
||||||
logger.warn(
|
|
||||||
'createExecutionEventWriter: Redis client unavailable, events will not be buffered',
|
|
||||||
{
|
|
||||||
executionId,
|
|
||||||
}
|
|
||||||
)
|
|
||||||
return {
|
|
||||||
write: async (event) => ({ eventId: 0, executionId, event }),
|
|
||||||
flush: async () => {},
|
|
||||||
close: async () => {},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let pending: ExecutionEventEntry[] = []
|
|
||||||
let nextEventId = 0
|
|
||||||
let maxReservedId = 0
|
|
||||||
let flushTimer: ReturnType<typeof setTimeout> | null = null
|
|
||||||
|
|
||||||
const scheduleFlush = () => {
|
|
||||||
if (flushTimer) return
|
|
||||||
flushTimer = setTimeout(() => {
|
|
||||||
flushTimer = null
|
|
||||||
void flush()
|
|
||||||
}, FLUSH_INTERVAL_MS)
|
|
||||||
}
|
|
||||||
|
|
||||||
const reserveIds = async (minCount: number) => {
|
|
||||||
const reserveCount = Math.max(RESERVE_BATCH, minCount)
|
|
||||||
const newMax = await redis.incrby(getSeqKey(executionId), reserveCount)
|
|
||||||
const startId = newMax - reserveCount + 1
|
|
||||||
if (nextEventId === 0 || nextEventId > maxReservedId) {
|
|
||||||
nextEventId = startId
|
|
||||||
maxReservedId = newMax
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let flushPromise: Promise<void> | null = null
|
|
||||||
let closed = false
|
|
||||||
const inflightWrites = new Set<Promise<ExecutionEventEntry>>()
|
|
||||||
|
|
||||||
const doFlush = async () => {
|
|
||||||
if (pending.length === 0) return
|
|
||||||
const batch = pending
|
|
||||||
pending = []
|
|
||||||
try {
|
|
||||||
const key = getEventsKey(executionId)
|
|
||||||
const zaddArgs: (string | number)[] = []
|
|
||||||
for (const entry of batch) {
|
|
||||||
zaddArgs.push(entry.eventId, JSON.stringify(entry))
|
|
||||||
}
|
|
||||||
const pipeline = redis.pipeline()
|
|
||||||
pipeline.zadd(key, ...zaddArgs)
|
|
||||||
pipeline.expire(key, TTL_SECONDS)
|
|
||||||
pipeline.expire(getSeqKey(executionId), TTL_SECONDS)
|
|
||||||
pipeline.zremrangebyrank(key, 0, -EVENT_LIMIT - 1)
|
|
||||||
await pipeline.exec()
|
|
||||||
} catch (error) {
|
|
||||||
logger.warn('Failed to flush execution events', {
|
|
||||||
executionId,
|
|
||||||
batchSize: batch.length,
|
|
||||||
error: error instanceof Error ? error.message : String(error),
|
|
||||||
stack: error instanceof Error ? error.stack : undefined,
|
|
||||||
})
|
|
||||||
pending = batch.concat(pending)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const flush = async () => {
|
|
||||||
if (flushPromise) {
|
|
||||||
await flushPromise
|
|
||||||
return
|
|
||||||
}
|
|
||||||
flushPromise = doFlush()
|
|
||||||
try {
|
|
||||||
await flushPromise
|
|
||||||
} finally {
|
|
||||||
flushPromise = null
|
|
||||||
if (pending.length > 0) scheduleFlush()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const writeCore = async (event: ExecutionEvent): Promise<ExecutionEventEntry> => {
|
|
||||||
if (closed) return { eventId: 0, executionId, event }
|
|
||||||
if (nextEventId === 0 || nextEventId > maxReservedId) {
|
|
||||||
await reserveIds(1)
|
|
||||||
}
|
|
||||||
const eventId = nextEventId++
|
|
||||||
const entry: ExecutionEventEntry = { eventId, executionId, event }
|
|
||||||
pending.push(entry)
|
|
||||||
if (pending.length >= FLUSH_MAX_BATCH) {
|
|
||||||
await flush()
|
|
||||||
} else {
|
|
||||||
scheduleFlush()
|
|
||||||
}
|
|
||||||
return entry
|
|
||||||
}
|
|
||||||
|
|
||||||
const write = (event: ExecutionEvent): Promise<ExecutionEventEntry> => {
|
|
||||||
const p = writeCore(event)
|
|
||||||
inflightWrites.add(p)
|
|
||||||
const remove = () => inflightWrites.delete(p)
|
|
||||||
p.then(remove, remove)
|
|
||||||
return p
|
|
||||||
}
|
|
||||||
|
|
||||||
const close = async () => {
|
|
||||||
closed = true
|
|
||||||
if (flushTimer) {
|
|
||||||
clearTimeout(flushTimer)
|
|
||||||
flushTimer = null
|
|
||||||
}
|
|
||||||
if (inflightWrites.size > 0) {
|
|
||||||
await Promise.allSettled(inflightWrites)
|
|
||||||
}
|
|
||||||
if (flushPromise) {
|
|
||||||
await flushPromise
|
|
||||||
}
|
|
||||||
if (pending.length > 0) {
|
|
||||||
await doFlush()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return { write, flush, close }
|
|
||||||
}
|
|
||||||
@@ -2364,261 +2364,6 @@ describe('hasWorkflowChanged', () => {
|
|||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe('Trigger Config Normalization (False Positive Prevention)', () => {
|
|
||||||
it.concurrent(
|
|
||||||
'should not detect change when deployed has null fields but current has values from triggerConfig',
|
|
||||||
() => {
|
|
||||||
// Core scenario: deployed state has null individual fields, current state has
|
|
||||||
// values populated from triggerConfig at runtime by populateTriggerFieldsFromConfig
|
|
||||||
const deployedState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
botToken: { id: 'botToken', type: 'short-input', value: null },
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const currentState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
|
||||||
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
it.concurrent(
|
|
||||||
'should detect change when user edits a trigger field to a different value',
|
|
||||||
() => {
|
|
||||||
const deployedState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'old-secret' },
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const currentState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'new-secret' },
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'old-secret' },
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(true)
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
it.concurrent('should not detect change when both sides have no triggerConfig', () => {
|
|
||||||
const deployedState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const currentState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent(
|
|
||||||
'should not detect change when deployed has empty fields and triggerConfig populates them',
|
|
||||||
() => {
|
|
||||||
// Empty string is also treated as "empty" by normalizeTriggerConfigValues
|
|
||||||
const deployedState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: '' },
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123' },
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const currentState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123' },
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
it.concurrent('should not detect change when triggerId differs', () => {
|
|
||||||
const deployedState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
model: { value: 'gpt-4' },
|
|
||||||
triggerId: { value: null },
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const currentState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
model: { value: 'gpt-4' },
|
|
||||||
triggerId: { value: 'slack_webhook' },
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent(
|
|
||||||
'should not detect change for namespaced system subBlock IDs like samplePayload_slack_webhook',
|
|
||||||
() => {
|
|
||||||
const deployedState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
model: { value: 'gpt-4' },
|
|
||||||
samplePayload_slack_webhook: { value: 'old payload' },
|
|
||||||
triggerInstructions_slack_webhook: { value: 'old instructions' },
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const currentState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
model: { value: 'gpt-4' },
|
|
||||||
samplePayload_slack_webhook: { value: 'new payload' },
|
|
||||||
triggerInstructions_slack_webhook: { value: 'new instructions' },
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
it.concurrent(
|
|
||||||
'should handle mixed scenario: some fields from triggerConfig, some user-edited',
|
|
||||||
() => {
|
|
||||||
const deployedState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
botToken: { id: 'botToken', type: 'short-input', value: null },
|
|
||||||
includeFiles: { id: 'includeFiles', type: 'switch', value: false },
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const currentState = createWorkflowState({
|
|
||||||
blocks: {
|
|
||||||
block1: createBlock('block1', {
|
|
||||||
type: 'starter',
|
|
||||||
subBlocks: {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
|
||||||
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
|
||||||
includeFiles: { id: 'includeFiles', type: 'switch', value: true },
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
// includeFiles changed from false to true — this IS a real change
|
|
||||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(true)
|
|
||||||
}
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
describe('Trigger Runtime Metadata (Should Not Trigger Change)', () => {
|
describe('Trigger Runtime Metadata (Should Not Trigger Change)', () => {
|
||||||
it.concurrent('should not detect change when webhookId differs', () => {
|
it.concurrent('should not detect change when webhookId differs', () => {
|
||||||
const deployedState = createWorkflowState({
|
const deployedState = createWorkflowState({
|
||||||
|
|||||||
@@ -9,7 +9,6 @@ import {
|
|||||||
normalizeLoop,
|
normalizeLoop,
|
||||||
normalizeParallel,
|
normalizeParallel,
|
||||||
normalizeSubBlockValue,
|
normalizeSubBlockValue,
|
||||||
normalizeTriggerConfigValues,
|
|
||||||
normalizeValue,
|
normalizeValue,
|
||||||
normalizeVariables,
|
normalizeVariables,
|
||||||
sanitizeVariable,
|
sanitizeVariable,
|
||||||
@@ -173,18 +172,14 @@ export function generateWorkflowDiffSummary(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Normalize trigger config values for both states before comparison
|
|
||||||
const normalizedCurrentSubs = normalizeTriggerConfigValues(currentSubBlocks)
|
|
||||||
const normalizedPreviousSubs = normalizeTriggerConfigValues(previousSubBlocks)
|
|
||||||
|
|
||||||
// Compare subBlocks using shared helper for filtering (single source of truth)
|
// Compare subBlocks using shared helper for filtering (single source of truth)
|
||||||
const allSubBlockIds = filterSubBlockIds([
|
const allSubBlockIds = filterSubBlockIds([
|
||||||
...new Set([...Object.keys(normalizedCurrentSubs), ...Object.keys(normalizedPreviousSubs)]),
|
...new Set([...Object.keys(currentSubBlocks), ...Object.keys(previousSubBlocks)]),
|
||||||
])
|
])
|
||||||
|
|
||||||
for (const subId of allSubBlockIds) {
|
for (const subId of allSubBlockIds) {
|
||||||
const currentSub = normalizedCurrentSubs[subId] as Record<string, unknown> | undefined
|
const currentSub = currentSubBlocks[subId] as Record<string, unknown> | undefined
|
||||||
const previousSub = normalizedPreviousSubs[subId] as Record<string, unknown> | undefined
|
const previousSub = previousSubBlocks[subId] as Record<string, unknown> | undefined
|
||||||
|
|
||||||
if (!currentSub || !previousSub) {
|
if (!currentSub || !previousSub) {
|
||||||
changes.push({
|
changes.push({
|
||||||
|
|||||||
@@ -4,12 +4,10 @@
|
|||||||
import { describe, expect, it } from 'vitest'
|
import { describe, expect, it } from 'vitest'
|
||||||
import type { Loop, Parallel } from '@/stores/workflows/workflow/types'
|
import type { Loop, Parallel } from '@/stores/workflows/workflow/types'
|
||||||
import {
|
import {
|
||||||
filterSubBlockIds,
|
|
||||||
normalizedStringify,
|
normalizedStringify,
|
||||||
normalizeEdge,
|
normalizeEdge,
|
||||||
normalizeLoop,
|
normalizeLoop,
|
||||||
normalizeParallel,
|
normalizeParallel,
|
||||||
normalizeTriggerConfigValues,
|
|
||||||
normalizeValue,
|
normalizeValue,
|
||||||
sanitizeInputFormat,
|
sanitizeInputFormat,
|
||||||
sanitizeTools,
|
sanitizeTools,
|
||||||
@@ -586,214 +584,4 @@ describe('Workflow Normalization Utilities', () => {
|
|||||||
expect(result2).toBe(result3)
|
expect(result2).toBe(result3)
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe('filterSubBlockIds', () => {
|
|
||||||
it.concurrent('should exclude exact SYSTEM_SUBBLOCK_IDS', () => {
|
|
||||||
const ids = ['signingSecret', 'samplePayload', 'triggerInstructions', 'botToken']
|
|
||||||
const result = filterSubBlockIds(ids)
|
|
||||||
expect(result).toEqual(['botToken', 'signingSecret'])
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should exclude namespaced SYSTEM_SUBBLOCK_IDS (prefix matching)', () => {
|
|
||||||
const ids = [
|
|
||||||
'signingSecret',
|
|
||||||
'samplePayload_slack_webhook',
|
|
||||||
'triggerInstructions_slack_webhook',
|
|
||||||
'webhookUrlDisplay_slack_webhook',
|
|
||||||
'botToken',
|
|
||||||
]
|
|
||||||
const result = filterSubBlockIds(ids)
|
|
||||||
expect(result).toEqual(['botToken', 'signingSecret'])
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should exclude exact TRIGGER_RUNTIME_SUBBLOCK_IDS', () => {
|
|
||||||
const ids = ['webhookId', 'triggerPath', 'triggerConfig', 'triggerId', 'signingSecret']
|
|
||||||
const result = filterSubBlockIds(ids)
|
|
||||||
expect(result).toEqual(['signingSecret'])
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should not exclude IDs that merely contain a system ID substring', () => {
|
|
||||||
const ids = ['mySamplePayload', 'notSamplePayload']
|
|
||||||
const result = filterSubBlockIds(ids)
|
|
||||||
expect(result).toEqual(['mySamplePayload', 'notSamplePayload'])
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should return sorted results', () => {
|
|
||||||
const ids = ['zebra', 'alpha', 'middle']
|
|
||||||
const result = filterSubBlockIds(ids)
|
|
||||||
expect(result).toEqual(['alpha', 'middle', 'zebra'])
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should handle empty array', () => {
|
|
||||||
expect(filterSubBlockIds([])).toEqual([])
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should handle all IDs being excluded', () => {
|
|
||||||
const ids = ['webhookId', 'triggerPath', 'samplePayload', 'triggerConfig']
|
|
||||||
const result = filterSubBlockIds(ids)
|
|
||||||
expect(result).toEqual([])
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should exclude setupScript and scheduleInfo namespaced variants', () => {
|
|
||||||
const ids = ['setupScript_google_sheets_row', 'scheduleInfo_cron_trigger', 'realField']
|
|
||||||
const result = filterSubBlockIds(ids)
|
|
||||||
expect(result).toEqual(['realField'])
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should exclude triggerCredentials namespaced variants', () => {
|
|
||||||
const ids = ['triggerCredentials_slack_webhook', 'signingSecret']
|
|
||||||
const result = filterSubBlockIds(ids)
|
|
||||||
expect(result).toEqual(['signingSecret'])
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe('normalizeTriggerConfigValues', () => {
|
|
||||||
it.concurrent('should return subBlocks unchanged when no triggerConfig exists', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
|
||||||
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect(result).toEqual(subBlocks)
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should return subBlocks unchanged when triggerConfig value is null', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: { id: 'triggerConfig', type: 'short-input', value: null },
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect(result).toEqual(subBlocks)
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent(
|
|
||||||
'should return subBlocks unchanged when triggerConfig value is not an object',
|
|
||||||
() => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: { id: 'triggerConfig', type: 'short-input', value: 'string-value' },
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect(result).toEqual(subBlocks)
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
it.concurrent('should populate null individual fields from triggerConfig', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
|
||||||
},
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
botToken: { id: 'botToken', type: 'short-input', value: null },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
|
||||||
expect((result.botToken as Record<string, unknown>).value).toBe('token456')
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should populate undefined individual fields from triggerConfig', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123' },
|
|
||||||
},
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: undefined },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should populate empty string individual fields from triggerConfig', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123' },
|
|
||||||
},
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: '' },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should NOT overwrite existing non-empty individual field values', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'old-secret' },
|
|
||||||
},
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'user-edited-secret' },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect((result.signingSecret as Record<string, unknown>).value).toBe('user-edited-secret')
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should skip triggerConfig fields that are null/undefined', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: null, botToken: undefined },
|
|
||||||
},
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
botToken: { id: 'botToken', type: 'short-input', value: null },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect((result.signingSecret as Record<string, unknown>).value).toBe(null)
|
|
||||||
expect((result.botToken as Record<string, unknown>).value).toBe(null)
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should skip fields from triggerConfig that have no matching subBlock', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { nonExistentField: 'value123' },
|
|
||||||
},
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
expect(result.nonExistentField).toBeUndefined()
|
|
||||||
expect((result.signingSecret as Record<string, unknown>).value).toBe(null)
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should not mutate the original subBlocks object', () => {
|
|
||||||
const original = {
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123' },
|
|
||||||
},
|
|
||||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
|
||||||
}
|
|
||||||
normalizeTriggerConfigValues(original)
|
|
||||||
expect((original.signingSecret as Record<string, unknown>).value).toBe(null)
|
|
||||||
})
|
|
||||||
|
|
||||||
it.concurrent('should preserve other subBlock properties when populating value', () => {
|
|
||||||
const subBlocks = {
|
|
||||||
triggerConfig: {
|
|
||||||
id: 'triggerConfig',
|
|
||||||
type: 'short-input',
|
|
||||||
value: { signingSecret: 'secret123' },
|
|
||||||
},
|
|
||||||
signingSecret: {
|
|
||||||
id: 'signingSecret',
|
|
||||||
type: 'short-input',
|
|
||||||
value: null,
|
|
||||||
placeholder: 'Enter signing secret',
|
|
||||||
},
|
|
||||||
}
|
|
||||||
const result = normalizeTriggerConfigValues(subBlocks)
|
|
||||||
const normalized = result.signingSecret as Record<string, unknown>
|
|
||||||
expect(normalized.value).toBe('secret123')
|
|
||||||
expect(normalized.id).toBe('signingSecret')
|
|
||||||
expect(normalized.type).toBe('short-input')
|
|
||||||
expect(normalized.placeholder).toBe('Enter signing secret')
|
|
||||||
})
|
|
||||||
})
|
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -418,48 +418,10 @@ export function extractBlockFieldsForComparison(block: BlockState): ExtractedBlo
|
|||||||
*/
|
*/
|
||||||
export function filterSubBlockIds(subBlockIds: string[]): string[] {
|
export function filterSubBlockIds(subBlockIds: string[]): string[] {
|
||||||
return subBlockIds
|
return subBlockIds
|
||||||
.filter((id) => {
|
.filter((id) => !SYSTEM_SUBBLOCK_IDS.includes(id) && !TRIGGER_RUNTIME_SUBBLOCK_IDS.includes(id))
|
||||||
if (TRIGGER_RUNTIME_SUBBLOCK_IDS.includes(id)) return false
|
|
||||||
if (SYSTEM_SUBBLOCK_IDS.some((sysId) => id === sysId || id.startsWith(`${sysId}_`)))
|
|
||||||
return false
|
|
||||||
return true
|
|
||||||
})
|
|
||||||
.sort()
|
.sort()
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Normalizes trigger block subBlocks by populating null/empty individual fields
|
|
||||||
* from the triggerConfig aggregate subBlock. This compensates for the runtime
|
|
||||||
* population done by populateTriggerFieldsFromConfig, ensuring consistent
|
|
||||||
* comparison between client state (with populated values) and deployed state
|
|
||||||
* (with null values from DB).
|
|
||||||
*/
|
|
||||||
export function normalizeTriggerConfigValues(
|
|
||||||
subBlocks: Record<string, unknown>
|
|
||||||
): Record<string, unknown> {
|
|
||||||
const triggerConfigSub = subBlocks.triggerConfig as Record<string, unknown> | undefined
|
|
||||||
const triggerConfigValue = triggerConfigSub?.value
|
|
||||||
if (!triggerConfigValue || typeof triggerConfigValue !== 'object') {
|
|
||||||
return subBlocks
|
|
||||||
}
|
|
||||||
|
|
||||||
const result = { ...subBlocks }
|
|
||||||
for (const [fieldId, configValue] of Object.entries(
|
|
||||||
triggerConfigValue as Record<string, unknown>
|
|
||||||
)) {
|
|
||||||
if (configValue === null || configValue === undefined) continue
|
|
||||||
const existingSub = result[fieldId] as Record<string, unknown> | undefined
|
|
||||||
if (
|
|
||||||
existingSub &&
|
|
||||||
(existingSub.value === null || existingSub.value === undefined || existingSub.value === '')
|
|
||||||
) {
|
|
||||||
result[fieldId] = { ...existingSub, value: configValue }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return result
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Normalizes a subBlock value with sanitization for specific subBlock types.
|
* Normalizes a subBlock value with sanitization for specific subBlock types.
|
||||||
* Sanitizes: tools (removes isExpanded), inputFormat (removes collapsed)
|
* Sanitizes: tools (removes isExpanded), inputFormat (removes collapsed)
|
||||||
|
|||||||
@@ -129,18 +129,6 @@ export const useExecutionStore = create<ExecutionState & ExecutionActions>()((se
|
|||||||
})
|
})
|
||||||
},
|
},
|
||||||
|
|
||||||
setCurrentExecutionId: (workflowId, executionId) => {
|
|
||||||
set({
|
|
||||||
workflowExecutions: updatedMap(get().workflowExecutions, workflowId, {
|
|
||||||
currentExecutionId: executionId,
|
|
||||||
}),
|
|
||||||
})
|
|
||||||
},
|
|
||||||
|
|
||||||
getCurrentExecutionId: (workflowId) => {
|
|
||||||
return getOrCreate(get().workflowExecutions, workflowId).currentExecutionId
|
|
||||||
},
|
|
||||||
|
|
||||||
clearRunPath: (workflowId) => {
|
clearRunPath: (workflowId) => {
|
||||||
set({
|
set({
|
||||||
workflowExecutions: updatedMap(get().workflowExecutions, workflowId, {
|
workflowExecutions: updatedMap(get().workflowExecutions, workflowId, {
|
||||||
|
|||||||
@@ -35,8 +35,6 @@ export interface WorkflowExecutionState {
|
|||||||
lastRunPath: Map<string, BlockRunStatus>
|
lastRunPath: Map<string, BlockRunStatus>
|
||||||
/** Maps edge IDs to their run result from the last execution */
|
/** Maps edge IDs to their run result from the last execution */
|
||||||
lastRunEdges: Map<string, EdgeRunStatus>
|
lastRunEdges: Map<string, EdgeRunStatus>
|
||||||
/** The execution ID of the currently running execution */
|
|
||||||
currentExecutionId: string | null
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -56,7 +54,6 @@ export const defaultWorkflowExecutionState: WorkflowExecutionState = {
|
|||||||
debugContext: null,
|
debugContext: null,
|
||||||
lastRunPath: new Map(),
|
lastRunPath: new Map(),
|
||||||
lastRunEdges: new Map(),
|
lastRunEdges: new Map(),
|
||||||
currentExecutionId: null,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -99,10 +96,6 @@ export interface ExecutionActions {
|
|||||||
setEdgeRunStatus: (workflowId: string, edgeId: string, status: EdgeRunStatus) => void
|
setEdgeRunStatus: (workflowId: string, edgeId: string, status: EdgeRunStatus) => void
|
||||||
/** Clears the run path and run edges for a workflow */
|
/** Clears the run path and run edges for a workflow */
|
||||||
clearRunPath: (workflowId: string) => void
|
clearRunPath: (workflowId: string) => void
|
||||||
/** Stores the current execution ID for a workflow */
|
|
||||||
setCurrentExecutionId: (workflowId: string, executionId: string | null) => void
|
|
||||||
/** Returns the current execution ID for a workflow */
|
|
||||||
getCurrentExecutionId: (workflowId: string) => string | null
|
|
||||||
/** Resets the entire store to its initial empty state */
|
/** Resets the entire store to its initial empty state */
|
||||||
reset: () => void
|
reset: () => void
|
||||||
/** Stores a serializable execution snapshot for a workflow */
|
/** Stores a serializable execution snapshot for a workflow */
|
||||||
|
|||||||
@@ -224,7 +224,7 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
|||||||
|
|
||||||
const newEntry = get().entries[0]
|
const newEntry = get().entries[0]
|
||||||
|
|
||||||
if (newEntry?.error && newEntry.blockType !== 'cancelled') {
|
if (newEntry?.error) {
|
||||||
notifyBlockError({
|
notifyBlockError({
|
||||||
error: newEntry.error,
|
error: newEntry.error,
|
||||||
blockName: newEntry.blockName || 'Unknown Block',
|
blockName: newEntry.blockName || 'Unknown Block',
|
||||||
@@ -243,11 +243,6 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
|||||||
useExecutionStore.getState().clearRunPath(workflowId)
|
useExecutionStore.getState().clearRunPath(workflowId)
|
||||||
},
|
},
|
||||||
|
|
||||||
clearExecutionEntries: (executionId: string) =>
|
|
||||||
set((state) => ({
|
|
||||||
entries: state.entries.filter((e) => e.executionId !== executionId),
|
|
||||||
})),
|
|
||||||
|
|
||||||
exportConsoleCSV: (workflowId: string) => {
|
exportConsoleCSV: (workflowId: string) => {
|
||||||
const entries = get().entries.filter((entry) => entry.workflowId === workflowId)
|
const entries = get().entries.filter((entry) => entry.workflowId === workflowId)
|
||||||
|
|
||||||
@@ -475,24 +470,12 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
|||||||
},
|
},
|
||||||
merge: (persistedState, currentState) => {
|
merge: (persistedState, currentState) => {
|
||||||
const persisted = persistedState as Partial<ConsoleStore> | undefined
|
const persisted = persistedState as Partial<ConsoleStore> | undefined
|
||||||
const rawEntries = persisted?.entries ?? currentState.entries
|
const entries = (persisted?.entries ?? currentState.entries).map((entry, index) => {
|
||||||
const oneHourAgo = Date.now() - 60 * 60 * 1000
|
|
||||||
|
|
||||||
const entries = rawEntries.map((entry, index) => {
|
|
||||||
let updated = entry
|
|
||||||
if (entry.executionOrder === undefined) {
|
if (entry.executionOrder === undefined) {
|
||||||
updated = { ...updated, executionOrder: index + 1 }
|
return { ...entry, executionOrder: index + 1 }
|
||||||
}
|
}
|
||||||
if (
|
return entry
|
||||||
entry.isRunning &&
|
|
||||||
entry.startedAt &&
|
|
||||||
new Date(entry.startedAt).getTime() < oneHourAgo
|
|
||||||
) {
|
|
||||||
updated = { ...updated, isRunning: false }
|
|
||||||
}
|
|
||||||
return updated
|
|
||||||
})
|
})
|
||||||
|
|
||||||
return {
|
return {
|
||||||
...currentState,
|
...currentState,
|
||||||
entries,
|
entries,
|
||||||
|
|||||||
@@ -51,7 +51,6 @@ export interface ConsoleStore {
|
|||||||
isOpen: boolean
|
isOpen: boolean
|
||||||
addConsole: (entry: Omit<ConsoleEntry, 'id' | 'timestamp'>) => ConsoleEntry
|
addConsole: (entry: Omit<ConsoleEntry, 'id' | 'timestamp'>) => ConsoleEntry
|
||||||
clearWorkflowConsole: (workflowId: string) => void
|
clearWorkflowConsole: (workflowId: string) => void
|
||||||
clearExecutionEntries: (executionId: string) => void
|
|
||||||
exportConsoleCSV: (workflowId: string) => void
|
exportConsoleCSV: (workflowId: string) => void
|
||||||
getWorkflowEntries: (workflowId: string) => ConsoleEntry[]
|
getWorkflowEntries: (workflowId: string) => ConsoleEntry[]
|
||||||
toggleConsole: () => void
|
toggleConsole: () => void
|
||||||
|
|||||||
@@ -1,7 +1,6 @@
|
|||||||
import {
|
import {
|
||||||
buildCanonicalIndex,
|
buildCanonicalIndex,
|
||||||
type CanonicalIndex,
|
type CanonicalIndex,
|
||||||
type CanonicalModeOverrides,
|
|
||||||
evaluateSubBlockCondition,
|
evaluateSubBlockCondition,
|
||||||
getCanonicalValues,
|
getCanonicalValues,
|
||||||
isCanonicalPair,
|
isCanonicalPair,
|
||||||
@@ -13,10 +12,7 @@ import type { SubBlockConfig as BlockSubBlockConfig } from '@/blocks/types'
|
|||||||
export {
|
export {
|
||||||
buildCanonicalIndex,
|
buildCanonicalIndex,
|
||||||
type CanonicalIndex,
|
type CanonicalIndex,
|
||||||
type CanonicalModeOverrides,
|
|
||||||
evaluateSubBlockCondition,
|
evaluateSubBlockCondition,
|
||||||
isCanonicalPair,
|
|
||||||
resolveCanonicalMode,
|
|
||||||
type SubBlockCondition,
|
type SubBlockCondition,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,17 +1,13 @@
|
|||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { extractInputFieldsFromBlocks } from '@/lib/workflows/input-format'
|
import { extractInputFieldsFromBlocks } from '@/lib/workflows/input-format'
|
||||||
import {
|
import {
|
||||||
buildCanonicalIndex,
|
|
||||||
type CanonicalModeOverrides,
|
|
||||||
evaluateSubBlockCondition,
|
evaluateSubBlockCondition,
|
||||||
isCanonicalPair,
|
|
||||||
resolveCanonicalMode,
|
|
||||||
type SubBlockCondition,
|
type SubBlockCondition,
|
||||||
} from '@/lib/workflows/subblocks/visibility'
|
} from '@/lib/workflows/subblocks/visibility'
|
||||||
import type { SubBlockConfig as BlockSubBlockConfig, GenerationType } from '@/blocks/types'
|
import type { SubBlockConfig as BlockSubBlockConfig } from '@/blocks/types'
|
||||||
import { safeAssign } from '@/tools/safe-assign'
|
import { safeAssign } from '@/tools/safe-assign'
|
||||||
import { isEmptyTagValue } from '@/tools/shared/tags'
|
import { isEmptyTagValue } from '@/tools/shared/tags'
|
||||||
import type { OAuthConfig, ParameterVisibility, ToolConfig } from '@/tools/types'
|
import type { ParameterVisibility, ToolConfig } from '@/tools/types'
|
||||||
import { getTool } from '@/tools/utils'
|
import { getTool } from '@/tools/utils'
|
||||||
|
|
||||||
const logger = createLogger('ToolsParams')
|
const logger = createLogger('ToolsParams')
|
||||||
@@ -68,14 +64,6 @@ export interface UIComponentConfig {
|
|||||||
mode?: 'basic' | 'advanced' | 'both' | 'trigger'
|
mode?: 'basic' | 'advanced' | 'both' | 'trigger'
|
||||||
/** The actual subblock ID this config was derived from */
|
/** The actual subblock ID this config was derived from */
|
||||||
actualSubBlockId?: string
|
actualSubBlockId?: string
|
||||||
/** Wand configuration for AI assistance */
|
|
||||||
wandConfig?: {
|
|
||||||
enabled: boolean
|
|
||||||
prompt: string
|
|
||||||
generationType?: GenerationType
|
|
||||||
placeholder?: string
|
|
||||||
maintainHistory?: boolean
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface SubBlockConfig {
|
export interface SubBlockConfig {
|
||||||
@@ -339,7 +327,6 @@ export function getToolParametersConfig(
|
|||||||
canonicalParamId: subBlock.canonicalParamId,
|
canonicalParamId: subBlock.canonicalParamId,
|
||||||
mode: subBlock.mode,
|
mode: subBlock.mode,
|
||||||
actualSubBlockId: subBlock.id,
|
actualSubBlockId: subBlock.id,
|
||||||
wandConfig: subBlock.wandConfig,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -825,200 +812,3 @@ export function formatParameterLabel(paramId: string): string {
|
|||||||
// Simple case - just capitalize first letter
|
// Simple case - just capitalize first letter
|
||||||
return paramId.charAt(0).toUpperCase() + paramId.slice(1)
|
return paramId.charAt(0).toUpperCase() + paramId.slice(1)
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* SubBlock IDs that are "structural" — they control tool routing or auth,
|
|
||||||
* not user-facing parameters. These are excluded from tool-input rendering
|
|
||||||
* unless they have an explicit paramVisibility set.
|
|
||||||
*/
|
|
||||||
const STRUCTURAL_SUBBLOCK_IDS = new Set(['operation', 'authMethod', 'destinationType'])
|
|
||||||
|
|
||||||
/**
|
|
||||||
* SubBlock types that represent auth/credential inputs handled separately
|
|
||||||
* by the tool-input OAuth credential selector.
|
|
||||||
*/
|
|
||||||
const AUTH_SUBBLOCK_TYPES = new Set(['oauth-input'])
|
|
||||||
|
|
||||||
/**
|
|
||||||
* SubBlock types that should never appear in tool-input context.
|
|
||||||
*/
|
|
||||||
const EXCLUDED_SUBBLOCK_TYPES = new Set([
|
|
||||||
'tool-input',
|
|
||||||
'skill-input',
|
|
||||||
'condition-input',
|
|
||||||
'eval-input',
|
|
||||||
'webhook-config',
|
|
||||||
'schedule-info',
|
|
||||||
'trigger-save',
|
|
||||||
'input-format',
|
|
||||||
'response-format',
|
|
||||||
'mcp-server-selector',
|
|
||||||
'mcp-tool-selector',
|
|
||||||
'mcp-dynamic-args',
|
|
||||||
'input-mapping',
|
|
||||||
'variables-input',
|
|
||||||
'messages-input',
|
|
||||||
'router-input',
|
|
||||||
'text',
|
|
||||||
])
|
|
||||||
|
|
||||||
export interface SubBlocksForToolInput {
|
|
||||||
toolConfig: ToolConfig
|
|
||||||
subBlocks: BlockSubBlockConfig[]
|
|
||||||
oauthConfig?: OAuthConfig
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Returns filtered SubBlockConfig[] for rendering in tool-input context.
|
|
||||||
* Uses subblock definitions as the primary source of UI metadata,
|
|
||||||
* getting all features (wandConfig, rich conditions, dependsOn, etc.) for free.
|
|
||||||
*
|
|
||||||
* For blocks without paramVisibility annotations, falls back to inferring
|
|
||||||
* visibility from the tool's param definitions.
|
|
||||||
*/
|
|
||||||
export function getSubBlocksForToolInput(
|
|
||||||
toolId: string,
|
|
||||||
blockType: string,
|
|
||||||
currentValues?: Record<string, unknown>,
|
|
||||||
canonicalModeOverrides?: CanonicalModeOverrides
|
|
||||||
): SubBlocksForToolInput | null {
|
|
||||||
try {
|
|
||||||
const toolConfig = getTool(toolId)
|
|
||||||
if (!toolConfig) {
|
|
||||||
logger.warn(`Tool not found: ${toolId}`)
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
|
|
||||||
const blockConfigs = getBlockConfigurations()
|
|
||||||
const blockConfig = blockConfigs[blockType]
|
|
||||||
if (!blockConfig?.subBlocks?.length) {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
|
|
||||||
const allSubBlocks = blockConfig.subBlocks as BlockSubBlockConfig[]
|
|
||||||
const canonicalIndex = buildCanonicalIndex(allSubBlocks)
|
|
||||||
|
|
||||||
// Build values for condition evaluation
|
|
||||||
const values = currentValues || {}
|
|
||||||
const valuesWithOperation = { ...values }
|
|
||||||
if (valuesWithOperation.operation === undefined) {
|
|
||||||
const parts = toolId.split('_')
|
|
||||||
valuesWithOperation.operation =
|
|
||||||
parts.length >= 3 ? parts.slice(2).join('_') : parts[parts.length - 1]
|
|
||||||
}
|
|
||||||
|
|
||||||
// Build a map of tool param IDs to their resolved visibility
|
|
||||||
const toolParamVisibility: Record<string, ParameterVisibility> = {}
|
|
||||||
for (const [paramId, param] of Object.entries(toolConfig.params || {})) {
|
|
||||||
toolParamVisibility[paramId] =
|
|
||||||
param.visibility ?? (param.required ? 'user-or-llm' : 'user-only')
|
|
||||||
}
|
|
||||||
|
|
||||||
// Track which canonical groups we've already included (to avoid duplicates)
|
|
||||||
const includedCanonicalIds = new Set<string>()
|
|
||||||
|
|
||||||
const filtered: BlockSubBlockConfig[] = []
|
|
||||||
|
|
||||||
for (const sb of allSubBlocks) {
|
|
||||||
// Skip excluded types
|
|
||||||
if (EXCLUDED_SUBBLOCK_TYPES.has(sb.type)) continue
|
|
||||||
|
|
||||||
// Skip trigger-mode-only subblocks
|
|
||||||
if (sb.mode === 'trigger') continue
|
|
||||||
|
|
||||||
// Determine the effective param ID (canonical or subblock id)
|
|
||||||
const effectiveParamId = sb.canonicalParamId || sb.id
|
|
||||||
|
|
||||||
// Resolve paramVisibility: explicit > inferred from tool params > skip
|
|
||||||
let visibility = sb.paramVisibility
|
|
||||||
if (!visibility) {
|
|
||||||
// Infer from structural checks
|
|
||||||
if (STRUCTURAL_SUBBLOCK_IDS.has(sb.id)) {
|
|
||||||
visibility = 'hidden'
|
|
||||||
} else if (AUTH_SUBBLOCK_TYPES.has(sb.type)) {
|
|
||||||
visibility = 'hidden'
|
|
||||||
} else if (
|
|
||||||
sb.password &&
|
|
||||||
(sb.id === 'botToken' || sb.id === 'accessToken' || sb.id === 'apiKey')
|
|
||||||
) {
|
|
||||||
// Auth tokens without explicit paramVisibility are hidden
|
|
||||||
// (they're handled by the OAuth credential selector or structurally)
|
|
||||||
// But only if they don't have a matching tool param
|
|
||||||
if (!(sb.id in toolParamVisibility)) {
|
|
||||||
visibility = 'hidden'
|
|
||||||
} else {
|
|
||||||
visibility = toolParamVisibility[sb.id] || 'user-or-llm'
|
|
||||||
}
|
|
||||||
} else if (effectiveParamId in toolParamVisibility) {
|
|
||||||
// Fallback: infer from tool param visibility
|
|
||||||
visibility = toolParamVisibility[effectiveParamId]
|
|
||||||
} else if (sb.id in toolParamVisibility) {
|
|
||||||
visibility = toolParamVisibility[sb.id]
|
|
||||||
} else if (sb.canonicalParamId) {
|
|
||||||
// SubBlock has a canonicalParamId that doesn't directly match a tool param.
|
|
||||||
// This means the block's params() function transforms it before sending to the tool
|
|
||||||
// (e.g. listFolderId → folderId). These are user-facing inputs, default to user-or-llm.
|
|
||||||
visibility = 'user-or-llm'
|
|
||||||
} else {
|
|
||||||
// SubBlock has no corresponding tool param — skip it
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Filter by visibility: exclude hidden and llm-only
|
|
||||||
if (visibility === 'hidden' || visibility === 'llm-only') continue
|
|
||||||
|
|
||||||
// Evaluate condition against current values
|
|
||||||
if (sb.condition) {
|
|
||||||
const conditionMet = evaluateSubBlockCondition(
|
|
||||||
sb.condition as SubBlockCondition,
|
|
||||||
valuesWithOperation
|
|
||||||
)
|
|
||||||
if (!conditionMet) continue
|
|
||||||
}
|
|
||||||
|
|
||||||
// Handle canonical pairs: only include the active mode variant
|
|
||||||
const canonicalId = canonicalIndex.canonicalIdBySubBlockId[sb.id]
|
|
||||||
if (canonicalId) {
|
|
||||||
const group = canonicalIndex.groupsById[canonicalId]
|
|
||||||
if (group && isCanonicalPair(group)) {
|
|
||||||
if (includedCanonicalIds.has(canonicalId)) continue
|
|
||||||
includedCanonicalIds.add(canonicalId)
|
|
||||||
|
|
||||||
// Determine active mode
|
|
||||||
const mode = resolveCanonicalMode(group, valuesWithOperation, canonicalModeOverrides)
|
|
||||||
if (mode === 'advanced') {
|
|
||||||
// Find the advanced variant
|
|
||||||
const advancedSb = allSubBlocks.find((s) => group.advancedIds.includes(s.id))
|
|
||||||
if (advancedSb) {
|
|
||||||
filtered.push({ ...advancedSb, paramVisibility: visibility })
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// Include basic variant (current sb if it's the basic one)
|
|
||||||
if (group.basicId === sb.id) {
|
|
||||||
filtered.push({ ...sb, paramVisibility: visibility })
|
|
||||||
} else {
|
|
||||||
const basicSb = allSubBlocks.find((s) => s.id === group.basicId)
|
|
||||||
if (basicSb) {
|
|
||||||
filtered.push({ ...basicSb, paramVisibility: visibility })
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Non-canonical, non-hidden, condition-passing subblock
|
|
||||||
filtered.push({ ...sb, paramVisibility: visibility })
|
|
||||||
}
|
|
||||||
|
|
||||||
return {
|
|
||||||
toolConfig,
|
|
||||||
subBlocks: filtered,
|
|
||||||
oauthConfig: toolConfig.oauth,
|
|
||||||
}
|
|
||||||
} catch (error) {
|
|
||||||
logger.error('Error getting subblocks for tool input:', error)
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -23,12 +23,7 @@ export const SYSTEM_SUBBLOCK_IDS: string[] = [
|
|||||||
* with default values from the trigger definition on load, which aren't present in
|
* with default values from the trigger definition on load, which aren't present in
|
||||||
* the deployed state, causing false positive change detection.
|
* the deployed state, causing false positive change detection.
|
||||||
*/
|
*/
|
||||||
export const TRIGGER_RUNTIME_SUBBLOCK_IDS: string[] = [
|
export const TRIGGER_RUNTIME_SUBBLOCK_IDS: string[] = ['webhookId', 'triggerPath', 'triggerConfig']
|
||||||
'webhookId',
|
|
||||||
'triggerPath',
|
|
||||||
'triggerConfig',
|
|
||||||
'triggerId',
|
|
||||||
]
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Maximum number of consecutive failures before a trigger (schedule/webhook) is auto-disabled.
|
* Maximum number of consecutive failures before a trigger (schedule/webhook) is auto-disabled.
|
||||||
|
|||||||
Reference in New Issue
Block a user