mirror of
https://github.com/simstudioai/sim.git
synced 2026-02-12 23:45:07 -05:00
Compare commits
55 Commits
feat/sim-p
...
v0.5.88
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
07d50f8fe1 | ||
|
|
27973953f6 | ||
|
|
50585273ce | ||
|
|
654cb2b407 | ||
|
|
6c66521d64 | ||
|
|
479cd347ad | ||
|
|
a3a99eda19 | ||
|
|
1a66d48add | ||
|
|
46822e91f3 | ||
|
|
2bb68335ee | ||
|
|
8528fbe2d2 | ||
|
|
31fdd2be13 | ||
|
|
028bc652c2 | ||
|
|
c6bf5cd58c | ||
|
|
11dc18a80d | ||
|
|
ab4e9dc72f | ||
|
|
1c58c35bd8 | ||
|
|
d63a5cb504 | ||
|
|
8bd5d41723 | ||
|
|
c12931bc50 | ||
|
|
e9c4251c1c | ||
|
|
cc2be33d6b | ||
|
|
45371e521e | ||
|
|
0ce0f98aa5 | ||
|
|
dff1c9d083 | ||
|
|
b09f683072 | ||
|
|
a8bb0db660 | ||
|
|
af82820a28 | ||
|
|
4372841797 | ||
|
|
5e8c843241 | ||
|
|
7bf3d73ee6 | ||
|
|
7ffc11a738 | ||
|
|
be578e2ed7 | ||
|
|
f415e5edc4 | ||
|
|
13a6e6c3fa | ||
|
|
f5ab7f21ae | ||
|
|
bfb6fffe38 | ||
|
|
4fbec0a43f | ||
|
|
585f5e365b | ||
|
|
3792bdd252 | ||
|
|
eb5d1f3e5b | ||
|
|
54ab82c8dd | ||
|
|
f895bf469b | ||
|
|
dd3209af06 | ||
|
|
b6ba3b50a7 | ||
|
|
b304233062 | ||
|
|
57e4b49bd6 | ||
|
|
e12dd204ed | ||
|
|
3d9d9cbc54 | ||
|
|
0f4ec962ad | ||
|
|
4827866f9a | ||
|
|
3e697d9ed9 | ||
|
|
4431a1a484 | ||
|
|
4d1a9a3f22 | ||
|
|
eb07a080fb |
@@ -4,10 +4,20 @@
|
||||
* @vitest-environment node
|
||||
*/
|
||||
|
||||
import { databaseMock, loggerMock } from '@sim/testing'
|
||||
import { loggerMock } from '@sim/testing'
|
||||
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
|
||||
vi.mock('@sim/db', () => databaseMock)
|
||||
vi.mock('@sim/db', () => ({
|
||||
db: {
|
||||
select: vi.fn().mockReturnThis(),
|
||||
from: vi.fn().mockReturnThis(),
|
||||
where: vi.fn().mockReturnThis(),
|
||||
limit: vi.fn().mockReturnValue([]),
|
||||
update: vi.fn().mockReturnThis(),
|
||||
set: vi.fn().mockReturnThis(),
|
||||
orderBy: vi.fn().mockReturnThis(),
|
||||
},
|
||||
}))
|
||||
|
||||
vi.mock('@/lib/oauth/oauth', () => ({
|
||||
refreshOAuthToken: vi.fn(),
|
||||
@@ -24,36 +34,13 @@ import {
|
||||
refreshTokenIfNeeded,
|
||||
} from '@/app/api/auth/oauth/utils'
|
||||
|
||||
const mockDb = db as any
|
||||
const mockDbTyped = db as any
|
||||
const mockRefreshOAuthToken = refreshOAuthToken as any
|
||||
|
||||
/**
|
||||
* Creates a chainable mock for db.select() calls.
|
||||
* Returns a nested chain: select() -> from() -> where() -> limit() / orderBy()
|
||||
*/
|
||||
function mockSelectChain(limitResult: unknown[]) {
|
||||
const mockLimit = vi.fn().mockReturnValue(limitResult)
|
||||
const mockOrderBy = vi.fn().mockReturnValue(limitResult)
|
||||
const mockWhere = vi.fn().mockReturnValue({ limit: mockLimit, orderBy: mockOrderBy })
|
||||
const mockFrom = vi.fn().mockReturnValue({ where: mockWhere })
|
||||
mockDb.select.mockReturnValueOnce({ from: mockFrom })
|
||||
return { mockFrom, mockWhere, mockLimit }
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a chainable mock for db.update() calls.
|
||||
* Returns a nested chain: update() -> set() -> where()
|
||||
*/
|
||||
function mockUpdateChain() {
|
||||
const mockWhere = vi.fn().mockResolvedValue({})
|
||||
const mockSet = vi.fn().mockReturnValue({ where: mockWhere })
|
||||
mockDb.update.mockReturnValueOnce({ set: mockSet })
|
||||
return { mockSet, mockWhere }
|
||||
}
|
||||
|
||||
describe('OAuth Utils', () => {
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks()
|
||||
mockDbTyped.limit.mockReturnValue([])
|
||||
})
|
||||
|
||||
afterEach(() => {
|
||||
@@ -63,20 +50,20 @@ describe('OAuth Utils', () => {
|
||||
describe('getCredential', () => {
|
||||
it('should return credential when found', async () => {
|
||||
const mockCredential = { id: 'credential-id', userId: 'test-user-id' }
|
||||
const { mockFrom, mockWhere, mockLimit } = mockSelectChain([mockCredential])
|
||||
mockDbTyped.limit.mockReturnValueOnce([mockCredential])
|
||||
|
||||
const credential = await getCredential('request-id', 'credential-id', 'test-user-id')
|
||||
|
||||
expect(mockDb.select).toHaveBeenCalled()
|
||||
expect(mockFrom).toHaveBeenCalled()
|
||||
expect(mockWhere).toHaveBeenCalled()
|
||||
expect(mockLimit).toHaveBeenCalledWith(1)
|
||||
expect(mockDbTyped.select).toHaveBeenCalled()
|
||||
expect(mockDbTyped.from).toHaveBeenCalled()
|
||||
expect(mockDbTyped.where).toHaveBeenCalled()
|
||||
expect(mockDbTyped.limit).toHaveBeenCalledWith(1)
|
||||
|
||||
expect(credential).toEqual(mockCredential)
|
||||
})
|
||||
|
||||
it('should return undefined when credential is not found', async () => {
|
||||
mockSelectChain([])
|
||||
mockDbTyped.limit.mockReturnValueOnce([])
|
||||
|
||||
const credential = await getCredential('request-id', 'nonexistent-id', 'test-user-id')
|
||||
|
||||
@@ -115,12 +102,11 @@ describe('OAuth Utils', () => {
|
||||
refreshToken: 'new-refresh-token',
|
||||
})
|
||||
|
||||
mockUpdateChain()
|
||||
|
||||
const result = await refreshTokenIfNeeded('request-id', mockCredential, 'credential-id')
|
||||
|
||||
expect(mockRefreshOAuthToken).toHaveBeenCalledWith('google', 'refresh-token')
|
||||
expect(mockDb.update).toHaveBeenCalled()
|
||||
expect(mockDbTyped.update).toHaveBeenCalled()
|
||||
expect(mockDbTyped.set).toHaveBeenCalled()
|
||||
expect(result).toEqual({ accessToken: 'new-token', refreshed: true })
|
||||
})
|
||||
|
||||
@@ -166,7 +152,7 @@ describe('OAuth Utils', () => {
|
||||
providerId: 'google',
|
||||
userId: 'test-user-id',
|
||||
}
|
||||
mockSelectChain([mockCredential])
|
||||
mockDbTyped.limit.mockReturnValueOnce([mockCredential])
|
||||
|
||||
const token = await refreshAccessTokenIfNeeded('credential-id', 'test-user-id', 'request-id')
|
||||
|
||||
@@ -183,8 +169,7 @@ describe('OAuth Utils', () => {
|
||||
providerId: 'google',
|
||||
userId: 'test-user-id',
|
||||
}
|
||||
mockSelectChain([mockCredential])
|
||||
mockUpdateChain()
|
||||
mockDbTyped.limit.mockReturnValueOnce([mockCredential])
|
||||
|
||||
mockRefreshOAuthToken.mockResolvedValueOnce({
|
||||
accessToken: 'new-token',
|
||||
@@ -195,12 +180,13 @@ describe('OAuth Utils', () => {
|
||||
const token = await refreshAccessTokenIfNeeded('credential-id', 'test-user-id', 'request-id')
|
||||
|
||||
expect(mockRefreshOAuthToken).toHaveBeenCalledWith('google', 'refresh-token')
|
||||
expect(mockDb.update).toHaveBeenCalled()
|
||||
expect(mockDbTyped.update).toHaveBeenCalled()
|
||||
expect(mockDbTyped.set).toHaveBeenCalled()
|
||||
expect(token).toBe('new-token')
|
||||
})
|
||||
|
||||
it('should return null if credential not found', async () => {
|
||||
mockSelectChain([])
|
||||
mockDbTyped.limit.mockReturnValueOnce([])
|
||||
|
||||
const token = await refreshAccessTokenIfNeeded('nonexistent-id', 'test-user-id', 'request-id')
|
||||
|
||||
@@ -216,7 +202,7 @@ describe('OAuth Utils', () => {
|
||||
providerId: 'google',
|
||||
userId: 'test-user-id',
|
||||
}
|
||||
mockSelectChain([mockCredential])
|
||||
mockDbTyped.limit.mockReturnValueOnce([mockCredential])
|
||||
|
||||
mockRefreshOAuthToken.mockResolvedValueOnce(null)
|
||||
|
||||
|
||||
@@ -85,7 +85,7 @@ const ChatMessageSchema = z.object({
|
||||
chatId: z.string().optional(),
|
||||
workflowId: z.string().optional(),
|
||||
workflowName: z.string().optional(),
|
||||
model: z.string().optional().default('claude-opus-4-5'),
|
||||
model: z.string().optional().default('claude-opus-4-6'),
|
||||
mode: z.enum(COPILOT_REQUEST_MODES).optional().default('agent'),
|
||||
prefetch: z.boolean().optional(),
|
||||
createNewChat: z.boolean().optional().default(false),
|
||||
@@ -238,7 +238,7 @@ export async function POST(req: NextRequest) {
|
||||
let currentChat: any = null
|
||||
let conversationHistory: any[] = []
|
||||
let actualChatId = chatId
|
||||
const selectedModel = model || 'claude-opus-4-5'
|
||||
const selectedModel = model || 'claude-opus-4-6'
|
||||
|
||||
if (chatId || createNewChat) {
|
||||
const chatResult = await resolveOrCreateChat({
|
||||
|
||||
@@ -4,12 +4,16 @@
|
||||
*
|
||||
* @vitest-environment node
|
||||
*/
|
||||
import { createEnvMock, databaseMock, loggerMock } from '@sim/testing'
|
||||
import { createEnvMock, createMockLogger } from '@sim/testing'
|
||||
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
|
||||
const loggerMock = vi.hoisted(() => ({
|
||||
createLogger: () => createMockLogger(),
|
||||
}))
|
||||
|
||||
vi.mock('drizzle-orm')
|
||||
vi.mock('@sim/logger', () => loggerMock)
|
||||
vi.mock('@sim/db', () => databaseMock)
|
||||
vi.mock('@sim/db')
|
||||
vi.mock('@/lib/knowledge/documents/utils', () => ({
|
||||
retryWithExponentialBackoff: (fn: any) => fn(),
|
||||
}))
|
||||
|
||||
@@ -38,7 +38,7 @@ import {
|
||||
|
||||
const logger = createLogger('CopilotMcpAPI')
|
||||
const mcpRateLimiter = new RateLimiter()
|
||||
const DEFAULT_COPILOT_MODEL = 'claude-opus-4-5'
|
||||
const DEFAULT_COPILOT_MODEL = 'claude-opus-4-6'
|
||||
|
||||
export const dynamic = 'force-dynamic'
|
||||
export const runtime = 'nodejs'
|
||||
|
||||
@@ -3,14 +3,17 @@
|
||||
*
|
||||
* @vitest-environment node
|
||||
*/
|
||||
import { databaseMock, loggerMock } from '@sim/testing'
|
||||
import { loggerMock } from '@sim/testing'
|
||||
import { NextRequest } from 'next/server'
|
||||
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
|
||||
const { mockGetSession, mockAuthorizeWorkflowByWorkspacePermission } = vi.hoisted(() => ({
|
||||
mockGetSession: vi.fn(),
|
||||
mockAuthorizeWorkflowByWorkspacePermission: vi.fn(),
|
||||
}))
|
||||
const { mockGetSession, mockAuthorizeWorkflowByWorkspacePermission, mockDbSelect, mockDbUpdate } =
|
||||
vi.hoisted(() => ({
|
||||
mockGetSession: vi.fn(),
|
||||
mockAuthorizeWorkflowByWorkspacePermission: vi.fn(),
|
||||
mockDbSelect: vi.fn(),
|
||||
mockDbUpdate: vi.fn(),
|
||||
}))
|
||||
|
||||
vi.mock('@/lib/auth', () => ({
|
||||
getSession: mockGetSession,
|
||||
@@ -20,7 +23,12 @@ vi.mock('@/lib/workflows/utils', () => ({
|
||||
authorizeWorkflowByWorkspacePermission: mockAuthorizeWorkflowByWorkspacePermission,
|
||||
}))
|
||||
|
||||
vi.mock('@sim/db', () => databaseMock)
|
||||
vi.mock('@sim/db', () => ({
|
||||
db: {
|
||||
select: mockDbSelect,
|
||||
update: mockDbUpdate,
|
||||
},
|
||||
}))
|
||||
|
||||
vi.mock('@sim/db/schema', () => ({
|
||||
workflow: { id: 'id', userId: 'userId', workspaceId: 'workspaceId' },
|
||||
@@ -51,9 +59,6 @@ function createParams(id: string): { params: Promise<{ id: string }> } {
|
||||
return { params: Promise.resolve({ id }) }
|
||||
}
|
||||
|
||||
const mockDbSelect = databaseMock.db.select as ReturnType<typeof vi.fn>
|
||||
const mockDbUpdate = databaseMock.db.update as ReturnType<typeof vi.fn>
|
||||
|
||||
function mockDbChain(selectResults: unknown[][]) {
|
||||
let selectCallIndex = 0
|
||||
mockDbSelect.mockImplementation(() => ({
|
||||
|
||||
@@ -3,14 +3,17 @@
|
||||
*
|
||||
* @vitest-environment node
|
||||
*/
|
||||
import { databaseMock, loggerMock } from '@sim/testing'
|
||||
import { loggerMock } from '@sim/testing'
|
||||
import { NextRequest } from 'next/server'
|
||||
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
|
||||
const { mockGetSession, mockAuthorizeWorkflowByWorkspacePermission } = vi.hoisted(() => ({
|
||||
mockGetSession: vi.fn(),
|
||||
mockAuthorizeWorkflowByWorkspacePermission: vi.fn(),
|
||||
}))
|
||||
const { mockGetSession, mockAuthorizeWorkflowByWorkspacePermission, mockDbSelect } = vi.hoisted(
|
||||
() => ({
|
||||
mockGetSession: vi.fn(),
|
||||
mockAuthorizeWorkflowByWorkspacePermission: vi.fn(),
|
||||
mockDbSelect: vi.fn(),
|
||||
})
|
||||
)
|
||||
|
||||
vi.mock('@/lib/auth', () => ({
|
||||
getSession: mockGetSession,
|
||||
@@ -20,7 +23,11 @@ vi.mock('@/lib/workflows/utils', () => ({
|
||||
authorizeWorkflowByWorkspacePermission: mockAuthorizeWorkflowByWorkspacePermission,
|
||||
}))
|
||||
|
||||
vi.mock('@sim/db', () => databaseMock)
|
||||
vi.mock('@sim/db', () => ({
|
||||
db: {
|
||||
select: mockDbSelect,
|
||||
},
|
||||
}))
|
||||
|
||||
vi.mock('@sim/db/schema', () => ({
|
||||
workflow: { id: 'id', userId: 'userId', workspaceId: 'workspaceId' },
|
||||
@@ -55,8 +62,6 @@ function createRequest(url: string): NextRequest {
|
||||
return new NextRequest(new URL(url), { method: 'GET' })
|
||||
}
|
||||
|
||||
const mockDbSelect = databaseMock.db.select as ReturnType<typeof vi.fn>
|
||||
|
||||
function mockDbChain(results: any[]) {
|
||||
let callIndex = 0
|
||||
mockDbSelect.mockImplementation(() => ({
|
||||
|
||||
@@ -8,7 +8,7 @@ import { resolveWorkflowIdForUser } from '@/lib/workflows/utils'
|
||||
import { authenticateV1Request } from '@/app/api/v1/auth'
|
||||
|
||||
const logger = createLogger('CopilotHeadlessAPI')
|
||||
const DEFAULT_COPILOT_MODEL = 'claude-opus-4-5'
|
||||
const DEFAULT_COPILOT_MODEL = 'claude-opus-4-6'
|
||||
|
||||
const RequestSchema = z.object({
|
||||
message: z.string().min(1, 'message is required'),
|
||||
|
||||
@@ -29,7 +29,7 @@ const patchBodySchema = z
|
||||
description: z
|
||||
.string()
|
||||
.trim()
|
||||
.max(2000, 'Description must be 2000 characters or less')
|
||||
.max(500, 'Description must be 500 characters or less')
|
||||
.nullable()
|
||||
.optional(),
|
||||
isActive: z.literal(true).optional(), // Set to true to activate this version
|
||||
|
||||
@@ -12,7 +12,7 @@ import {
|
||||
import { generateRequestId } from '@/lib/core/utils/request'
|
||||
import { SSE_HEADERS } from '@/lib/core/utils/sse'
|
||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
||||
import { createExecutionEventWriter, setExecutionMeta } from '@/lib/execution/event-buffer'
|
||||
import { markExecutionCancelled } from '@/lib/execution/cancellation'
|
||||
import { processInputFileFields } from '@/lib/execution/files'
|
||||
import { preprocessExecution } from '@/lib/execution/preprocessing'
|
||||
import { LoggingSession } from '@/lib/logs/execution/logging-session'
|
||||
@@ -700,27 +700,15 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
||||
const timeoutController = createTimeoutAbortController(preprocessResult.executionTimeout?.sync)
|
||||
let isStreamClosed = false
|
||||
|
||||
const eventWriter = createExecutionEventWriter(executionId)
|
||||
setExecutionMeta(executionId, {
|
||||
status: 'active',
|
||||
userId: actorUserId,
|
||||
workflowId,
|
||||
}).catch(() => {})
|
||||
|
||||
const stream = new ReadableStream<Uint8Array>({
|
||||
async start(controller) {
|
||||
let finalMetaStatus: 'complete' | 'error' | 'cancelled' | null = null
|
||||
|
||||
const sendEvent = (event: ExecutionEvent) => {
|
||||
if (!isStreamClosed) {
|
||||
try {
|
||||
controller.enqueue(encodeSSEEvent(event))
|
||||
} catch {
|
||||
isStreamClosed = true
|
||||
}
|
||||
}
|
||||
if (event.type !== 'stream:chunk' && event.type !== 'stream:done') {
|
||||
eventWriter.write(event).catch(() => {})
|
||||
if (isStreamClosed) return
|
||||
|
||||
try {
|
||||
controller.enqueue(encodeSSEEvent(event))
|
||||
} catch {
|
||||
isStreamClosed = true
|
||||
}
|
||||
}
|
||||
|
||||
@@ -841,12 +829,14 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
||||
|
||||
const reader = streamingExec.stream.getReader()
|
||||
const decoder = new TextDecoder()
|
||||
let chunkCount = 0
|
||||
|
||||
try {
|
||||
while (true) {
|
||||
const { done, value } = await reader.read()
|
||||
if (done) break
|
||||
|
||||
chunkCount++
|
||||
const chunk = decoder.decode(value, { stream: true })
|
||||
sendEvent({
|
||||
type: 'stream:chunk',
|
||||
@@ -961,7 +951,6 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
||||
duration: result.metadata?.duration || 0,
|
||||
},
|
||||
})
|
||||
finalMetaStatus = 'error'
|
||||
} else {
|
||||
logger.info(`[${requestId}] Workflow execution was cancelled`)
|
||||
|
||||
@@ -974,7 +963,6 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
||||
duration: result.metadata?.duration || 0,
|
||||
},
|
||||
})
|
||||
finalMetaStatus = 'cancelled'
|
||||
}
|
||||
return
|
||||
}
|
||||
@@ -998,7 +986,6 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
||||
endTime: result.metadata?.endTime || new Date().toISOString(),
|
||||
},
|
||||
})
|
||||
finalMetaStatus = 'complete'
|
||||
} catch (error: unknown) {
|
||||
const isTimeout = isTimeoutError(error) || timeoutController.isTimedOut()
|
||||
const errorMessage = isTimeout
|
||||
@@ -1030,18 +1017,7 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
||||
duration: executionResult?.metadata?.duration || 0,
|
||||
},
|
||||
})
|
||||
finalMetaStatus = 'error'
|
||||
} finally {
|
||||
try {
|
||||
await eventWriter.close()
|
||||
} catch (closeError) {
|
||||
logger.warn(`[${requestId}] Failed to close event writer`, {
|
||||
error: closeError instanceof Error ? closeError.message : String(closeError),
|
||||
})
|
||||
}
|
||||
if (finalMetaStatus) {
|
||||
setExecutionMeta(executionId, { status: finalMetaStatus }).catch(() => {})
|
||||
}
|
||||
timeoutController.cleanup()
|
||||
if (executionId) {
|
||||
await cleanupExecutionBase64Cache(executionId)
|
||||
@@ -1056,7 +1032,10 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
||||
},
|
||||
cancel() {
|
||||
isStreamClosed = true
|
||||
logger.info(`[${requestId}] Client disconnected from SSE stream`)
|
||||
timeoutController.cleanup()
|
||||
logger.info(`[${requestId}] Client aborted SSE stream, signalling cancellation`)
|
||||
timeoutController.abort()
|
||||
markExecutionCancelled(executionId).catch(() => {})
|
||||
},
|
||||
})
|
||||
|
||||
|
||||
@@ -1,170 +0,0 @@
|
||||
import { createLogger } from '@sim/logger'
|
||||
import { type NextRequest, NextResponse } from 'next/server'
|
||||
import { checkHybridAuth } from '@/lib/auth/hybrid'
|
||||
import { SSE_HEADERS } from '@/lib/core/utils/sse'
|
||||
import {
|
||||
type ExecutionStreamStatus,
|
||||
getExecutionMeta,
|
||||
readExecutionEvents,
|
||||
} from '@/lib/execution/event-buffer'
|
||||
import { formatSSEEvent } from '@/lib/workflows/executor/execution-events'
|
||||
import { authorizeWorkflowByWorkspacePermission } from '@/lib/workflows/utils'
|
||||
|
||||
const logger = createLogger('ExecutionStreamReconnectAPI')
|
||||
|
||||
const POLL_INTERVAL_MS = 500
|
||||
const MAX_POLL_DURATION_MS = 10 * 60 * 1000 // 10 minutes
|
||||
|
||||
function isTerminalStatus(status: ExecutionStreamStatus): boolean {
|
||||
return status === 'complete' || status === 'error' || status === 'cancelled'
|
||||
}
|
||||
|
||||
export const runtime = 'nodejs'
|
||||
export const dynamic = 'force-dynamic'
|
||||
|
||||
export async function GET(
|
||||
req: NextRequest,
|
||||
{ params }: { params: Promise<{ id: string; executionId: string }> }
|
||||
) {
|
||||
const { id: workflowId, executionId } = await params
|
||||
|
||||
try {
|
||||
const auth = await checkHybridAuth(req, { requireWorkflowId: false })
|
||||
if (!auth.success || !auth.userId) {
|
||||
return NextResponse.json({ error: auth.error || 'Unauthorized' }, { status: 401 })
|
||||
}
|
||||
|
||||
const workflowAuthorization = await authorizeWorkflowByWorkspacePermission({
|
||||
workflowId,
|
||||
userId: auth.userId,
|
||||
action: 'read',
|
||||
})
|
||||
if (!workflowAuthorization.allowed) {
|
||||
return NextResponse.json(
|
||||
{ error: workflowAuthorization.message || 'Access denied' },
|
||||
{ status: workflowAuthorization.status }
|
||||
)
|
||||
}
|
||||
|
||||
const meta = await getExecutionMeta(executionId)
|
||||
if (!meta) {
|
||||
return NextResponse.json({ error: 'Execution buffer not found or expired' }, { status: 404 })
|
||||
}
|
||||
|
||||
if (meta.workflowId && meta.workflowId !== workflowId) {
|
||||
return NextResponse.json(
|
||||
{ error: 'Execution does not belong to this workflow' },
|
||||
{ status: 403 }
|
||||
)
|
||||
}
|
||||
|
||||
const fromParam = req.nextUrl.searchParams.get('from')
|
||||
const parsed = fromParam ? Number.parseInt(fromParam, 10) : 0
|
||||
const fromEventId = Number.isFinite(parsed) && parsed >= 0 ? parsed : 0
|
||||
|
||||
logger.info('Reconnection stream requested', {
|
||||
workflowId,
|
||||
executionId,
|
||||
fromEventId,
|
||||
metaStatus: meta.status,
|
||||
})
|
||||
|
||||
const encoder = new TextEncoder()
|
||||
|
||||
let closed = false
|
||||
|
||||
const stream = new ReadableStream<Uint8Array>({
|
||||
async start(controller) {
|
||||
let lastEventId = fromEventId
|
||||
const pollDeadline = Date.now() + MAX_POLL_DURATION_MS
|
||||
|
||||
const enqueue = (text: string) => {
|
||||
if (closed) return
|
||||
try {
|
||||
controller.enqueue(encoder.encode(text))
|
||||
} catch {
|
||||
closed = true
|
||||
}
|
||||
}
|
||||
|
||||
try {
|
||||
const events = await readExecutionEvents(executionId, lastEventId)
|
||||
for (const entry of events) {
|
||||
if (closed) return
|
||||
enqueue(formatSSEEvent(entry.event))
|
||||
lastEventId = entry.eventId
|
||||
}
|
||||
|
||||
const currentMeta = await getExecutionMeta(executionId)
|
||||
if (!currentMeta || isTerminalStatus(currentMeta.status)) {
|
||||
enqueue('data: [DONE]\n\n')
|
||||
if (!closed) controller.close()
|
||||
return
|
||||
}
|
||||
|
||||
while (!closed && Date.now() < pollDeadline) {
|
||||
await new Promise((resolve) => setTimeout(resolve, POLL_INTERVAL_MS))
|
||||
if (closed) return
|
||||
|
||||
const newEvents = await readExecutionEvents(executionId, lastEventId)
|
||||
for (const entry of newEvents) {
|
||||
if (closed) return
|
||||
enqueue(formatSSEEvent(entry.event))
|
||||
lastEventId = entry.eventId
|
||||
}
|
||||
|
||||
const polledMeta = await getExecutionMeta(executionId)
|
||||
if (!polledMeta || isTerminalStatus(polledMeta.status)) {
|
||||
const finalEvents = await readExecutionEvents(executionId, lastEventId)
|
||||
for (const entry of finalEvents) {
|
||||
if (closed) return
|
||||
enqueue(formatSSEEvent(entry.event))
|
||||
lastEventId = entry.eventId
|
||||
}
|
||||
enqueue('data: [DONE]\n\n')
|
||||
if (!closed) controller.close()
|
||||
return
|
||||
}
|
||||
}
|
||||
|
||||
if (!closed) {
|
||||
logger.warn('Reconnection stream poll deadline reached', { executionId })
|
||||
enqueue('data: [DONE]\n\n')
|
||||
controller.close()
|
||||
}
|
||||
} catch (error) {
|
||||
logger.error('Error in reconnection stream', {
|
||||
executionId,
|
||||
error: error instanceof Error ? error.message : String(error),
|
||||
})
|
||||
if (!closed) {
|
||||
try {
|
||||
controller.close()
|
||||
} catch {}
|
||||
}
|
||||
}
|
||||
},
|
||||
cancel() {
|
||||
closed = true
|
||||
logger.info('Client disconnected from reconnection stream', { executionId })
|
||||
},
|
||||
})
|
||||
|
||||
return new NextResponse(stream, {
|
||||
headers: {
|
||||
...SSE_HEADERS,
|
||||
'X-Execution-Id': executionId,
|
||||
},
|
||||
})
|
||||
} catch (error: any) {
|
||||
logger.error('Failed to start reconnection stream', {
|
||||
workflowId,
|
||||
executionId,
|
||||
error: error.message,
|
||||
})
|
||||
return NextResponse.json(
|
||||
{ error: error.message || 'Failed to start reconnection stream' },
|
||||
{ status: 500 }
|
||||
)
|
||||
}
|
||||
}
|
||||
@@ -5,7 +5,7 @@
|
||||
* @vitest-environment node
|
||||
*/
|
||||
|
||||
import { loggerMock, setupGlobalFetchMock } from '@sim/testing'
|
||||
import { loggerMock } from '@sim/testing'
|
||||
import { NextRequest } from 'next/server'
|
||||
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
|
||||
@@ -284,7 +284,9 @@ describe('Workflow By ID API Route', () => {
|
||||
where: vi.fn().mockResolvedValue([{ id: 'workflow-123' }]),
|
||||
})
|
||||
|
||||
setupGlobalFetchMock({ ok: true })
|
||||
global.fetch = vi.fn().mockResolvedValue({
|
||||
ok: true,
|
||||
})
|
||||
|
||||
const req = new NextRequest('http://localhost:3000/api/workflows/workflow-123', {
|
||||
method: 'DELETE',
|
||||
@@ -329,7 +331,9 @@ describe('Workflow By ID API Route', () => {
|
||||
where: vi.fn().mockResolvedValue([{ id: 'workflow-123' }]),
|
||||
})
|
||||
|
||||
setupGlobalFetchMock({ ok: true })
|
||||
global.fetch = vi.fn().mockResolvedValue({
|
||||
ok: true,
|
||||
})
|
||||
|
||||
const req = new NextRequest('http://localhost:3000/api/workflows/workflow-123', {
|
||||
method: 'DELETE',
|
||||
|
||||
@@ -12,7 +12,7 @@ import { getUserEntityPermissions, getWorkspaceById } from '@/lib/workspaces/per
|
||||
|
||||
const logger = createLogger('WorkspaceBYOKKeysAPI')
|
||||
|
||||
const VALID_PROVIDERS = ['openai', 'anthropic', 'google', 'mistral', 'serper'] as const
|
||||
const VALID_PROVIDERS = ['openai', 'anthropic', 'google', 'mistral'] as const
|
||||
|
||||
const UpsertKeySchema = z.object({
|
||||
providerId: z.enum(VALID_PROVIDERS),
|
||||
|
||||
@@ -113,7 +113,7 @@ export function VersionDescriptionModal({
|
||||
className='min-h-[120px] resize-none'
|
||||
value={description}
|
||||
onChange={(e) => setDescription(e.target.value)}
|
||||
maxLength={2000}
|
||||
maxLength={500}
|
||||
disabled={isGenerating}
|
||||
/>
|
||||
<div className='flex items-center justify-between'>
|
||||
@@ -123,7 +123,7 @@ export function VersionDescriptionModal({
|
||||
</p>
|
||||
)}
|
||||
{!updateMutation.error && !generateMutation.error && <div />}
|
||||
<p className='text-[11px] text-[var(--text-tertiary)]'>{description.length}/2000</p>
|
||||
<p className='text-[11px] text-[var(--text-tertiary)]'>{description.length}/500</p>
|
||||
</div>
|
||||
</ModalBody>
|
||||
<ModalFooter>
|
||||
|
||||
@@ -57,21 +57,6 @@ export function useChangeDetection({
|
||||
}
|
||||
}
|
||||
|
||||
if (block.triggerMode) {
|
||||
const triggerConfigValue = blockSubValues?.triggerConfig
|
||||
if (
|
||||
triggerConfigValue &&
|
||||
typeof triggerConfigValue === 'object' &&
|
||||
!subBlocks.triggerConfig
|
||||
) {
|
||||
subBlocks.triggerConfig = {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: triggerConfigValue,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
blocksWithSubBlocks[blockId] = {
|
||||
...block,
|
||||
subBlocks,
|
||||
|
||||
@@ -3,7 +3,6 @@ import {
|
||||
buildCanonicalIndex,
|
||||
evaluateSubBlockCondition,
|
||||
isSubBlockFeatureEnabled,
|
||||
isSubBlockHiddenByHostedKey,
|
||||
isSubBlockVisibleForMode,
|
||||
} from '@/lib/workflows/subblocks/visibility'
|
||||
import type { BlockConfig, SubBlockConfig, SubBlockType } from '@/blocks/types'
|
||||
@@ -109,9 +108,6 @@ export function useEditorSubblockLayout(
|
||||
// Check required feature if specified - declarative feature gating
|
||||
if (!isSubBlockFeatureEnabled(block)) return false
|
||||
|
||||
// Hide tool API key fields when hosted key is available
|
||||
if (isSubBlockHiddenByHostedKey(block)) return false
|
||||
|
||||
// Special handling for trigger-config type (legacy trigger configuration UI)
|
||||
if (block.type === ('trigger-config' as SubBlockType)) {
|
||||
const isPureTriggerBlock = config?.triggers?.enabled && config.category === 'triggers'
|
||||
|
||||
@@ -15,7 +15,6 @@ import {
|
||||
evaluateSubBlockCondition,
|
||||
hasAdvancedValues,
|
||||
isSubBlockFeatureEnabled,
|
||||
isSubBlockHiddenByHostedKey,
|
||||
isSubBlockVisibleForMode,
|
||||
resolveDependencyValue,
|
||||
} from '@/lib/workflows/subblocks/visibility'
|
||||
@@ -829,7 +828,6 @@ export const WorkflowBlock = memo(function WorkflowBlock({
|
||||
if (block.hidden) return false
|
||||
if (block.hideFromPreview) return false
|
||||
if (!isSubBlockFeatureEnabled(block)) return false
|
||||
if (isSubBlockHiddenByHostedKey(block)) return false
|
||||
|
||||
const isPureTriggerBlock = config?.triggers?.enabled && config.category === 'triggers'
|
||||
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
import { useCallback, useEffect, useRef, useState } from 'react'
|
||||
import { useCallback, useRef, useState } from 'react'
|
||||
import { createLogger } from '@sim/logger'
|
||||
import { useQueryClient } from '@tanstack/react-query'
|
||||
import { v4 as uuidv4 } from 'uuid'
|
||||
@@ -46,13 +46,7 @@ import { useWorkflowStore } from '@/stores/workflows/workflow/store'
|
||||
|
||||
const logger = createLogger('useWorkflowExecution')
|
||||
|
||||
/**
|
||||
* Module-level Set tracking which workflows have an active reconnection effect.
|
||||
* Prevents multiple hook instances (from different components) from starting
|
||||
* concurrent reconnection streams for the same workflow during the same mount cycle.
|
||||
*/
|
||||
const activeReconnections = new Set<string>()
|
||||
|
||||
// Debug state validation result
|
||||
interface DebugValidationResult {
|
||||
isValid: boolean
|
||||
error?: string
|
||||
@@ -60,7 +54,7 @@ interface DebugValidationResult {
|
||||
|
||||
interface BlockEventHandlerConfig {
|
||||
workflowId?: string
|
||||
executionIdRef: { current: string }
|
||||
executionId?: string
|
||||
workflowEdges: Array<{ id: string; target: string; sourceHandle?: string | null }>
|
||||
activeBlocksSet: Set<string>
|
||||
accumulatedBlockLogs: BlockLog[]
|
||||
@@ -114,15 +108,12 @@ export function useWorkflowExecution() {
|
||||
const queryClient = useQueryClient()
|
||||
const currentWorkflow = useCurrentWorkflow()
|
||||
const { activeWorkflowId, workflows } = useWorkflowRegistry()
|
||||
const { toggleConsole, addConsole, updateConsole, cancelRunningEntries, clearExecutionEntries } =
|
||||
const { toggleConsole, addConsole, updateConsole, cancelRunningEntries } =
|
||||
useTerminalConsoleStore()
|
||||
const hasHydrated = useTerminalConsoleStore((s) => s._hasHydrated)
|
||||
const { getAllVariables } = useEnvironmentStore()
|
||||
const { getVariablesByWorkflowId, variables } = useVariablesStore()
|
||||
const { isExecuting, isDebugging, pendingBlocks, executor, debugContext } =
|
||||
useCurrentWorkflowExecution()
|
||||
const setCurrentExecutionId = useExecutionStore((s) => s.setCurrentExecutionId)
|
||||
const getCurrentExecutionId = useExecutionStore((s) => s.getCurrentExecutionId)
|
||||
const setIsExecuting = useExecutionStore((s) => s.setIsExecuting)
|
||||
const setIsDebugging = useExecutionStore((s) => s.setIsDebugging)
|
||||
const setPendingBlocks = useExecutionStore((s) => s.setPendingBlocks)
|
||||
@@ -306,7 +297,7 @@ export function useWorkflowExecution() {
|
||||
(config: BlockEventHandlerConfig) => {
|
||||
const {
|
||||
workflowId,
|
||||
executionIdRef,
|
||||
executionId,
|
||||
workflowEdges,
|
||||
activeBlocksSet,
|
||||
accumulatedBlockLogs,
|
||||
@@ -317,14 +308,6 @@ export function useWorkflowExecution() {
|
||||
onBlockCompleteCallback,
|
||||
} = config
|
||||
|
||||
/** Returns true if this execution was cancelled or superseded by another run. */
|
||||
const isStaleExecution = () =>
|
||||
!!(
|
||||
workflowId &&
|
||||
executionIdRef.current &&
|
||||
useExecutionStore.getState().getCurrentExecutionId(workflowId) !== executionIdRef.current
|
||||
)
|
||||
|
||||
const updateActiveBlocks = (blockId: string, isActive: boolean) => {
|
||||
if (!workflowId) return
|
||||
if (isActive) {
|
||||
@@ -377,7 +360,7 @@ export function useWorkflowExecution() {
|
||||
endedAt: data.endedAt,
|
||||
workflowId,
|
||||
blockId: data.blockId,
|
||||
executionId: executionIdRef.current,
|
||||
executionId,
|
||||
blockName: data.blockName || 'Unknown Block',
|
||||
blockType: data.blockType || 'unknown',
|
||||
iterationCurrent: data.iterationCurrent,
|
||||
@@ -400,7 +383,7 @@ export function useWorkflowExecution() {
|
||||
endedAt: data.endedAt,
|
||||
workflowId,
|
||||
blockId: data.blockId,
|
||||
executionId: executionIdRef.current,
|
||||
executionId,
|
||||
blockName: data.blockName || 'Unknown Block',
|
||||
blockType: data.blockType || 'unknown',
|
||||
iterationCurrent: data.iterationCurrent,
|
||||
@@ -427,7 +410,7 @@ export function useWorkflowExecution() {
|
||||
iterationType: data.iterationType,
|
||||
iterationContainerId: data.iterationContainerId,
|
||||
},
|
||||
executionIdRef.current
|
||||
executionId
|
||||
)
|
||||
}
|
||||
|
||||
@@ -449,12 +432,11 @@ export function useWorkflowExecution() {
|
||||
iterationType: data.iterationType,
|
||||
iterationContainerId: data.iterationContainerId,
|
||||
},
|
||||
executionIdRef.current
|
||||
executionId
|
||||
)
|
||||
}
|
||||
|
||||
const onBlockStarted = (data: BlockStartedData) => {
|
||||
if (isStaleExecution()) return
|
||||
updateActiveBlocks(data.blockId, true)
|
||||
markIncomingEdges(data.blockId)
|
||||
|
||||
@@ -471,7 +453,7 @@ export function useWorkflowExecution() {
|
||||
endedAt: undefined,
|
||||
workflowId,
|
||||
blockId: data.blockId,
|
||||
executionId: executionIdRef.current,
|
||||
executionId,
|
||||
blockName: data.blockName || 'Unknown Block',
|
||||
blockType: data.blockType || 'unknown',
|
||||
isRunning: true,
|
||||
@@ -483,7 +465,6 @@ export function useWorkflowExecution() {
|
||||
}
|
||||
|
||||
const onBlockCompleted = (data: BlockCompletedData) => {
|
||||
if (isStaleExecution()) return
|
||||
updateActiveBlocks(data.blockId, false)
|
||||
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'success')
|
||||
|
||||
@@ -514,7 +495,6 @@ export function useWorkflowExecution() {
|
||||
}
|
||||
|
||||
const onBlockError = (data: BlockErrorData) => {
|
||||
if (isStaleExecution()) return
|
||||
updateActiveBlocks(data.blockId, false)
|
||||
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'error')
|
||||
|
||||
@@ -922,6 +902,10 @@ export function useWorkflowExecution() {
|
||||
|
||||
// Update block logs with actual stream completion times
|
||||
if (result.logs && streamCompletionTimes.size > 0) {
|
||||
const streamCompletionEndTime = new Date(
|
||||
Math.max(...Array.from(streamCompletionTimes.values()))
|
||||
).toISOString()
|
||||
|
||||
result.logs.forEach((log: BlockLog) => {
|
||||
if (streamCompletionTimes.has(log.blockId)) {
|
||||
const completionTime = streamCompletionTimes.get(log.blockId)!
|
||||
@@ -1003,6 +987,7 @@ export function useWorkflowExecution() {
|
||||
return { success: true, stream }
|
||||
}
|
||||
|
||||
// For manual (non-chat) execution
|
||||
const manualExecutionId = uuidv4()
|
||||
try {
|
||||
const result = await executeWorkflow(
|
||||
@@ -1017,10 +1002,29 @@ export function useWorkflowExecution() {
|
||||
if (result.metadata.pendingBlocks) {
|
||||
setPendingBlocks(activeWorkflowId, result.metadata.pendingBlocks)
|
||||
}
|
||||
} else if (result && 'success' in result) {
|
||||
setExecutionResult(result)
|
||||
// Reset execution state after successful non-debug execution
|
||||
setIsExecuting(activeWorkflowId, false)
|
||||
setIsDebugging(activeWorkflowId, false)
|
||||
setActiveBlocks(activeWorkflowId, new Set())
|
||||
|
||||
if (isChatExecution) {
|
||||
if (!result.metadata) {
|
||||
result.metadata = { duration: 0, startTime: new Date().toISOString() }
|
||||
}
|
||||
;(result.metadata as any).source = 'chat'
|
||||
}
|
||||
|
||||
// Invalidate subscription queries to update usage
|
||||
setTimeout(() => {
|
||||
queryClient.invalidateQueries({ queryKey: subscriptionKeys.all })
|
||||
}, 1000)
|
||||
}
|
||||
return result
|
||||
} catch (error: any) {
|
||||
const errorResult = handleExecutionError(error, { executionId: manualExecutionId })
|
||||
// Note: Error logs are already persisted server-side via execution-core.ts
|
||||
return errorResult
|
||||
}
|
||||
},
|
||||
@@ -1271,7 +1275,7 @@ export function useWorkflowExecution() {
|
||||
if (activeWorkflowId) {
|
||||
logger.info('Using server-side executor')
|
||||
|
||||
const executionIdRef = { current: '' }
|
||||
const executionId = uuidv4()
|
||||
|
||||
let executionResult: ExecutionResult = {
|
||||
success: false,
|
||||
@@ -1289,7 +1293,7 @@ export function useWorkflowExecution() {
|
||||
try {
|
||||
const blockHandlers = buildBlockEventHandlers({
|
||||
workflowId: activeWorkflowId,
|
||||
executionIdRef,
|
||||
executionId,
|
||||
workflowEdges,
|
||||
activeBlocksSet,
|
||||
accumulatedBlockLogs,
|
||||
@@ -1322,10 +1326,6 @@ export function useWorkflowExecution() {
|
||||
loops: clientWorkflowState.loops,
|
||||
parallels: clientWorkflowState.parallels,
|
||||
},
|
||||
onExecutionId: (id) => {
|
||||
executionIdRef.current = id
|
||||
setCurrentExecutionId(activeWorkflowId, id)
|
||||
},
|
||||
callbacks: {
|
||||
onExecutionStarted: (data) => {
|
||||
logger.info('Server execution started:', data)
|
||||
@@ -1368,18 +1368,6 @@ export function useWorkflowExecution() {
|
||||
},
|
||||
|
||||
onExecutionCompleted: (data) => {
|
||||
if (
|
||||
activeWorkflowId &&
|
||||
executionIdRef.current &&
|
||||
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
||||
executionIdRef.current
|
||||
)
|
||||
return
|
||||
|
||||
if (activeWorkflowId) {
|
||||
setCurrentExecutionId(activeWorkflowId, null)
|
||||
}
|
||||
|
||||
executionResult = {
|
||||
success: data.success,
|
||||
output: data.output,
|
||||
@@ -1437,33 +1425,9 @@ export function useWorkflowExecution() {
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
const workflowExecState = activeWorkflowId
|
||||
? useExecutionStore.getState().getWorkflowExecution(activeWorkflowId)
|
||||
: null
|
||||
if (activeWorkflowId && !workflowExecState?.isDebugging) {
|
||||
setExecutionResult(executionResult)
|
||||
setIsExecuting(activeWorkflowId, false)
|
||||
setActiveBlocks(activeWorkflowId, new Set())
|
||||
setTimeout(() => {
|
||||
queryClient.invalidateQueries({ queryKey: subscriptionKeys.all })
|
||||
}, 1000)
|
||||
}
|
||||
},
|
||||
|
||||
onExecutionError: (data) => {
|
||||
if (
|
||||
activeWorkflowId &&
|
||||
executionIdRef.current &&
|
||||
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
||||
executionIdRef.current
|
||||
)
|
||||
return
|
||||
|
||||
if (activeWorkflowId) {
|
||||
setCurrentExecutionId(activeWorkflowId, null)
|
||||
}
|
||||
|
||||
executionResult = {
|
||||
success: false,
|
||||
output: {},
|
||||
@@ -1477,53 +1441,43 @@ export function useWorkflowExecution() {
|
||||
const isPreExecutionError = accumulatedBlockLogs.length === 0
|
||||
handleExecutionErrorConsole({
|
||||
workflowId: activeWorkflowId,
|
||||
executionId: executionIdRef.current,
|
||||
executionId,
|
||||
error: data.error,
|
||||
durationMs: data.duration,
|
||||
blockLogs: accumulatedBlockLogs,
|
||||
isPreExecutionError,
|
||||
})
|
||||
|
||||
if (activeWorkflowId) {
|
||||
setIsExecuting(activeWorkflowId, false)
|
||||
setIsDebugging(activeWorkflowId, false)
|
||||
setActiveBlocks(activeWorkflowId, new Set())
|
||||
}
|
||||
},
|
||||
|
||||
onExecutionCancelled: (data) => {
|
||||
if (
|
||||
activeWorkflowId &&
|
||||
executionIdRef.current &&
|
||||
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
||||
executionIdRef.current
|
||||
)
|
||||
return
|
||||
|
||||
if (activeWorkflowId) {
|
||||
setCurrentExecutionId(activeWorkflowId, null)
|
||||
}
|
||||
|
||||
handleExecutionCancelledConsole({
|
||||
workflowId: activeWorkflowId,
|
||||
executionId: executionIdRef.current,
|
||||
executionId,
|
||||
durationMs: data?.duration,
|
||||
})
|
||||
|
||||
if (activeWorkflowId) {
|
||||
setIsExecuting(activeWorkflowId, false)
|
||||
setIsDebugging(activeWorkflowId, false)
|
||||
setActiveBlocks(activeWorkflowId, new Set())
|
||||
}
|
||||
},
|
||||
},
|
||||
})
|
||||
|
||||
return executionResult
|
||||
} catch (error: any) {
|
||||
// Don't log abort errors - they're intentional user actions
|
||||
if (error.name === 'AbortError' || error.message?.includes('aborted')) {
|
||||
logger.info('Execution aborted by user')
|
||||
return executionResult
|
||||
|
||||
// Reset execution state
|
||||
if (activeWorkflowId) {
|
||||
setIsExecuting(activeWorkflowId, false)
|
||||
setActiveBlocks(activeWorkflowId, new Set())
|
||||
}
|
||||
|
||||
// Return gracefully without error
|
||||
return {
|
||||
success: false,
|
||||
output: {},
|
||||
metadata: { duration: 0 },
|
||||
logs: [],
|
||||
}
|
||||
}
|
||||
|
||||
logger.error('Server-side execution failed:', error)
|
||||
@@ -1531,6 +1485,7 @@ export function useWorkflowExecution() {
|
||||
}
|
||||
}
|
||||
|
||||
// Fallback: should never reach here
|
||||
throw new Error('Server-side execution is required')
|
||||
}
|
||||
|
||||
@@ -1762,28 +1717,25 @@ export function useWorkflowExecution() {
|
||||
* Handles cancelling the current workflow execution
|
||||
*/
|
||||
const handleCancelExecution = useCallback(() => {
|
||||
if (!activeWorkflowId) return
|
||||
logger.info('Workflow execution cancellation requested')
|
||||
|
||||
const storedExecutionId = getCurrentExecutionId(activeWorkflowId)
|
||||
// Cancel the execution stream for this workflow (server-side)
|
||||
executionStream.cancel(activeWorkflowId ?? undefined)
|
||||
|
||||
if (storedExecutionId) {
|
||||
setCurrentExecutionId(activeWorkflowId, null)
|
||||
fetch(`/api/workflows/${activeWorkflowId}/executions/${storedExecutionId}/cancel`, {
|
||||
method: 'POST',
|
||||
}).catch(() => {})
|
||||
handleExecutionCancelledConsole({
|
||||
workflowId: activeWorkflowId,
|
||||
executionId: storedExecutionId,
|
||||
})
|
||||
// Mark current chat execution as superseded so its cleanup won't affect new executions
|
||||
currentChatExecutionIdRef.current = null
|
||||
|
||||
// Mark all running entries as canceled in the terminal
|
||||
if (activeWorkflowId) {
|
||||
cancelRunningEntries(activeWorkflowId)
|
||||
|
||||
// Reset execution state - this triggers chat stream cleanup via useEffect in chat.tsx
|
||||
setIsExecuting(activeWorkflowId, false)
|
||||
setIsDebugging(activeWorkflowId, false)
|
||||
setActiveBlocks(activeWorkflowId, new Set())
|
||||
}
|
||||
|
||||
executionStream.cancel(activeWorkflowId)
|
||||
currentChatExecutionIdRef.current = null
|
||||
setIsExecuting(activeWorkflowId, false)
|
||||
setIsDebugging(activeWorkflowId, false)
|
||||
setActiveBlocks(activeWorkflowId, new Set())
|
||||
|
||||
// If in debug mode, also reset debug state
|
||||
if (isDebugging) {
|
||||
resetDebugState()
|
||||
}
|
||||
@@ -1795,9 +1747,7 @@ export function useWorkflowExecution() {
|
||||
setIsDebugging,
|
||||
setActiveBlocks,
|
||||
activeWorkflowId,
|
||||
getCurrentExecutionId,
|
||||
setCurrentExecutionId,
|
||||
handleExecutionCancelledConsole,
|
||||
cancelRunningEntries,
|
||||
])
|
||||
|
||||
/**
|
||||
@@ -1897,7 +1847,7 @@ export function useWorkflowExecution() {
|
||||
}
|
||||
|
||||
setIsExecuting(workflowId, true)
|
||||
const executionIdRef = { current: '' }
|
||||
const executionId = uuidv4()
|
||||
const accumulatedBlockLogs: BlockLog[] = []
|
||||
const accumulatedBlockStates = new Map<string, BlockState>()
|
||||
const executedBlockIds = new Set<string>()
|
||||
@@ -1906,7 +1856,7 @@ export function useWorkflowExecution() {
|
||||
try {
|
||||
const blockHandlers = buildBlockEventHandlers({
|
||||
workflowId,
|
||||
executionIdRef,
|
||||
executionId,
|
||||
workflowEdges,
|
||||
activeBlocksSet,
|
||||
accumulatedBlockLogs,
|
||||
@@ -1921,10 +1871,6 @@ export function useWorkflowExecution() {
|
||||
startBlockId: blockId,
|
||||
sourceSnapshot: effectiveSnapshot,
|
||||
input: workflowInput,
|
||||
onExecutionId: (id) => {
|
||||
executionIdRef.current = id
|
||||
setCurrentExecutionId(workflowId, id)
|
||||
},
|
||||
callbacks: {
|
||||
onBlockStarted: blockHandlers.onBlockStarted,
|
||||
onBlockCompleted: blockHandlers.onBlockCompleted,
|
||||
@@ -1932,6 +1878,7 @@ export function useWorkflowExecution() {
|
||||
|
||||
onExecutionCompleted: (data) => {
|
||||
if (data.success) {
|
||||
// Add the start block (trigger) to executed blocks
|
||||
executedBlockIds.add(blockId)
|
||||
|
||||
const mergedBlockStates: Record<string, BlockState> = {
|
||||
@@ -1955,10 +1902,6 @@ export function useWorkflowExecution() {
|
||||
}
|
||||
setLastExecutionSnapshot(workflowId, updatedSnapshot)
|
||||
}
|
||||
|
||||
setCurrentExecutionId(workflowId, null)
|
||||
setIsExecuting(workflowId, false)
|
||||
setActiveBlocks(workflowId, new Set())
|
||||
},
|
||||
|
||||
onExecutionError: (data) => {
|
||||
@@ -1978,27 +1921,19 @@ export function useWorkflowExecution() {
|
||||
|
||||
handleExecutionErrorConsole({
|
||||
workflowId,
|
||||
executionId: executionIdRef.current,
|
||||
executionId,
|
||||
error: data.error,
|
||||
durationMs: data.duration,
|
||||
blockLogs: accumulatedBlockLogs,
|
||||
})
|
||||
|
||||
setCurrentExecutionId(workflowId, null)
|
||||
setIsExecuting(workflowId, false)
|
||||
setActiveBlocks(workflowId, new Set())
|
||||
},
|
||||
|
||||
onExecutionCancelled: (data) => {
|
||||
handleExecutionCancelledConsole({
|
||||
workflowId,
|
||||
executionId: executionIdRef.current,
|
||||
executionId,
|
||||
durationMs: data?.duration,
|
||||
})
|
||||
|
||||
setCurrentExecutionId(workflowId, null)
|
||||
setIsExecuting(workflowId, false)
|
||||
setActiveBlocks(workflowId, new Set())
|
||||
},
|
||||
},
|
||||
})
|
||||
@@ -2007,20 +1942,14 @@ export function useWorkflowExecution() {
|
||||
logger.error('Run-from-block failed:', error)
|
||||
}
|
||||
} finally {
|
||||
const currentId = getCurrentExecutionId(workflowId)
|
||||
if (currentId === null || currentId === executionIdRef.current) {
|
||||
setCurrentExecutionId(workflowId, null)
|
||||
setIsExecuting(workflowId, false)
|
||||
setActiveBlocks(workflowId, new Set())
|
||||
}
|
||||
setIsExecuting(workflowId, false)
|
||||
setActiveBlocks(workflowId, new Set())
|
||||
}
|
||||
},
|
||||
[
|
||||
getLastExecutionSnapshot,
|
||||
setLastExecutionSnapshot,
|
||||
clearLastExecutionSnapshot,
|
||||
getCurrentExecutionId,
|
||||
setCurrentExecutionId,
|
||||
setIsExecuting,
|
||||
setActiveBlocks,
|
||||
setBlockRunStatus,
|
||||
@@ -2050,213 +1979,29 @@ export function useWorkflowExecution() {
|
||||
|
||||
const executionId = uuidv4()
|
||||
try {
|
||||
await executeWorkflow(undefined, undefined, executionId, undefined, 'manual', blockId)
|
||||
const result = await executeWorkflow(
|
||||
undefined,
|
||||
undefined,
|
||||
executionId,
|
||||
undefined,
|
||||
'manual',
|
||||
blockId
|
||||
)
|
||||
if (result && 'success' in result) {
|
||||
setExecutionResult(result)
|
||||
}
|
||||
} catch (error) {
|
||||
const errorResult = handleExecutionError(error, { executionId })
|
||||
return errorResult
|
||||
} finally {
|
||||
setCurrentExecutionId(workflowId, null)
|
||||
setIsExecuting(workflowId, false)
|
||||
setIsDebugging(workflowId, false)
|
||||
setActiveBlocks(workflowId, new Set())
|
||||
}
|
||||
},
|
||||
[
|
||||
activeWorkflowId,
|
||||
setCurrentExecutionId,
|
||||
setExecutionResult,
|
||||
setIsExecuting,
|
||||
setIsDebugging,
|
||||
setActiveBlocks,
|
||||
]
|
||||
[activeWorkflowId, setExecutionResult, setIsExecuting, setIsDebugging, setActiveBlocks]
|
||||
)
|
||||
|
||||
useEffect(() => {
|
||||
if (!activeWorkflowId || !hasHydrated) return
|
||||
|
||||
const entries = useTerminalConsoleStore.getState().entries
|
||||
const runningEntries = entries.filter(
|
||||
(e) => e.isRunning && e.workflowId === activeWorkflowId && e.executionId
|
||||
)
|
||||
if (runningEntries.length === 0) return
|
||||
|
||||
if (activeReconnections.has(activeWorkflowId)) return
|
||||
activeReconnections.add(activeWorkflowId)
|
||||
|
||||
executionStream.cancel(activeWorkflowId)
|
||||
|
||||
const sorted = [...runningEntries].sort((a, b) => {
|
||||
const aTime = a.startedAt ? new Date(a.startedAt).getTime() : 0
|
||||
const bTime = b.startedAt ? new Date(b.startedAt).getTime() : 0
|
||||
return bTime - aTime
|
||||
})
|
||||
const executionId = sorted[0].executionId!
|
||||
|
||||
const otherExecutionIds = new Set(
|
||||
sorted.filter((e) => e.executionId !== executionId).map((e) => e.executionId!)
|
||||
)
|
||||
if (otherExecutionIds.size > 0) {
|
||||
cancelRunningEntries(activeWorkflowId)
|
||||
}
|
||||
|
||||
setCurrentExecutionId(activeWorkflowId, executionId)
|
||||
setIsExecuting(activeWorkflowId, true)
|
||||
|
||||
const workflowEdges = useWorkflowStore.getState().edges
|
||||
const activeBlocksSet = new Set<string>()
|
||||
const accumulatedBlockLogs: BlockLog[] = []
|
||||
const accumulatedBlockStates = new Map<string, BlockState>()
|
||||
const executedBlockIds = new Set<string>()
|
||||
|
||||
const executionIdRef = { current: executionId }
|
||||
|
||||
const handlers = buildBlockEventHandlers({
|
||||
workflowId: activeWorkflowId,
|
||||
executionIdRef,
|
||||
workflowEdges,
|
||||
activeBlocksSet,
|
||||
accumulatedBlockLogs,
|
||||
accumulatedBlockStates,
|
||||
executedBlockIds,
|
||||
consoleMode: 'update',
|
||||
includeStartConsoleEntry: true,
|
||||
})
|
||||
|
||||
const originalEntries = entries
|
||||
.filter((e) => e.executionId === executionId)
|
||||
.map((e) => ({ ...e }))
|
||||
|
||||
let cleared = false
|
||||
let reconnectionComplete = false
|
||||
let cleanupRan = false
|
||||
const clearOnce = () => {
|
||||
if (!cleared) {
|
||||
cleared = true
|
||||
clearExecutionEntries(executionId)
|
||||
}
|
||||
}
|
||||
|
||||
const reconnectWorkflowId = activeWorkflowId
|
||||
|
||||
executionStream
|
||||
.reconnect({
|
||||
workflowId: reconnectWorkflowId,
|
||||
executionId,
|
||||
callbacks: {
|
||||
onBlockStarted: (data) => {
|
||||
clearOnce()
|
||||
handlers.onBlockStarted(data)
|
||||
},
|
||||
onBlockCompleted: (data) => {
|
||||
clearOnce()
|
||||
handlers.onBlockCompleted(data)
|
||||
},
|
||||
onBlockError: (data) => {
|
||||
clearOnce()
|
||||
handlers.onBlockError(data)
|
||||
},
|
||||
onExecutionCompleted: () => {
|
||||
const currentId = useExecutionStore
|
||||
.getState()
|
||||
.getCurrentExecutionId(reconnectWorkflowId)
|
||||
if (currentId !== executionId) {
|
||||
reconnectionComplete = true
|
||||
activeReconnections.delete(reconnectWorkflowId)
|
||||
return
|
||||
}
|
||||
clearOnce()
|
||||
reconnectionComplete = true
|
||||
activeReconnections.delete(reconnectWorkflowId)
|
||||
setCurrentExecutionId(reconnectWorkflowId, null)
|
||||
setIsExecuting(reconnectWorkflowId, false)
|
||||
setActiveBlocks(reconnectWorkflowId, new Set())
|
||||
},
|
||||
onExecutionError: (data) => {
|
||||
const currentId = useExecutionStore
|
||||
.getState()
|
||||
.getCurrentExecutionId(reconnectWorkflowId)
|
||||
if (currentId !== executionId) {
|
||||
reconnectionComplete = true
|
||||
activeReconnections.delete(reconnectWorkflowId)
|
||||
return
|
||||
}
|
||||
clearOnce()
|
||||
reconnectionComplete = true
|
||||
activeReconnections.delete(reconnectWorkflowId)
|
||||
setCurrentExecutionId(reconnectWorkflowId, null)
|
||||
setIsExecuting(reconnectWorkflowId, false)
|
||||
setActiveBlocks(reconnectWorkflowId, new Set())
|
||||
handleExecutionErrorConsole({
|
||||
workflowId: reconnectWorkflowId,
|
||||
executionId,
|
||||
error: data.error,
|
||||
blockLogs: accumulatedBlockLogs,
|
||||
})
|
||||
},
|
||||
onExecutionCancelled: () => {
|
||||
const currentId = useExecutionStore
|
||||
.getState()
|
||||
.getCurrentExecutionId(reconnectWorkflowId)
|
||||
if (currentId !== executionId) {
|
||||
reconnectionComplete = true
|
||||
activeReconnections.delete(reconnectWorkflowId)
|
||||
return
|
||||
}
|
||||
clearOnce()
|
||||
reconnectionComplete = true
|
||||
activeReconnections.delete(reconnectWorkflowId)
|
||||
setCurrentExecutionId(reconnectWorkflowId, null)
|
||||
setIsExecuting(reconnectWorkflowId, false)
|
||||
setActiveBlocks(reconnectWorkflowId, new Set())
|
||||
handleExecutionCancelledConsole({
|
||||
workflowId: reconnectWorkflowId,
|
||||
executionId,
|
||||
})
|
||||
},
|
||||
},
|
||||
})
|
||||
.catch((error) => {
|
||||
logger.warn('Execution reconnection failed', { executionId, error })
|
||||
})
|
||||
.finally(() => {
|
||||
if (reconnectionComplete || cleanupRan) return
|
||||
const currentId = useExecutionStore.getState().getCurrentExecutionId(reconnectWorkflowId)
|
||||
if (currentId !== executionId) return
|
||||
reconnectionComplete = true
|
||||
activeReconnections.delete(reconnectWorkflowId)
|
||||
clearExecutionEntries(executionId)
|
||||
for (const entry of originalEntries) {
|
||||
addConsole({
|
||||
workflowId: entry.workflowId,
|
||||
blockId: entry.blockId,
|
||||
blockName: entry.blockName,
|
||||
blockType: entry.blockType,
|
||||
executionId: entry.executionId,
|
||||
executionOrder: entry.executionOrder,
|
||||
isRunning: false,
|
||||
warning: 'Execution result unavailable — check the logs page',
|
||||
})
|
||||
}
|
||||
setCurrentExecutionId(reconnectWorkflowId, null)
|
||||
setIsExecuting(reconnectWorkflowId, false)
|
||||
setActiveBlocks(reconnectWorkflowId, new Set())
|
||||
})
|
||||
|
||||
return () => {
|
||||
cleanupRan = true
|
||||
executionStream.cancel(reconnectWorkflowId)
|
||||
activeReconnections.delete(reconnectWorkflowId)
|
||||
|
||||
if (cleared && !reconnectionComplete) {
|
||||
clearExecutionEntries(executionId)
|
||||
for (const entry of originalEntries) {
|
||||
addConsole(entry)
|
||||
}
|
||||
}
|
||||
}
|
||||
// eslint-disable-next-line react-hooks/exhaustive-deps
|
||||
}, [activeWorkflowId, hasHydrated])
|
||||
|
||||
return {
|
||||
isExecuting,
|
||||
isDebugging,
|
||||
|
||||
@@ -13,7 +13,7 @@ import {
|
||||
ModalFooter,
|
||||
ModalHeader,
|
||||
} from '@/components/emcn'
|
||||
import { AnthropicIcon, GeminiIcon, MistralIcon, OpenAIIcon, SerperIcon } from '@/components/icons'
|
||||
import { AnthropicIcon, GeminiIcon, MistralIcon, OpenAIIcon } from '@/components/icons'
|
||||
import { Skeleton } from '@/components/ui'
|
||||
import {
|
||||
type BYOKKey,
|
||||
@@ -60,13 +60,6 @@ const PROVIDERS: {
|
||||
description: 'LLM calls and Knowledge Base OCR',
|
||||
placeholder: 'Enter your API key',
|
||||
},
|
||||
{
|
||||
id: 'serper',
|
||||
name: 'Serper',
|
||||
icon: SerperIcon,
|
||||
description: 'Web search tool',
|
||||
placeholder: 'Enter your Serper API key',
|
||||
},
|
||||
]
|
||||
|
||||
function BYOKKeySkeleton() {
|
||||
|
||||
@@ -58,16 +58,6 @@ export const S3Block: BlockConfig<S3Response> = {
|
||||
},
|
||||
required: true,
|
||||
},
|
||||
{
|
||||
id: 'getObjectRegion',
|
||||
title: 'AWS Region',
|
||||
type: 'short-input',
|
||||
placeholder: 'Used when S3 URL does not include region',
|
||||
condition: {
|
||||
field: 'operation',
|
||||
value: ['get_object'],
|
||||
},
|
||||
},
|
||||
{
|
||||
id: 'bucketName',
|
||||
title: 'Bucket Name',
|
||||
@@ -301,11 +291,34 @@ export const S3Block: BlockConfig<S3Response> = {
|
||||
if (!params.s3Uri) {
|
||||
throw new Error('S3 Object URL is required')
|
||||
}
|
||||
return {
|
||||
accessKeyId: params.accessKeyId,
|
||||
secretAccessKey: params.secretAccessKey,
|
||||
region: params.getObjectRegion || params.region,
|
||||
s3Uri: params.s3Uri,
|
||||
|
||||
// Parse S3 URI for get_object
|
||||
try {
|
||||
const url = new URL(params.s3Uri)
|
||||
const hostname = url.hostname
|
||||
const bucketName = hostname.split('.')[0]
|
||||
const regionMatch = hostname.match(/s3[.-]([^.]+)\.amazonaws\.com/)
|
||||
const region = regionMatch ? regionMatch[1] : params.region
|
||||
const objectKey = url.pathname.startsWith('/')
|
||||
? url.pathname.substring(1)
|
||||
: url.pathname
|
||||
|
||||
if (!bucketName || !objectKey) {
|
||||
throw new Error('Could not parse S3 URL')
|
||||
}
|
||||
|
||||
return {
|
||||
accessKeyId: params.accessKeyId,
|
||||
secretAccessKey: params.secretAccessKey,
|
||||
region,
|
||||
bucketName,
|
||||
objectKey,
|
||||
s3Uri: params.s3Uri,
|
||||
}
|
||||
} catch (_error) {
|
||||
throw new Error(
|
||||
'Invalid S3 Object URL format. Expected: https://bucket-name.s3.region.amazonaws.com/path/to/file'
|
||||
)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -388,7 +401,6 @@ export const S3Block: BlockConfig<S3Response> = {
|
||||
acl: { type: 'string', description: 'Access control list' },
|
||||
// Download inputs
|
||||
s3Uri: { type: 'string', description: 'S3 object URL' },
|
||||
getObjectRegion: { type: 'string', description: 'Optional AWS region override for downloads' },
|
||||
// List inputs
|
||||
prefix: { type: 'string', description: 'Prefix filter' },
|
||||
maxKeys: { type: 'number', description: 'Maximum results' },
|
||||
|
||||
@@ -78,7 +78,6 @@ export const SerperBlock: BlockConfig<SearchResponse> = {
|
||||
placeholder: 'Enter your Serper API key',
|
||||
password: true,
|
||||
required: true,
|
||||
hideWhenHosted: true,
|
||||
},
|
||||
],
|
||||
tools: {
|
||||
|
||||
@@ -243,7 +243,6 @@ export interface SubBlockConfig {
|
||||
hidden?: boolean
|
||||
hideFromPreview?: boolean // Hide this subblock from the workflow block preview
|
||||
requiresFeature?: string // Environment variable name that must be truthy for this subblock to be visible
|
||||
hideWhenHosted?: boolean // Hide this subblock when running on hosted sim
|
||||
description?: string
|
||||
tooltip?: string // Tooltip text displayed via info icon next to the title
|
||||
value?: (params: Record<string, any>) => string
|
||||
|
||||
@@ -1,4 +1,3 @@
|
||||
import { setupGlobalFetchMock } from '@sim/testing'
|
||||
import { afterEach, beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
||||
import { getAllBlocks } from '@/blocks'
|
||||
import { BlockType, isMcpTool } from '@/executor/constants'
|
||||
@@ -62,30 +61,6 @@ vi.mock('@/providers', () => ({
|
||||
}),
|
||||
}))
|
||||
|
||||
vi.mock('@/executor/utils/http', () => ({
|
||||
buildAuthHeaders: vi.fn().mockResolvedValue({ 'Content-Type': 'application/json' }),
|
||||
buildAPIUrl: vi.fn((path: string, params?: Record<string, string>) => {
|
||||
const url = new URL(path, 'http://localhost:3000')
|
||||
if (params) {
|
||||
for (const [key, value] of Object.entries(params)) {
|
||||
if (value !== undefined && value !== null) {
|
||||
url.searchParams.set(key, value)
|
||||
}
|
||||
}
|
||||
}
|
||||
return url
|
||||
}),
|
||||
extractAPIErrorMessage: vi.fn(async (response: Response) => {
|
||||
const defaultMessage = `API request failed with status ${response.status}`
|
||||
try {
|
||||
const errorData = await response.json()
|
||||
return errorData.error || defaultMessage
|
||||
} catch {
|
||||
return defaultMessage
|
||||
}
|
||||
}),
|
||||
}))
|
||||
|
||||
vi.mock('@sim/db', () => ({
|
||||
db: {
|
||||
select: vi.fn().mockReturnValue({
|
||||
@@ -109,7 +84,7 @@ vi.mock('@sim/db/schema', () => ({
|
||||
},
|
||||
}))
|
||||
|
||||
setupGlobalFetchMock()
|
||||
global.fetch = Object.assign(vi.fn(), { preconnect: vi.fn() }) as typeof fetch
|
||||
|
||||
const mockGetAllBlocks = getAllBlocks as Mock
|
||||
const mockExecuteTool = executeTool as Mock
|
||||
@@ -1926,301 +1901,5 @@ describe('AgentBlockHandler', () => {
|
||||
|
||||
expect(discoveryCalls[0].url).toContain('serverId=mcp-legacy-server')
|
||||
})
|
||||
|
||||
describe('customToolId resolution - DB as source of truth', () => {
|
||||
const staleInlineSchema = {
|
||||
function: {
|
||||
name: 'formatReport',
|
||||
description: 'Formats a report',
|
||||
parameters: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
title: { type: 'string', description: 'Report title' },
|
||||
content: { type: 'string', description: 'Report content' },
|
||||
},
|
||||
required: ['title', 'content'],
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
const dbSchema = {
|
||||
function: {
|
||||
name: 'formatReport',
|
||||
description: 'Formats a report',
|
||||
parameters: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
title: { type: 'string', description: 'Report title' },
|
||||
content: { type: 'string', description: 'Report content' },
|
||||
format: { type: 'string', description: 'Output format' },
|
||||
},
|
||||
required: ['title', 'content', 'format'],
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
const staleInlineCode = 'return { title, content };'
|
||||
const dbCode = 'return { title, content, format };'
|
||||
|
||||
function mockFetchForCustomTool(toolId: string) {
|
||||
mockFetch.mockImplementation((url: string) => {
|
||||
if (typeof url === 'string' && url.includes('/api/tools/custom')) {
|
||||
return Promise.resolve({
|
||||
ok: true,
|
||||
headers: { get: () => null },
|
||||
json: () =>
|
||||
Promise.resolve({
|
||||
data: [
|
||||
{
|
||||
id: toolId,
|
||||
title: 'formatReport',
|
||||
schema: dbSchema,
|
||||
code: dbCode,
|
||||
},
|
||||
],
|
||||
}),
|
||||
})
|
||||
}
|
||||
return Promise.resolve({
|
||||
ok: true,
|
||||
headers: { get: () => null },
|
||||
json: () => Promise.resolve({}),
|
||||
})
|
||||
})
|
||||
}
|
||||
|
||||
function mockFetchFailure() {
|
||||
mockFetch.mockImplementation((url: string) => {
|
||||
if (typeof url === 'string' && url.includes('/api/tools/custom')) {
|
||||
return Promise.resolve({
|
||||
ok: false,
|
||||
status: 500,
|
||||
headers: { get: () => null },
|
||||
json: () => Promise.resolve({}),
|
||||
})
|
||||
}
|
||||
return Promise.resolve({
|
||||
ok: true,
|
||||
headers: { get: () => null },
|
||||
json: () => Promise.resolve({}),
|
||||
})
|
||||
})
|
||||
}
|
||||
|
||||
beforeEach(() => {
|
||||
Object.defineProperty(global, 'window', {
|
||||
value: undefined,
|
||||
writable: true,
|
||||
configurable: true,
|
||||
})
|
||||
})
|
||||
|
||||
it('should always fetch latest schema from DB when customToolId is present', async () => {
|
||||
const toolId = 'custom-tool-123'
|
||||
mockFetchForCustomTool(toolId)
|
||||
|
||||
const inputs = {
|
||||
model: 'gpt-4o',
|
||||
userPrompt: 'Format a report',
|
||||
apiKey: 'test-api-key',
|
||||
tools: [
|
||||
{
|
||||
type: 'custom-tool',
|
||||
customToolId: toolId,
|
||||
title: 'formatReport',
|
||||
schema: staleInlineSchema,
|
||||
code: staleInlineCode,
|
||||
usageControl: 'auto' as const,
|
||||
},
|
||||
],
|
||||
}
|
||||
|
||||
mockGetProviderFromModel.mockReturnValue('openai')
|
||||
|
||||
await handler.execute(mockContext, mockBlock, inputs)
|
||||
|
||||
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||
const tools = providerCall[1].tools
|
||||
|
||||
expect(tools.length).toBe(1)
|
||||
// DB schema wins over stale inline — includes format param
|
||||
expect(tools[0].parameters.required).toContain('format')
|
||||
expect(tools[0].parameters.properties).toHaveProperty('format')
|
||||
})
|
||||
|
||||
it('should fetch from DB when customToolId has no inline schema', async () => {
|
||||
const toolId = 'custom-tool-123'
|
||||
mockFetchForCustomTool(toolId)
|
||||
|
||||
const inputs = {
|
||||
model: 'gpt-4o',
|
||||
userPrompt: 'Format a report',
|
||||
apiKey: 'test-api-key',
|
||||
tools: [
|
||||
{
|
||||
type: 'custom-tool',
|
||||
customToolId: toolId,
|
||||
usageControl: 'auto' as const,
|
||||
},
|
||||
],
|
||||
}
|
||||
|
||||
mockGetProviderFromModel.mockReturnValue('openai')
|
||||
|
||||
await handler.execute(mockContext, mockBlock, inputs)
|
||||
|
||||
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||
const tools = providerCall[1].tools
|
||||
|
||||
expect(tools.length).toBe(1)
|
||||
expect(tools[0].name).toBe('formatReport')
|
||||
expect(tools[0].parameters.required).toContain('format')
|
||||
})
|
||||
|
||||
it('should fall back to inline schema when DB fetch fails and inline exists', async () => {
|
||||
mockFetchFailure()
|
||||
|
||||
const inputs = {
|
||||
model: 'gpt-4o',
|
||||
userPrompt: 'Format a report',
|
||||
apiKey: 'test-api-key',
|
||||
tools: [
|
||||
{
|
||||
type: 'custom-tool',
|
||||
customToolId: 'custom-tool-123',
|
||||
title: 'formatReport',
|
||||
schema: staleInlineSchema,
|
||||
code: staleInlineCode,
|
||||
usageControl: 'auto' as const,
|
||||
},
|
||||
],
|
||||
}
|
||||
|
||||
mockGetProviderFromModel.mockReturnValue('openai')
|
||||
|
||||
await handler.execute(mockContext, mockBlock, inputs)
|
||||
|
||||
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||
const tools = providerCall[1].tools
|
||||
|
||||
expect(tools.length).toBe(1)
|
||||
expect(tools[0].name).toBe('formatReport')
|
||||
expect(tools[0].parameters.required).not.toContain('format')
|
||||
})
|
||||
|
||||
it('should return null when DB fetch fails and no inline schema exists', async () => {
|
||||
mockFetchFailure()
|
||||
|
||||
const inputs = {
|
||||
model: 'gpt-4o',
|
||||
userPrompt: 'Format a report',
|
||||
apiKey: 'test-api-key',
|
||||
tools: [
|
||||
{
|
||||
type: 'custom-tool',
|
||||
customToolId: 'custom-tool-123',
|
||||
usageControl: 'auto' as const,
|
||||
},
|
||||
],
|
||||
}
|
||||
|
||||
mockGetProviderFromModel.mockReturnValue('openai')
|
||||
|
||||
await handler.execute(mockContext, mockBlock, inputs)
|
||||
|
||||
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||
const tools = providerCall[1].tools
|
||||
|
||||
expect(tools.length).toBe(0)
|
||||
})
|
||||
|
||||
it('should use DB code for executeFunction when customToolId resolves', async () => {
|
||||
const toolId = 'custom-tool-123'
|
||||
mockFetchForCustomTool(toolId)
|
||||
|
||||
let capturedTools: any[] = []
|
||||
Promise.all = vi.fn().mockImplementation((promises: Promise<any>[]) => {
|
||||
const result = originalPromiseAll.call(Promise, promises)
|
||||
result.then((tools: any[]) => {
|
||||
if (tools?.length) {
|
||||
capturedTools = tools.filter((t) => t !== null)
|
||||
}
|
||||
})
|
||||
return result
|
||||
})
|
||||
|
||||
const inputs = {
|
||||
model: 'gpt-4o',
|
||||
userPrompt: 'Format a report',
|
||||
apiKey: 'test-api-key',
|
||||
tools: [
|
||||
{
|
||||
type: 'custom-tool',
|
||||
customToolId: toolId,
|
||||
title: 'formatReport',
|
||||
schema: staleInlineSchema,
|
||||
code: staleInlineCode,
|
||||
usageControl: 'auto' as const,
|
||||
},
|
||||
],
|
||||
}
|
||||
|
||||
mockGetProviderFromModel.mockReturnValue('openai')
|
||||
|
||||
await handler.execute(mockContext, mockBlock, inputs)
|
||||
|
||||
expect(capturedTools.length).toBe(1)
|
||||
expect(typeof capturedTools[0].executeFunction).toBe('function')
|
||||
|
||||
await capturedTools[0].executeFunction({ title: 'Q1', format: 'pdf' })
|
||||
|
||||
expect(mockExecuteTool).toHaveBeenCalledWith(
|
||||
'function_execute',
|
||||
expect.objectContaining({
|
||||
code: dbCode,
|
||||
}),
|
||||
false,
|
||||
expect.any(Object)
|
||||
)
|
||||
})
|
||||
|
||||
it('should not fetch from DB when no customToolId is present', async () => {
|
||||
const inputs = {
|
||||
model: 'gpt-4o',
|
||||
userPrompt: 'Use the tool',
|
||||
apiKey: 'test-api-key',
|
||||
tools: [
|
||||
{
|
||||
type: 'custom-tool',
|
||||
title: 'formatReport',
|
||||
schema: staleInlineSchema,
|
||||
code: staleInlineCode,
|
||||
usageControl: 'auto' as const,
|
||||
},
|
||||
],
|
||||
}
|
||||
|
||||
mockGetProviderFromModel.mockReturnValue('openai')
|
||||
|
||||
await handler.execute(mockContext, mockBlock, inputs)
|
||||
|
||||
const customToolFetches = mockFetch.mock.calls.filter(
|
||||
(call: any[]) => typeof call[0] === 'string' && call[0].includes('/api/tools/custom')
|
||||
)
|
||||
expect(customToolFetches.length).toBe(0)
|
||||
|
||||
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||
const tools = providerCall[1].tools
|
||||
|
||||
expect(tools.length).toBe(1)
|
||||
expect(tools[0].name).toBe('formatReport')
|
||||
expect(tools[0].parameters.required).not.toContain('format')
|
||||
})
|
||||
})
|
||||
})
|
||||
})
|
||||
|
||||
@@ -272,16 +272,15 @@ export class AgentBlockHandler implements BlockHandler {
|
||||
let code = tool.code
|
||||
let title = tool.title
|
||||
|
||||
if (tool.customToolId) {
|
||||
if (tool.customToolId && !schema) {
|
||||
const resolved = await this.fetchCustomToolById(ctx, tool.customToolId)
|
||||
if (resolved) {
|
||||
schema = resolved.schema
|
||||
code = resolved.code
|
||||
title = resolved.title
|
||||
} else if (!schema) {
|
||||
if (!resolved) {
|
||||
logger.error(`Custom tool not found: ${tool.customToolId}`)
|
||||
return null
|
||||
}
|
||||
schema = resolved.schema
|
||||
code = resolved.code
|
||||
title = resolved.title
|
||||
}
|
||||
|
||||
if (!schema?.function) {
|
||||
|
||||
@@ -1,4 +1,3 @@
|
||||
import { setupGlobalFetchMock } from '@sim/testing'
|
||||
import { beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
||||
import { BlockType } from '@/executor/constants'
|
||||
import { WorkflowBlockHandler } from '@/executor/handlers/workflow/workflow-handler'
|
||||
@@ -10,7 +9,7 @@ vi.mock('@/lib/auth/internal', () => ({
|
||||
}))
|
||||
|
||||
// Mock fetch globally
|
||||
setupGlobalFetchMock()
|
||||
global.fetch = vi.fn()
|
||||
|
||||
describe('WorkflowBlockHandler', () => {
|
||||
let handler: WorkflowBlockHandler
|
||||
|
||||
@@ -4,7 +4,7 @@ import { API_ENDPOINTS } from '@/stores/constants'
|
||||
|
||||
const logger = createLogger('BYOKKeysQueries')
|
||||
|
||||
export type BYOKProviderId = 'openai' | 'anthropic' | 'google' | 'mistral' | 'serper'
|
||||
export type BYOKProviderId = 'openai' | 'anthropic' | 'google' | 'mistral'
|
||||
|
||||
export interface BYOKKey {
|
||||
id: string
|
||||
|
||||
@@ -423,7 +423,7 @@ interface GenerateVersionDescriptionVariables {
|
||||
|
||||
const VERSION_DESCRIPTION_SYSTEM_PROMPT = `You are writing deployment version descriptions for a workflow automation platform.
|
||||
|
||||
Write a brief, factual description (1-3 sentences, under 2000 characters) that states what changed between versions.
|
||||
Write a brief, factual description (1-3 sentences, under 400 characters) that states what changed between versions.
|
||||
|
||||
Guidelines:
|
||||
- Use the specific values provided (credential names, channel names, model names)
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
import { useCallback } from 'react'
|
||||
import { useCallback, useRef } from 'react'
|
||||
import { createLogger } from '@sim/logger'
|
||||
import type {
|
||||
BlockCompletedData,
|
||||
@@ -16,18 +16,6 @@ import type { SerializableExecutionState } from '@/executor/execution/types'
|
||||
|
||||
const logger = createLogger('useExecutionStream')
|
||||
|
||||
/**
|
||||
* Detects errors caused by the browser killing a fetch (page refresh, navigation, tab close).
|
||||
* These should be treated as clean disconnects, not execution errors.
|
||||
*/
|
||||
function isClientDisconnectError(error: any): boolean {
|
||||
if (error.name === 'AbortError') return true
|
||||
const msg = (error.message ?? '').toLowerCase()
|
||||
return (
|
||||
msg.includes('network error') || msg.includes('failed to fetch') || msg.includes('load failed')
|
||||
)
|
||||
}
|
||||
|
||||
/**
|
||||
* Processes SSE events from a response body and invokes appropriate callbacks.
|
||||
*/
|
||||
@@ -133,7 +121,6 @@ export interface ExecuteStreamOptions {
|
||||
parallels?: Record<string, any>
|
||||
}
|
||||
stopAfterBlockId?: string
|
||||
onExecutionId?: (executionId: string) => void
|
||||
callbacks?: ExecutionStreamCallbacks
|
||||
}
|
||||
|
||||
@@ -142,40 +129,30 @@ export interface ExecuteFromBlockOptions {
|
||||
startBlockId: string
|
||||
sourceSnapshot: SerializableExecutionState
|
||||
input?: any
|
||||
onExecutionId?: (executionId: string) => void
|
||||
callbacks?: ExecutionStreamCallbacks
|
||||
}
|
||||
|
||||
export interface ReconnectStreamOptions {
|
||||
workflowId: string
|
||||
executionId: string
|
||||
fromEventId?: number
|
||||
callbacks?: ExecutionStreamCallbacks
|
||||
}
|
||||
|
||||
/**
|
||||
* Module-level map shared across all hook instances.
|
||||
* Ensures ANY instance can cancel streams started by ANY other instance,
|
||||
* which is critical for SPA navigation where the original hook instance unmounts
|
||||
* but the SSE stream must be cancellable from the new instance.
|
||||
*/
|
||||
const sharedAbortControllers = new Map<string, AbortController>()
|
||||
|
||||
/**
|
||||
* Hook for executing workflows via server-side SSE streaming.
|
||||
* Supports concurrent executions via per-workflow AbortController maps.
|
||||
*/
|
||||
export function useExecutionStream() {
|
||||
const execute = useCallback(async (options: ExecuteStreamOptions) => {
|
||||
const { workflowId, callbacks = {}, onExecutionId, ...payload } = options
|
||||
const abortControllersRef = useRef<Map<string, AbortController>>(new Map())
|
||||
const currentExecutionsRef = useRef<Map<string, { workflowId: string; executionId: string }>>(
|
||||
new Map()
|
||||
)
|
||||
|
||||
const existing = sharedAbortControllers.get(workflowId)
|
||||
const execute = useCallback(async (options: ExecuteStreamOptions) => {
|
||||
const { workflowId, callbacks = {}, ...payload } = options
|
||||
|
||||
const existing = abortControllersRef.current.get(workflowId)
|
||||
if (existing) {
|
||||
existing.abort()
|
||||
}
|
||||
|
||||
const abortController = new AbortController()
|
||||
sharedAbortControllers.set(workflowId, abortController)
|
||||
abortControllersRef.current.set(workflowId, abortController)
|
||||
currentExecutionsRef.current.delete(workflowId)
|
||||
|
||||
try {
|
||||
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
||||
@@ -200,48 +177,42 @@ export function useExecutionStream() {
|
||||
throw new Error('No response body')
|
||||
}
|
||||
|
||||
const serverExecutionId = response.headers.get('X-Execution-Id')
|
||||
if (serverExecutionId) {
|
||||
onExecutionId?.(serverExecutionId)
|
||||
const executionId = response.headers.get('X-Execution-Id')
|
||||
if (executionId) {
|
||||
currentExecutionsRef.current.set(workflowId, { workflowId, executionId })
|
||||
}
|
||||
|
||||
const reader = response.body.getReader()
|
||||
await processSSEStream(reader, callbacks, 'Execution')
|
||||
} catch (error: any) {
|
||||
if (isClientDisconnectError(error)) {
|
||||
logger.info('Execution stream disconnected (page unload or abort)')
|
||||
return
|
||||
if (error.name === 'AbortError') {
|
||||
logger.info('Execution stream cancelled')
|
||||
callbacks.onExecutionCancelled?.({ duration: 0 })
|
||||
} else {
|
||||
logger.error('Execution stream error:', error)
|
||||
callbacks.onExecutionError?.({
|
||||
error: error.message || 'Unknown error',
|
||||
duration: 0,
|
||||
})
|
||||
}
|
||||
logger.error('Execution stream error:', error)
|
||||
callbacks.onExecutionError?.({
|
||||
error: error.message || 'Unknown error',
|
||||
duration: 0,
|
||||
})
|
||||
throw error
|
||||
} finally {
|
||||
if (sharedAbortControllers.get(workflowId) === abortController) {
|
||||
sharedAbortControllers.delete(workflowId)
|
||||
}
|
||||
abortControllersRef.current.delete(workflowId)
|
||||
currentExecutionsRef.current.delete(workflowId)
|
||||
}
|
||||
}, [])
|
||||
|
||||
const executeFromBlock = useCallback(async (options: ExecuteFromBlockOptions) => {
|
||||
const {
|
||||
workflowId,
|
||||
startBlockId,
|
||||
sourceSnapshot,
|
||||
input,
|
||||
onExecutionId,
|
||||
callbacks = {},
|
||||
} = options
|
||||
const { workflowId, startBlockId, sourceSnapshot, input, callbacks = {} } = options
|
||||
|
||||
const existing = sharedAbortControllers.get(workflowId)
|
||||
const existing = abortControllersRef.current.get(workflowId)
|
||||
if (existing) {
|
||||
existing.abort()
|
||||
}
|
||||
|
||||
const abortController = new AbortController()
|
||||
sharedAbortControllers.set(workflowId, abortController)
|
||||
abortControllersRef.current.set(workflowId, abortController)
|
||||
currentExecutionsRef.current.delete(workflowId)
|
||||
|
||||
try {
|
||||
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
||||
@@ -275,80 +246,64 @@ export function useExecutionStream() {
|
||||
throw new Error('No response body')
|
||||
}
|
||||
|
||||
const serverExecutionId = response.headers.get('X-Execution-Id')
|
||||
if (serverExecutionId) {
|
||||
onExecutionId?.(serverExecutionId)
|
||||
const executionId = response.headers.get('X-Execution-Id')
|
||||
if (executionId) {
|
||||
currentExecutionsRef.current.set(workflowId, { workflowId, executionId })
|
||||
}
|
||||
|
||||
const reader = response.body.getReader()
|
||||
await processSSEStream(reader, callbacks, 'Run-from-block')
|
||||
} catch (error: any) {
|
||||
if (isClientDisconnectError(error)) {
|
||||
logger.info('Run-from-block stream disconnected (page unload or abort)')
|
||||
return
|
||||
if (error.name === 'AbortError') {
|
||||
logger.info('Run-from-block execution cancelled')
|
||||
callbacks.onExecutionCancelled?.({ duration: 0 })
|
||||
} else {
|
||||
logger.error('Run-from-block execution error:', error)
|
||||
callbacks.onExecutionError?.({
|
||||
error: error.message || 'Unknown error',
|
||||
duration: 0,
|
||||
})
|
||||
}
|
||||
logger.error('Run-from-block execution error:', error)
|
||||
callbacks.onExecutionError?.({
|
||||
error: error.message || 'Unknown error',
|
||||
duration: 0,
|
||||
})
|
||||
throw error
|
||||
} finally {
|
||||
if (sharedAbortControllers.get(workflowId) === abortController) {
|
||||
sharedAbortControllers.delete(workflowId)
|
||||
}
|
||||
}
|
||||
}, [])
|
||||
|
||||
const reconnect = useCallback(async (options: ReconnectStreamOptions) => {
|
||||
const { workflowId, executionId, fromEventId = 0, callbacks = {} } = options
|
||||
|
||||
const existing = sharedAbortControllers.get(workflowId)
|
||||
if (existing) {
|
||||
existing.abort()
|
||||
}
|
||||
|
||||
const abortController = new AbortController()
|
||||
sharedAbortControllers.set(workflowId, abortController)
|
||||
try {
|
||||
const response = await fetch(
|
||||
`/api/workflows/${workflowId}/executions/${executionId}/stream?from=${fromEventId}`,
|
||||
{ signal: abortController.signal }
|
||||
)
|
||||
if (!response.ok) throw new Error(`Reconnect failed (${response.status})`)
|
||||
if (!response.body) throw new Error('No response body')
|
||||
|
||||
await processSSEStream(response.body.getReader(), callbacks, 'Reconnect')
|
||||
} catch (error: any) {
|
||||
if (isClientDisconnectError(error)) return
|
||||
logger.error('Reconnection stream error:', error)
|
||||
throw error
|
||||
} finally {
|
||||
if (sharedAbortControllers.get(workflowId) === abortController) {
|
||||
sharedAbortControllers.delete(workflowId)
|
||||
}
|
||||
abortControllersRef.current.delete(workflowId)
|
||||
currentExecutionsRef.current.delete(workflowId)
|
||||
}
|
||||
}, [])
|
||||
|
||||
const cancel = useCallback((workflowId?: string) => {
|
||||
if (workflowId) {
|
||||
const controller = sharedAbortControllers.get(workflowId)
|
||||
const execution = currentExecutionsRef.current.get(workflowId)
|
||||
if (execution) {
|
||||
fetch(`/api/workflows/${execution.workflowId}/executions/${execution.executionId}/cancel`, {
|
||||
method: 'POST',
|
||||
}).catch(() => {})
|
||||
}
|
||||
|
||||
const controller = abortControllersRef.current.get(workflowId)
|
||||
if (controller) {
|
||||
controller.abort()
|
||||
sharedAbortControllers.delete(workflowId)
|
||||
abortControllersRef.current.delete(workflowId)
|
||||
}
|
||||
currentExecutionsRef.current.delete(workflowId)
|
||||
} else {
|
||||
for (const [, controller] of sharedAbortControllers) {
|
||||
for (const [, execution] of currentExecutionsRef.current) {
|
||||
fetch(`/api/workflows/${execution.workflowId}/executions/${execution.executionId}/cancel`, {
|
||||
method: 'POST',
|
||||
}).catch(() => {})
|
||||
}
|
||||
|
||||
for (const [, controller] of abortControllersRef.current) {
|
||||
controller.abort()
|
||||
}
|
||||
sharedAbortControllers.clear()
|
||||
abortControllersRef.current.clear()
|
||||
currentExecutionsRef.current.clear()
|
||||
}
|
||||
}, [])
|
||||
|
||||
return {
|
||||
execute,
|
||||
executeFromBlock,
|
||||
reconnect,
|
||||
cancel,
|
||||
}
|
||||
}
|
||||
|
||||
@@ -10,7 +10,7 @@ import { useProvidersStore } from '@/stores/providers/store'
|
||||
|
||||
const logger = createLogger('BYOKKeys')
|
||||
|
||||
export type BYOKProviderId = 'openai' | 'anthropic' | 'google' | 'mistral' | 'serper'
|
||||
export type BYOKProviderId = 'openai' | 'anthropic' | 'google' | 'mistral'
|
||||
|
||||
export interface BYOKKeyResult {
|
||||
apiKey: string
|
||||
|
||||
@@ -1,246 +0,0 @@
|
||||
import { createLogger } from '@sim/logger'
|
||||
import { getRedisClient } from '@/lib/core/config/redis'
|
||||
import type { ExecutionEvent } from '@/lib/workflows/executor/execution-events'
|
||||
|
||||
const logger = createLogger('ExecutionEventBuffer')
|
||||
|
||||
const REDIS_PREFIX = 'execution:stream:'
|
||||
const TTL_SECONDS = 60 * 60 // 1 hour
|
||||
const EVENT_LIMIT = 1000
|
||||
const RESERVE_BATCH = 100
|
||||
const FLUSH_INTERVAL_MS = 15
|
||||
const FLUSH_MAX_BATCH = 200
|
||||
|
||||
function getEventsKey(executionId: string) {
|
||||
return `${REDIS_PREFIX}${executionId}:events`
|
||||
}
|
||||
|
||||
function getSeqKey(executionId: string) {
|
||||
return `${REDIS_PREFIX}${executionId}:seq`
|
||||
}
|
||||
|
||||
function getMetaKey(executionId: string) {
|
||||
return `${REDIS_PREFIX}${executionId}:meta`
|
||||
}
|
||||
|
||||
export type ExecutionStreamStatus = 'active' | 'complete' | 'error' | 'cancelled'
|
||||
|
||||
export interface ExecutionStreamMeta {
|
||||
status: ExecutionStreamStatus
|
||||
userId?: string
|
||||
workflowId?: string
|
||||
updatedAt?: string
|
||||
}
|
||||
|
||||
export interface ExecutionEventEntry {
|
||||
eventId: number
|
||||
executionId: string
|
||||
event: ExecutionEvent
|
||||
}
|
||||
|
||||
export interface ExecutionEventWriter {
|
||||
write: (event: ExecutionEvent) => Promise<ExecutionEventEntry>
|
||||
flush: () => Promise<void>
|
||||
close: () => Promise<void>
|
||||
}
|
||||
|
||||
export async function setExecutionMeta(
|
||||
executionId: string,
|
||||
meta: Partial<ExecutionStreamMeta>
|
||||
): Promise<void> {
|
||||
const redis = getRedisClient()
|
||||
if (!redis) {
|
||||
logger.warn('setExecutionMeta: Redis client unavailable', { executionId })
|
||||
return
|
||||
}
|
||||
try {
|
||||
const key = getMetaKey(executionId)
|
||||
const payload: Record<string, string> = {
|
||||
updatedAt: new Date().toISOString(),
|
||||
}
|
||||
if (meta.status) payload.status = meta.status
|
||||
if (meta.userId) payload.userId = meta.userId
|
||||
if (meta.workflowId) payload.workflowId = meta.workflowId
|
||||
await redis.hset(key, payload)
|
||||
await redis.expire(key, TTL_SECONDS)
|
||||
} catch (error) {
|
||||
logger.warn('Failed to update execution meta', {
|
||||
executionId,
|
||||
error: error instanceof Error ? error.message : String(error),
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
export async function getExecutionMeta(executionId: string): Promise<ExecutionStreamMeta | null> {
|
||||
const redis = getRedisClient()
|
||||
if (!redis) {
|
||||
logger.warn('getExecutionMeta: Redis client unavailable', { executionId })
|
||||
return null
|
||||
}
|
||||
try {
|
||||
const key = getMetaKey(executionId)
|
||||
const meta = await redis.hgetall(key)
|
||||
if (!meta || Object.keys(meta).length === 0) return null
|
||||
return meta as unknown as ExecutionStreamMeta
|
||||
} catch (error) {
|
||||
logger.warn('Failed to read execution meta', {
|
||||
executionId,
|
||||
error: error instanceof Error ? error.message : String(error),
|
||||
})
|
||||
return null
|
||||
}
|
||||
}
|
||||
|
||||
export async function readExecutionEvents(
|
||||
executionId: string,
|
||||
afterEventId: number
|
||||
): Promise<ExecutionEventEntry[]> {
|
||||
const redis = getRedisClient()
|
||||
if (!redis) return []
|
||||
try {
|
||||
const raw = await redis.zrangebyscore(getEventsKey(executionId), afterEventId + 1, '+inf')
|
||||
return raw
|
||||
.map((entry) => {
|
||||
try {
|
||||
return JSON.parse(entry) as ExecutionEventEntry
|
||||
} catch {
|
||||
return null
|
||||
}
|
||||
})
|
||||
.filter((entry): entry is ExecutionEventEntry => Boolean(entry))
|
||||
} catch (error) {
|
||||
logger.warn('Failed to read execution events', {
|
||||
executionId,
|
||||
error: error instanceof Error ? error.message : String(error),
|
||||
})
|
||||
return []
|
||||
}
|
||||
}
|
||||
|
||||
export function createExecutionEventWriter(executionId: string): ExecutionEventWriter {
|
||||
const redis = getRedisClient()
|
||||
if (!redis) {
|
||||
logger.warn(
|
||||
'createExecutionEventWriter: Redis client unavailable, events will not be buffered',
|
||||
{
|
||||
executionId,
|
||||
}
|
||||
)
|
||||
return {
|
||||
write: async (event) => ({ eventId: 0, executionId, event }),
|
||||
flush: async () => {},
|
||||
close: async () => {},
|
||||
}
|
||||
}
|
||||
|
||||
let pending: ExecutionEventEntry[] = []
|
||||
let nextEventId = 0
|
||||
let maxReservedId = 0
|
||||
let flushTimer: ReturnType<typeof setTimeout> | null = null
|
||||
|
||||
const scheduleFlush = () => {
|
||||
if (flushTimer) return
|
||||
flushTimer = setTimeout(() => {
|
||||
flushTimer = null
|
||||
void flush()
|
||||
}, FLUSH_INTERVAL_MS)
|
||||
}
|
||||
|
||||
const reserveIds = async (minCount: number) => {
|
||||
const reserveCount = Math.max(RESERVE_BATCH, minCount)
|
||||
const newMax = await redis.incrby(getSeqKey(executionId), reserveCount)
|
||||
const startId = newMax - reserveCount + 1
|
||||
if (nextEventId === 0 || nextEventId > maxReservedId) {
|
||||
nextEventId = startId
|
||||
maxReservedId = newMax
|
||||
}
|
||||
}
|
||||
|
||||
let flushPromise: Promise<void> | null = null
|
||||
let closed = false
|
||||
const inflightWrites = new Set<Promise<ExecutionEventEntry>>()
|
||||
|
||||
const doFlush = async () => {
|
||||
if (pending.length === 0) return
|
||||
const batch = pending
|
||||
pending = []
|
||||
try {
|
||||
const key = getEventsKey(executionId)
|
||||
const zaddArgs: (string | number)[] = []
|
||||
for (const entry of batch) {
|
||||
zaddArgs.push(entry.eventId, JSON.stringify(entry))
|
||||
}
|
||||
const pipeline = redis.pipeline()
|
||||
pipeline.zadd(key, ...zaddArgs)
|
||||
pipeline.expire(key, TTL_SECONDS)
|
||||
pipeline.expire(getSeqKey(executionId), TTL_SECONDS)
|
||||
pipeline.zremrangebyrank(key, 0, -EVENT_LIMIT - 1)
|
||||
await pipeline.exec()
|
||||
} catch (error) {
|
||||
logger.warn('Failed to flush execution events', {
|
||||
executionId,
|
||||
batchSize: batch.length,
|
||||
error: error instanceof Error ? error.message : String(error),
|
||||
stack: error instanceof Error ? error.stack : undefined,
|
||||
})
|
||||
pending = batch.concat(pending)
|
||||
}
|
||||
}
|
||||
|
||||
const flush = async () => {
|
||||
if (flushPromise) {
|
||||
await flushPromise
|
||||
return
|
||||
}
|
||||
flushPromise = doFlush()
|
||||
try {
|
||||
await flushPromise
|
||||
} finally {
|
||||
flushPromise = null
|
||||
if (pending.length > 0) scheduleFlush()
|
||||
}
|
||||
}
|
||||
|
||||
const writeCore = async (event: ExecutionEvent): Promise<ExecutionEventEntry> => {
|
||||
if (closed) return { eventId: 0, executionId, event }
|
||||
if (nextEventId === 0 || nextEventId > maxReservedId) {
|
||||
await reserveIds(1)
|
||||
}
|
||||
const eventId = nextEventId++
|
||||
const entry: ExecutionEventEntry = { eventId, executionId, event }
|
||||
pending.push(entry)
|
||||
if (pending.length >= FLUSH_MAX_BATCH) {
|
||||
await flush()
|
||||
} else {
|
||||
scheduleFlush()
|
||||
}
|
||||
return entry
|
||||
}
|
||||
|
||||
const write = (event: ExecutionEvent): Promise<ExecutionEventEntry> => {
|
||||
const p = writeCore(event)
|
||||
inflightWrites.add(p)
|
||||
const remove = () => inflightWrites.delete(p)
|
||||
p.then(remove, remove)
|
||||
return p
|
||||
}
|
||||
|
||||
const close = async () => {
|
||||
closed = true
|
||||
if (flushTimer) {
|
||||
clearTimeout(flushTimer)
|
||||
flushTimer = null
|
||||
}
|
||||
if (inflightWrites.size > 0) {
|
||||
await Promise.allSettled(inflightWrites)
|
||||
}
|
||||
if (flushPromise) {
|
||||
await flushPromise
|
||||
}
|
||||
if (pending.length > 0) {
|
||||
await doFlush()
|
||||
}
|
||||
}
|
||||
|
||||
return { write, flush, close }
|
||||
}
|
||||
@@ -1,4 +1,4 @@
|
||||
import { createEnvMock, loggerMock } from '@sim/testing'
|
||||
import { createEnvMock, createMockLogger } from '@sim/testing'
|
||||
import { beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
||||
|
||||
/**
|
||||
@@ -10,6 +10,10 @@ import { beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
||||
* mock functions can intercept.
|
||||
*/
|
||||
|
||||
const loggerMock = vi.hoisted(() => ({
|
||||
createLogger: () => createMockLogger(),
|
||||
}))
|
||||
|
||||
const mockSend = vi.fn()
|
||||
const mockBatchSend = vi.fn()
|
||||
const mockAzureBeginSend = vi.fn()
|
||||
|
||||
@@ -1,8 +1,20 @@
|
||||
import { createEnvMock, databaseMock, loggerMock } from '@sim/testing'
|
||||
import { createEnvMock, createMockLogger } from '@sim/testing'
|
||||
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
import type { EmailType } from '@/lib/messaging/email/mailer'
|
||||
|
||||
vi.mock('@sim/db', () => databaseMock)
|
||||
const loggerMock = vi.hoisted(() => ({
|
||||
createLogger: () => createMockLogger(),
|
||||
}))
|
||||
|
||||
const mockDb = vi.hoisted(() => ({
|
||||
select: vi.fn(),
|
||||
insert: vi.fn(),
|
||||
update: vi.fn(),
|
||||
}))
|
||||
|
||||
vi.mock('@sim/db', () => ({
|
||||
db: mockDb,
|
||||
}))
|
||||
|
||||
vi.mock('@sim/db/schema', () => ({
|
||||
user: { id: 'id', email: 'email' },
|
||||
@@ -18,8 +30,6 @@ vi.mock('drizzle-orm', () => ({
|
||||
eq: vi.fn((a, b) => ({ type: 'eq', left: a, right: b })),
|
||||
}))
|
||||
|
||||
const mockDb = databaseMock.db as Record<string, ReturnType<typeof vi.fn>>
|
||||
|
||||
vi.mock('@/lib/core/config/env', () => createEnvMock({ BETTER_AUTH_SECRET: 'test-secret-key' }))
|
||||
|
||||
vi.mock('@sim/logger', () => loggerMock)
|
||||
|
||||
@@ -2364,261 +2364,6 @@ describe('hasWorkflowChanged', () => {
|
||||
})
|
||||
})
|
||||
|
||||
describe('Trigger Config Normalization (False Positive Prevention)', () => {
|
||||
it.concurrent(
|
||||
'should not detect change when deployed has null fields but current has values from triggerConfig',
|
||||
() => {
|
||||
// Core scenario: deployed state has null individual fields, current state has
|
||||
// values populated from triggerConfig at runtime by populateTriggerFieldsFromConfig
|
||||
const deployedState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
botToken: { id: 'botToken', type: 'short-input', value: null },
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||
},
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
const currentState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
||||
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||
},
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||
}
|
||||
)
|
||||
|
||||
it.concurrent(
|
||||
'should detect change when user edits a trigger field to a different value',
|
||||
() => {
|
||||
const deployedState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'old-secret' },
|
||||
},
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
const currentState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'new-secret' },
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'old-secret' },
|
||||
},
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(true)
|
||||
}
|
||||
)
|
||||
|
||||
it.concurrent('should not detect change when both sides have no triggerConfig', () => {
|
||||
const deployedState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
const currentState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||
})
|
||||
|
||||
it.concurrent(
|
||||
'should not detect change when deployed has empty fields and triggerConfig populates them',
|
||||
() => {
|
||||
// Empty string is also treated as "empty" by normalizeTriggerConfigValues
|
||||
const deployedState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: '' },
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123' },
|
||||
},
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
const currentState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123' },
|
||||
},
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||
}
|
||||
)
|
||||
|
||||
it.concurrent('should not detect change when triggerId differs', () => {
|
||||
const deployedState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
model: { value: 'gpt-4' },
|
||||
triggerId: { value: null },
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
const currentState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
model: { value: 'gpt-4' },
|
||||
triggerId: { value: 'slack_webhook' },
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||
})
|
||||
|
||||
it.concurrent(
|
||||
'should not detect change for namespaced system subBlock IDs like samplePayload_slack_webhook',
|
||||
() => {
|
||||
const deployedState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
model: { value: 'gpt-4' },
|
||||
samplePayload_slack_webhook: { value: 'old payload' },
|
||||
triggerInstructions_slack_webhook: { value: 'old instructions' },
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
const currentState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
model: { value: 'gpt-4' },
|
||||
samplePayload_slack_webhook: { value: 'new payload' },
|
||||
triggerInstructions_slack_webhook: { value: 'new instructions' },
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||
}
|
||||
)
|
||||
|
||||
it.concurrent(
|
||||
'should handle mixed scenario: some fields from triggerConfig, some user-edited',
|
||||
() => {
|
||||
const deployedState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
botToken: { id: 'botToken', type: 'short-input', value: null },
|
||||
includeFiles: { id: 'includeFiles', type: 'switch', value: false },
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||
},
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
const currentState = createWorkflowState({
|
||||
blocks: {
|
||||
block1: createBlock('block1', {
|
||||
type: 'starter',
|
||||
subBlocks: {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
||||
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
||||
includeFiles: { id: 'includeFiles', type: 'switch', value: true },
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||
},
|
||||
},
|
||||
}),
|
||||
},
|
||||
})
|
||||
|
||||
// includeFiles changed from false to true — this IS a real change
|
||||
expect(hasWorkflowChanged(currentState, deployedState)).toBe(true)
|
||||
}
|
||||
)
|
||||
})
|
||||
|
||||
describe('Trigger Runtime Metadata (Should Not Trigger Change)', () => {
|
||||
it.concurrent('should not detect change when webhookId differs', () => {
|
||||
const deployedState = createWorkflowState({
|
||||
|
||||
@@ -9,7 +9,6 @@ import {
|
||||
normalizeLoop,
|
||||
normalizeParallel,
|
||||
normalizeSubBlockValue,
|
||||
normalizeTriggerConfigValues,
|
||||
normalizeValue,
|
||||
normalizeVariables,
|
||||
sanitizeVariable,
|
||||
@@ -173,18 +172,14 @@ export function generateWorkflowDiffSummary(
|
||||
}
|
||||
}
|
||||
|
||||
// Normalize trigger config values for both states before comparison
|
||||
const normalizedCurrentSubs = normalizeTriggerConfigValues(currentSubBlocks)
|
||||
const normalizedPreviousSubs = normalizeTriggerConfigValues(previousSubBlocks)
|
||||
|
||||
// Compare subBlocks using shared helper for filtering (single source of truth)
|
||||
const allSubBlockIds = filterSubBlockIds([
|
||||
...new Set([...Object.keys(normalizedCurrentSubs), ...Object.keys(normalizedPreviousSubs)]),
|
||||
...new Set([...Object.keys(currentSubBlocks), ...Object.keys(previousSubBlocks)]),
|
||||
])
|
||||
|
||||
for (const subId of allSubBlockIds) {
|
||||
const currentSub = normalizedCurrentSubs[subId] as Record<string, unknown> | undefined
|
||||
const previousSub = normalizedPreviousSubs[subId] as Record<string, unknown> | undefined
|
||||
const currentSub = currentSubBlocks[subId] as Record<string, unknown> | undefined
|
||||
const previousSub = previousSubBlocks[subId] as Record<string, unknown> | undefined
|
||||
|
||||
if (!currentSub || !previousSub) {
|
||||
changes.push({
|
||||
|
||||
@@ -4,12 +4,10 @@
|
||||
import { describe, expect, it } from 'vitest'
|
||||
import type { Loop, Parallel } from '@/stores/workflows/workflow/types'
|
||||
import {
|
||||
filterSubBlockIds,
|
||||
normalizedStringify,
|
||||
normalizeEdge,
|
||||
normalizeLoop,
|
||||
normalizeParallel,
|
||||
normalizeTriggerConfigValues,
|
||||
normalizeValue,
|
||||
sanitizeInputFormat,
|
||||
sanitizeTools,
|
||||
@@ -586,214 +584,4 @@ describe('Workflow Normalization Utilities', () => {
|
||||
expect(result2).toBe(result3)
|
||||
})
|
||||
})
|
||||
|
||||
describe('filterSubBlockIds', () => {
|
||||
it.concurrent('should exclude exact SYSTEM_SUBBLOCK_IDS', () => {
|
||||
const ids = ['signingSecret', 'samplePayload', 'triggerInstructions', 'botToken']
|
||||
const result = filterSubBlockIds(ids)
|
||||
expect(result).toEqual(['botToken', 'signingSecret'])
|
||||
})
|
||||
|
||||
it.concurrent('should exclude namespaced SYSTEM_SUBBLOCK_IDS (prefix matching)', () => {
|
||||
const ids = [
|
||||
'signingSecret',
|
||||
'samplePayload_slack_webhook',
|
||||
'triggerInstructions_slack_webhook',
|
||||
'webhookUrlDisplay_slack_webhook',
|
||||
'botToken',
|
||||
]
|
||||
const result = filterSubBlockIds(ids)
|
||||
expect(result).toEqual(['botToken', 'signingSecret'])
|
||||
})
|
||||
|
||||
it.concurrent('should exclude exact TRIGGER_RUNTIME_SUBBLOCK_IDS', () => {
|
||||
const ids = ['webhookId', 'triggerPath', 'triggerConfig', 'triggerId', 'signingSecret']
|
||||
const result = filterSubBlockIds(ids)
|
||||
expect(result).toEqual(['signingSecret'])
|
||||
})
|
||||
|
||||
it.concurrent('should not exclude IDs that merely contain a system ID substring', () => {
|
||||
const ids = ['mySamplePayload', 'notSamplePayload']
|
||||
const result = filterSubBlockIds(ids)
|
||||
expect(result).toEqual(['mySamplePayload', 'notSamplePayload'])
|
||||
})
|
||||
|
||||
it.concurrent('should return sorted results', () => {
|
||||
const ids = ['zebra', 'alpha', 'middle']
|
||||
const result = filterSubBlockIds(ids)
|
||||
expect(result).toEqual(['alpha', 'middle', 'zebra'])
|
||||
})
|
||||
|
||||
it.concurrent('should handle empty array', () => {
|
||||
expect(filterSubBlockIds([])).toEqual([])
|
||||
})
|
||||
|
||||
it.concurrent('should handle all IDs being excluded', () => {
|
||||
const ids = ['webhookId', 'triggerPath', 'samplePayload', 'triggerConfig']
|
||||
const result = filterSubBlockIds(ids)
|
||||
expect(result).toEqual([])
|
||||
})
|
||||
|
||||
it.concurrent('should exclude setupScript and scheduleInfo namespaced variants', () => {
|
||||
const ids = ['setupScript_google_sheets_row', 'scheduleInfo_cron_trigger', 'realField']
|
||||
const result = filterSubBlockIds(ids)
|
||||
expect(result).toEqual(['realField'])
|
||||
})
|
||||
|
||||
it.concurrent('should exclude triggerCredentials namespaced variants', () => {
|
||||
const ids = ['triggerCredentials_slack_webhook', 'signingSecret']
|
||||
const result = filterSubBlockIds(ids)
|
||||
expect(result).toEqual(['signingSecret'])
|
||||
})
|
||||
})
|
||||
|
||||
describe('normalizeTriggerConfigValues', () => {
|
||||
it.concurrent('should return subBlocks unchanged when no triggerConfig exists', () => {
|
||||
const subBlocks = {
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
||||
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect(result).toEqual(subBlocks)
|
||||
})
|
||||
|
||||
it.concurrent('should return subBlocks unchanged when triggerConfig value is null', () => {
|
||||
const subBlocks = {
|
||||
triggerConfig: { id: 'triggerConfig', type: 'short-input', value: null },
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect(result).toEqual(subBlocks)
|
||||
})
|
||||
|
||||
it.concurrent(
|
||||
'should return subBlocks unchanged when triggerConfig value is not an object',
|
||||
() => {
|
||||
const subBlocks = {
|
||||
triggerConfig: { id: 'triggerConfig', type: 'short-input', value: 'string-value' },
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect(result).toEqual(subBlocks)
|
||||
}
|
||||
)
|
||||
|
||||
it.concurrent('should populate null individual fields from triggerConfig', () => {
|
||||
const subBlocks = {
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||
},
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
botToken: { id: 'botToken', type: 'short-input', value: null },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
||||
expect((result.botToken as Record<string, unknown>).value).toBe('token456')
|
||||
})
|
||||
|
||||
it.concurrent('should populate undefined individual fields from triggerConfig', () => {
|
||||
const subBlocks = {
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123' },
|
||||
},
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: undefined },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
||||
})
|
||||
|
||||
it.concurrent('should populate empty string individual fields from triggerConfig', () => {
|
||||
const subBlocks = {
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123' },
|
||||
},
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: '' },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
||||
})
|
||||
|
||||
it.concurrent('should NOT overwrite existing non-empty individual field values', () => {
|
||||
const subBlocks = {
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'old-secret' },
|
||||
},
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'user-edited-secret' },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect((result.signingSecret as Record<string, unknown>).value).toBe('user-edited-secret')
|
||||
})
|
||||
|
||||
it.concurrent('should skip triggerConfig fields that are null/undefined', () => {
|
||||
const subBlocks = {
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: null, botToken: undefined },
|
||||
},
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
botToken: { id: 'botToken', type: 'short-input', value: null },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect((result.signingSecret as Record<string, unknown>).value).toBe(null)
|
||||
expect((result.botToken as Record<string, unknown>).value).toBe(null)
|
||||
})
|
||||
|
||||
it.concurrent('should skip fields from triggerConfig that have no matching subBlock', () => {
|
||||
const subBlocks = {
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { nonExistentField: 'value123' },
|
||||
},
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
expect(result.nonExistentField).toBeUndefined()
|
||||
expect((result.signingSecret as Record<string, unknown>).value).toBe(null)
|
||||
})
|
||||
|
||||
it.concurrent('should not mutate the original subBlocks object', () => {
|
||||
const original = {
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123' },
|
||||
},
|
||||
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||
}
|
||||
normalizeTriggerConfigValues(original)
|
||||
expect((original.signingSecret as Record<string, unknown>).value).toBe(null)
|
||||
})
|
||||
|
||||
it.concurrent('should preserve other subBlock properties when populating value', () => {
|
||||
const subBlocks = {
|
||||
triggerConfig: {
|
||||
id: 'triggerConfig',
|
||||
type: 'short-input',
|
||||
value: { signingSecret: 'secret123' },
|
||||
},
|
||||
signingSecret: {
|
||||
id: 'signingSecret',
|
||||
type: 'short-input',
|
||||
value: null,
|
||||
placeholder: 'Enter signing secret',
|
||||
},
|
||||
}
|
||||
const result = normalizeTriggerConfigValues(subBlocks)
|
||||
const normalized = result.signingSecret as Record<string, unknown>
|
||||
expect(normalized.value).toBe('secret123')
|
||||
expect(normalized.id).toBe('signingSecret')
|
||||
expect(normalized.type).toBe('short-input')
|
||||
expect(normalized.placeholder).toBe('Enter signing secret')
|
||||
})
|
||||
})
|
||||
})
|
||||
|
||||
@@ -418,48 +418,10 @@ export function extractBlockFieldsForComparison(block: BlockState): ExtractedBlo
|
||||
*/
|
||||
export function filterSubBlockIds(subBlockIds: string[]): string[] {
|
||||
return subBlockIds
|
||||
.filter((id) => {
|
||||
if (TRIGGER_RUNTIME_SUBBLOCK_IDS.includes(id)) return false
|
||||
if (SYSTEM_SUBBLOCK_IDS.some((sysId) => id === sysId || id.startsWith(`${sysId}_`)))
|
||||
return false
|
||||
return true
|
||||
})
|
||||
.filter((id) => !SYSTEM_SUBBLOCK_IDS.includes(id) && !TRIGGER_RUNTIME_SUBBLOCK_IDS.includes(id))
|
||||
.sort()
|
||||
}
|
||||
|
||||
/**
|
||||
* Normalizes trigger block subBlocks by populating null/empty individual fields
|
||||
* from the triggerConfig aggregate subBlock. This compensates for the runtime
|
||||
* population done by populateTriggerFieldsFromConfig, ensuring consistent
|
||||
* comparison between client state (with populated values) and deployed state
|
||||
* (with null values from DB).
|
||||
*/
|
||||
export function normalizeTriggerConfigValues(
|
||||
subBlocks: Record<string, unknown>
|
||||
): Record<string, unknown> {
|
||||
const triggerConfigSub = subBlocks.triggerConfig as Record<string, unknown> | undefined
|
||||
const triggerConfigValue = triggerConfigSub?.value
|
||||
if (!triggerConfigValue || typeof triggerConfigValue !== 'object') {
|
||||
return subBlocks
|
||||
}
|
||||
|
||||
const result = { ...subBlocks }
|
||||
for (const [fieldId, configValue] of Object.entries(
|
||||
triggerConfigValue as Record<string, unknown>
|
||||
)) {
|
||||
if (configValue === null || configValue === undefined) continue
|
||||
const existingSub = result[fieldId] as Record<string, unknown> | undefined
|
||||
if (
|
||||
existingSub &&
|
||||
(existingSub.value === null || existingSub.value === undefined || existingSub.value === '')
|
||||
) {
|
||||
result[fieldId] = { ...existingSub, value: configValue }
|
||||
}
|
||||
}
|
||||
|
||||
return result
|
||||
}
|
||||
|
||||
/**
|
||||
* Normalizes a subBlock value with sanitization for specific subBlock types.
|
||||
* Sanitizes: tools (removes isExpanded), inputFormat (removes collapsed)
|
||||
|
||||
@@ -1,11 +1,18 @@
|
||||
/**
|
||||
* @vitest-environment node
|
||||
*/
|
||||
import { loggerMock } from '@sim/testing'
|
||||
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
import type { BlockState, WorkflowState } from '@/stores/workflows/workflow/types'
|
||||
|
||||
vi.mock('@sim/logger', () => loggerMock)
|
||||
// Mock all external dependencies before imports
|
||||
vi.mock('@sim/logger', () => ({
|
||||
createLogger: () => ({
|
||||
info: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
error: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
}),
|
||||
}))
|
||||
|
||||
vi.mock('@/stores/workflows/workflow/store', () => ({
|
||||
useWorkflowStore: {
|
||||
|
||||
@@ -1,5 +1,4 @@
|
||||
import { getEnv, isTruthy } from '@/lib/core/config/env'
|
||||
import { isHosted } from '@/lib/core/config/feature-flags'
|
||||
import type { SubBlockConfig } from '@/blocks/types'
|
||||
|
||||
export type CanonicalMode = 'basic' | 'advanced'
|
||||
@@ -271,12 +270,3 @@ export function isSubBlockFeatureEnabled(subBlock: SubBlockConfig): boolean {
|
||||
if (!subBlock.requiresFeature) return true
|
||||
return isTruthy(getEnv(subBlock.requiresFeature))
|
||||
}
|
||||
|
||||
/**
|
||||
* Check if a subblock should be hidden because we're running on hosted Sim.
|
||||
* Used for tool API key fields that should be hidden when Sim provides hosted keys.
|
||||
*/
|
||||
export function isSubBlockHiddenByHostedKey(subBlock: SubBlockConfig): boolean {
|
||||
if (!subBlock.hideWhenHosted) return false
|
||||
return isHosted
|
||||
}
|
||||
|
||||
@@ -14,15 +14,22 @@ import {
|
||||
databaseMock,
|
||||
expectWorkflowAccessDenied,
|
||||
expectWorkflowAccessGranted,
|
||||
mockAuth,
|
||||
} from '@sim/testing'
|
||||
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
|
||||
const mockDb = databaseMock.db
|
||||
vi.mock('@sim/db', () => databaseMock)
|
||||
|
||||
// Mock the auth module
|
||||
vi.mock('@/lib/auth', () => ({
|
||||
getSession: vi.fn(),
|
||||
}))
|
||||
|
||||
import { db } from '@sim/db'
|
||||
import { getSession } from '@/lib/auth'
|
||||
// Import after mocks are set up
|
||||
import { validateWorkflowPermissions } from '@/lib/workflows/utils'
|
||||
|
||||
describe('validateWorkflowPermissions', () => {
|
||||
const auth = mockAuth()
|
||||
|
||||
const mockSession = createSession({ userId: 'user-1', email: 'user1@test.com' })
|
||||
const mockWorkflow = createWorkflowRecord({
|
||||
id: 'wf-1',
|
||||
@@ -35,17 +42,13 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
|
||||
beforeEach(() => {
|
||||
vi.resetModules()
|
||||
vi.clearAllMocks()
|
||||
|
||||
vi.doMock('@sim/db', () => databaseMock)
|
||||
})
|
||||
|
||||
describe('authentication', () => {
|
||||
it('should return 401 when no session exists', async () => {
|
||||
auth.setUnauthenticated()
|
||||
vi.mocked(getSession).mockResolvedValue(null)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'read')
|
||||
|
||||
expectWorkflowAccessDenied(result, 401)
|
||||
@@ -53,9 +56,8 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
|
||||
it('should return 401 when session has no user id', async () => {
|
||||
auth.mockGetSession.mockResolvedValue({ user: {} } as any)
|
||||
vi.mocked(getSession).mockResolvedValue({ user: {} } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'read')
|
||||
|
||||
expectWorkflowAccessDenied(result, 401)
|
||||
@@ -64,14 +66,14 @@ describe('validateWorkflowPermissions', () => {
|
||||
|
||||
describe('workflow not found', () => {
|
||||
it('should return 404 when workflow does not exist', async () => {
|
||||
auth.mockGetSession.mockResolvedValue(mockSession as any)
|
||||
vi.mocked(getSession).mockResolvedValue(mockSession as any)
|
||||
|
||||
// Mock workflow query to return empty
|
||||
const mockLimit = vi.fn().mockResolvedValue([])
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('non-existent', 'req-1', 'read')
|
||||
|
||||
expectWorkflowAccessDenied(result, 404)
|
||||
@@ -81,42 +83,43 @@ describe('validateWorkflowPermissions', () => {
|
||||
|
||||
describe('owner access', () => {
|
||||
it('should deny access to workflow owner without workspace permissions for read action', async () => {
|
||||
auth.setAuthenticated({ id: 'owner-1', email: 'owner-1@test.com' })
|
||||
const ownerSession = createSession({ userId: 'owner-1' })
|
||||
vi.mocked(getSession).mockResolvedValue(ownerSession as any)
|
||||
|
||||
// Mock workflow query
|
||||
const mockLimit = vi.fn().mockResolvedValue([mockWorkflow])
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'read')
|
||||
|
||||
expectWorkflowAccessDenied(result, 403)
|
||||
})
|
||||
|
||||
it('should deny access to workflow owner without workspace permissions for write action', async () => {
|
||||
auth.setAuthenticated({ id: 'owner-1', email: 'owner-1@test.com' })
|
||||
const ownerSession = createSession({ userId: 'owner-1' })
|
||||
vi.mocked(getSession).mockResolvedValue(ownerSession as any)
|
||||
|
||||
const mockLimit = vi.fn().mockResolvedValue([mockWorkflow])
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'write')
|
||||
|
||||
expectWorkflowAccessDenied(result, 403)
|
||||
})
|
||||
|
||||
it('should deny access to workflow owner without workspace permissions for admin action', async () => {
|
||||
auth.setAuthenticated({ id: 'owner-1', email: 'owner-1@test.com' })
|
||||
const ownerSession = createSession({ userId: 'owner-1' })
|
||||
vi.mocked(getSession).mockResolvedValue(ownerSession as any)
|
||||
|
||||
const mockLimit = vi.fn().mockResolvedValue([mockWorkflow])
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'admin')
|
||||
|
||||
expectWorkflowAccessDenied(result, 403)
|
||||
@@ -125,10 +128,11 @@ describe('validateWorkflowPermissions', () => {
|
||||
|
||||
describe('workspace member access with permissions', () => {
|
||||
beforeEach(() => {
|
||||
auth.mockGetSession.mockResolvedValue(mockSession as any)
|
||||
vi.mocked(getSession).mockResolvedValue(mockSession as any)
|
||||
})
|
||||
|
||||
it('should grant read access to user with read permission', async () => {
|
||||
// First call: workflow query, second call: workspace owner, third call: permission
|
||||
let callCount = 0
|
||||
const mockLimit = vi.fn().mockImplementation(() => {
|
||||
callCount++
|
||||
@@ -137,9 +141,8 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'read')
|
||||
|
||||
expectWorkflowAccessGranted(result)
|
||||
@@ -154,9 +157,8 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'write')
|
||||
|
||||
expectWorkflowAccessDenied(result, 403)
|
||||
@@ -172,9 +174,8 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'write')
|
||||
|
||||
expectWorkflowAccessGranted(result)
|
||||
@@ -189,9 +190,8 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'write')
|
||||
|
||||
expectWorkflowAccessGranted(result)
|
||||
@@ -206,9 +206,8 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'admin')
|
||||
|
||||
expectWorkflowAccessDenied(result, 403)
|
||||
@@ -224,9 +223,8 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'admin')
|
||||
|
||||
expectWorkflowAccessGranted(result)
|
||||
@@ -235,19 +233,18 @@ describe('validateWorkflowPermissions', () => {
|
||||
|
||||
describe('no workspace permission', () => {
|
||||
it('should deny access to user without any workspace permission', async () => {
|
||||
auth.mockGetSession.mockResolvedValue(mockSession as any)
|
||||
vi.mocked(getSession).mockResolvedValue(mockSession as any)
|
||||
|
||||
let callCount = 0
|
||||
const mockLimit = vi.fn().mockImplementation(() => {
|
||||
callCount++
|
||||
if (callCount === 1) return Promise.resolve([mockWorkflow])
|
||||
return Promise.resolve([])
|
||||
return Promise.resolve([]) // No permission record
|
||||
})
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1', 'read')
|
||||
|
||||
expectWorkflowAccessDenied(result, 403)
|
||||
@@ -262,14 +259,13 @@ describe('validateWorkflowPermissions', () => {
|
||||
workspaceId: null,
|
||||
})
|
||||
|
||||
auth.mockGetSession.mockResolvedValue(mockSession as any)
|
||||
vi.mocked(getSession).mockResolvedValue(mockSession as any)
|
||||
|
||||
const mockLimit = vi.fn().mockResolvedValue([workflowWithoutWorkspace])
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-2', 'req-1', 'read')
|
||||
|
||||
expectWorkflowAccessDenied(result, 403)
|
||||
@@ -282,14 +278,13 @@ describe('validateWorkflowPermissions', () => {
|
||||
workspaceId: null,
|
||||
})
|
||||
|
||||
auth.mockGetSession.mockResolvedValue(mockSession as any)
|
||||
vi.mocked(getSession).mockResolvedValue(mockSession as any)
|
||||
|
||||
const mockLimit = vi.fn().mockResolvedValue([workflowWithoutWorkspace])
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-2', 'req-1', 'read')
|
||||
|
||||
expectWorkflowAccessDenied(result, 403)
|
||||
@@ -298,7 +293,7 @@ describe('validateWorkflowPermissions', () => {
|
||||
|
||||
describe('default action', () => {
|
||||
it('should default to read action when not specified', async () => {
|
||||
auth.mockGetSession.mockResolvedValue(mockSession as any)
|
||||
vi.mocked(getSession).mockResolvedValue(mockSession as any)
|
||||
|
||||
let callCount = 0
|
||||
const mockLimit = vi.fn().mockImplementation(() => {
|
||||
@@ -308,9 +303,8 @@ describe('validateWorkflowPermissions', () => {
|
||||
})
|
||||
const mockWhere = vi.fn(() => ({ limit: mockLimit }))
|
||||
const mockFrom = vi.fn(() => ({ where: mockWhere }))
|
||||
vi.mocked(mockDb.select).mockReturnValue({ from: mockFrom } as any)
|
||||
vi.mocked(db.select).mockReturnValue({ from: mockFrom } as any)
|
||||
|
||||
const { validateWorkflowPermissions } = await import('@/lib/workflows/utils')
|
||||
const result = await validateWorkflowPermissions('wf-1', 'req-1')
|
||||
|
||||
expectWorkflowAccessGranted(result)
|
||||
|
||||
@@ -1,7 +1,17 @@
|
||||
import { databaseMock, drizzleOrmMock } from '@sim/testing'
|
||||
import { drizzleOrmMock } from '@sim/testing/mocks'
|
||||
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
||||
|
||||
vi.mock('@sim/db', () => databaseMock)
|
||||
vi.mock('@sim/db', () => ({
|
||||
db: {
|
||||
select: vi.fn(),
|
||||
from: vi.fn(),
|
||||
where: vi.fn(),
|
||||
limit: vi.fn(),
|
||||
innerJoin: vi.fn(),
|
||||
leftJoin: vi.fn(),
|
||||
orderBy: vi.fn(),
|
||||
},
|
||||
}))
|
||||
|
||||
vi.mock('@sim/db/schema', () => ({
|
||||
permissions: {
|
||||
|
||||
@@ -10,7 +10,6 @@ import {
|
||||
isCanonicalPair,
|
||||
isNonEmptyValue,
|
||||
isSubBlockFeatureEnabled,
|
||||
isSubBlockHiddenByHostedKey,
|
||||
resolveCanonicalMode,
|
||||
} from '@/lib/workflows/subblocks/visibility'
|
||||
import { getBlock } from '@/blocks'
|
||||
@@ -50,7 +49,6 @@ function shouldSerializeSubBlock(
|
||||
canonicalModeOverrides?: CanonicalModeOverrides
|
||||
): boolean {
|
||||
if (!isSubBlockFeatureEnabled(subBlockConfig)) return false
|
||||
if (isSubBlockHiddenByHostedKey(subBlockConfig)) return false
|
||||
|
||||
if (subBlockConfig.mode === 'trigger') {
|
||||
if (!isTriggerContext && !isTriggerCategory) return false
|
||||
|
||||
@@ -129,18 +129,6 @@ export const useExecutionStore = create<ExecutionState & ExecutionActions>()((se
|
||||
})
|
||||
},
|
||||
|
||||
setCurrentExecutionId: (workflowId, executionId) => {
|
||||
set({
|
||||
workflowExecutions: updatedMap(get().workflowExecutions, workflowId, {
|
||||
currentExecutionId: executionId,
|
||||
}),
|
||||
})
|
||||
},
|
||||
|
||||
getCurrentExecutionId: (workflowId) => {
|
||||
return getOrCreate(get().workflowExecutions, workflowId).currentExecutionId
|
||||
},
|
||||
|
||||
clearRunPath: (workflowId) => {
|
||||
set({
|
||||
workflowExecutions: updatedMap(get().workflowExecutions, workflowId, {
|
||||
|
||||
@@ -35,8 +35,6 @@ export interface WorkflowExecutionState {
|
||||
lastRunPath: Map<string, BlockRunStatus>
|
||||
/** Maps edge IDs to their run result from the last execution */
|
||||
lastRunEdges: Map<string, EdgeRunStatus>
|
||||
/** The execution ID of the currently running execution */
|
||||
currentExecutionId: string | null
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -56,7 +54,6 @@ export const defaultWorkflowExecutionState: WorkflowExecutionState = {
|
||||
debugContext: null,
|
||||
lastRunPath: new Map(),
|
||||
lastRunEdges: new Map(),
|
||||
currentExecutionId: null,
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -99,10 +96,6 @@ export interface ExecutionActions {
|
||||
setEdgeRunStatus: (workflowId: string, edgeId: string, status: EdgeRunStatus) => void
|
||||
/** Clears the run path and run edges for a workflow */
|
||||
clearRunPath: (workflowId: string) => void
|
||||
/** Stores the current execution ID for a workflow */
|
||||
setCurrentExecutionId: (workflowId: string, executionId: string | null) => void
|
||||
/** Returns the current execution ID for a workflow */
|
||||
getCurrentExecutionId: (workflowId: string) => string | null
|
||||
/** Resets the entire store to its initial empty state */
|
||||
reset: () => void
|
||||
/** Stores a serializable execution snapshot for a workflow */
|
||||
|
||||
@@ -1042,7 +1042,7 @@ const cachedAutoAllowedTools = readAutoAllowedToolsFromStorage()
|
||||
// Initial state (subset required for UI/streaming)
|
||||
const initialState = {
|
||||
mode: 'build' as const,
|
||||
selectedModel: 'anthropic/claude-opus-4-5' as CopilotStore['selectedModel'],
|
||||
selectedModel: 'anthropic/claude-opus-4-6' as CopilotStore['selectedModel'],
|
||||
agentPrefetch: false,
|
||||
availableModels: [] as AvailableModel[],
|
||||
isLoadingModels: false,
|
||||
@@ -2381,17 +2381,17 @@ export const useCopilotStore = create<CopilotStore>()(
|
||||
(model) => model.id === normalizedSelectedModel
|
||||
)
|
||||
|
||||
// Pick the best default: prefer claude-opus-4-5 with provider priority:
|
||||
// Pick the best default: prefer claude-opus-4-6 with provider priority:
|
||||
// direct anthropic > bedrock > azure-anthropic > any other.
|
||||
let nextSelectedModel = normalizedSelectedModel
|
||||
if (!selectedModelExists && normalizedModels.length > 0) {
|
||||
let opus45: AvailableModel | undefined
|
||||
let opus46: AvailableModel | undefined
|
||||
for (const prov of MODEL_PROVIDER_PRIORITY) {
|
||||
opus45 = normalizedModels.find((m) => m.id === `${prov}/claude-opus-4-5`)
|
||||
if (opus45) break
|
||||
opus46 = normalizedModels.find((m) => m.id === `${prov}/claude-opus-4-6`)
|
||||
if (opus46) break
|
||||
}
|
||||
if (!opus45) opus45 = normalizedModels.find((m) => m.id.endsWith('/claude-opus-4-5'))
|
||||
nextSelectedModel = opus45 ? opus45.id : normalizedModels[0].id
|
||||
if (!opus46) opus46 = normalizedModels.find((m) => m.id.endsWith('/claude-opus-4-6'))
|
||||
nextSelectedModel = opus46 ? opus46.id : normalizedModels[0].id
|
||||
}
|
||||
|
||||
set({
|
||||
|
||||
@@ -224,7 +224,7 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
||||
|
||||
const newEntry = get().entries[0]
|
||||
|
||||
if (newEntry?.error && newEntry.blockType !== 'cancelled') {
|
||||
if (newEntry?.error) {
|
||||
notifyBlockError({
|
||||
error: newEntry.error,
|
||||
blockName: newEntry.blockName || 'Unknown Block',
|
||||
@@ -243,11 +243,6 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
||||
useExecutionStore.getState().clearRunPath(workflowId)
|
||||
},
|
||||
|
||||
clearExecutionEntries: (executionId: string) =>
|
||||
set((state) => ({
|
||||
entries: state.entries.filter((e) => e.executionId !== executionId),
|
||||
})),
|
||||
|
||||
exportConsoleCSV: (workflowId: string) => {
|
||||
const entries = get().entries.filter((entry) => entry.workflowId === workflowId)
|
||||
|
||||
@@ -475,24 +470,12 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
||||
},
|
||||
merge: (persistedState, currentState) => {
|
||||
const persisted = persistedState as Partial<ConsoleStore> | undefined
|
||||
const rawEntries = persisted?.entries ?? currentState.entries
|
||||
const oneHourAgo = Date.now() - 60 * 60 * 1000
|
||||
|
||||
const entries = rawEntries.map((entry, index) => {
|
||||
let updated = entry
|
||||
const entries = (persisted?.entries ?? currentState.entries).map((entry, index) => {
|
||||
if (entry.executionOrder === undefined) {
|
||||
updated = { ...updated, executionOrder: index + 1 }
|
||||
return { ...entry, executionOrder: index + 1 }
|
||||
}
|
||||
if (
|
||||
entry.isRunning &&
|
||||
entry.startedAt &&
|
||||
new Date(entry.startedAt).getTime() < oneHourAgo
|
||||
) {
|
||||
updated = { ...updated, isRunning: false }
|
||||
}
|
||||
return updated
|
||||
return entry
|
||||
})
|
||||
|
||||
return {
|
||||
...currentState,
|
||||
entries,
|
||||
|
||||
@@ -51,7 +51,6 @@ export interface ConsoleStore {
|
||||
isOpen: boolean
|
||||
addConsole: (entry: Omit<ConsoleEntry, 'id' | 'timestamp'>) => ConsoleEntry
|
||||
clearWorkflowConsole: (workflowId: string) => void
|
||||
clearExecutionEntries: (executionId: string) => void
|
||||
exportConsoleCSV: (workflowId: string) => void
|
||||
getWorkflowEntries: (workflowId: string) => ConsoleEntry[]
|
||||
toggleConsole: () => void
|
||||
|
||||
@@ -1,9 +1,5 @@
|
||||
import { createLogger } from '@sim/logger'
|
||||
import { generateInternalToken } from '@/lib/auth/internal'
|
||||
import { getBYOKKey } from '@/lib/api-key/byok'
|
||||
import { logFixedUsage } from '@/lib/billing/core/usage-log'
|
||||
import { env } from '@/lib/core/config/env'
|
||||
import { isHosted } from '@/lib/core/config/feature-flags'
|
||||
import { DEFAULT_EXECUTION_TIMEOUT_MS } from '@/lib/core/execution-limits'
|
||||
import {
|
||||
secureFetchWithPinnedIP,
|
||||
@@ -17,12 +13,7 @@ import { resolveSkillContent } from '@/executor/handlers/agent/skills-resolver'
|
||||
import type { ExecutionContext } from '@/executor/types'
|
||||
import type { ErrorInfo } from '@/tools/error-extractors'
|
||||
import { extractErrorMessage } from '@/tools/error-extractors'
|
||||
import type {
|
||||
OAuthTokenPayload,
|
||||
ToolConfig,
|
||||
ToolHostingPricing,
|
||||
ToolResponse,
|
||||
} from '@/tools/types'
|
||||
import type { OAuthTokenPayload, ToolConfig, ToolResponse } from '@/tools/types'
|
||||
import {
|
||||
formatRequestParams,
|
||||
getTool,
|
||||
@@ -32,195 +23,6 @@ import {
|
||||
|
||||
const logger = createLogger('Tools')
|
||||
|
||||
/**
|
||||
* Get a hosted API key from environment variables
|
||||
* Supports rotation when multiple keys are configured
|
||||
*/
|
||||
function getHostedKeyFromEnv(envKeys: string[]): string | null {
|
||||
const keys = envKeys
|
||||
.map((key) => env[key as keyof typeof env])
|
||||
.filter((value): value is string => Boolean(value))
|
||||
|
||||
if (keys.length === 0) return null
|
||||
|
||||
// Round-robin rotation based on current minute
|
||||
const currentMinute = Math.floor(Date.now() / 60000)
|
||||
const keyIndex = currentMinute % keys.length
|
||||
|
||||
return keys[keyIndex]
|
||||
}
|
||||
|
||||
/**
|
||||
* Inject hosted API key if tool supports it and user didn't provide one.
|
||||
* Checks BYOK workspace keys first, then falls back to hosted env keys.
|
||||
* Returns whether a hosted (billable) key was injected.
|
||||
*/
|
||||
async function injectHostedKeyIfNeeded(
|
||||
tool: ToolConfig,
|
||||
params: Record<string, unknown>,
|
||||
executionContext: ExecutionContext | undefined,
|
||||
requestId: string
|
||||
): Promise<boolean> {
|
||||
if (!tool.hosting) return false
|
||||
if (!isHosted) return false
|
||||
|
||||
const { envKeys, apiKeyParam, byokProviderId } = tool.hosting
|
||||
const userProvidedKey = params[apiKeyParam]
|
||||
|
||||
if (userProvidedKey) {
|
||||
logger.debug(`[${requestId}] User provided API key for ${tool.id}, skipping hosted key`)
|
||||
return false
|
||||
}
|
||||
|
||||
// Check BYOK workspace key first
|
||||
if (byokProviderId && executionContext?.workspaceId) {
|
||||
try {
|
||||
const byokResult = await getBYOKKey(
|
||||
executionContext.workspaceId,
|
||||
byokProviderId as 'openai' | 'anthropic' | 'google' | 'mistral' | 'serper'
|
||||
)
|
||||
if (byokResult) {
|
||||
params[apiKeyParam] = byokResult.apiKey
|
||||
logger.info(`[${requestId}] Using BYOK key for ${tool.id}`)
|
||||
return false // Don't bill - user's own key
|
||||
}
|
||||
} catch (error) {
|
||||
logger.error(`[${requestId}] Failed to get BYOK key for ${tool.id}:`, error)
|
||||
// Fall through to hosted key
|
||||
}
|
||||
}
|
||||
|
||||
// Fall back to hosted env key
|
||||
const hostedKey = getHostedKeyFromEnv(envKeys)
|
||||
if (!hostedKey) {
|
||||
logger.debug(`[${requestId}] No hosted key available for ${tool.id}`)
|
||||
return false
|
||||
}
|
||||
|
||||
params[apiKeyParam] = hostedKey
|
||||
logger.info(`[${requestId}] Using hosted key for ${tool.id}`)
|
||||
return true // Bill the user
|
||||
}
|
||||
|
||||
/**
|
||||
* Check if an error is a rate limit (throttling) error
|
||||
*/
|
||||
function isRateLimitError(error: unknown): boolean {
|
||||
if (error && typeof error === 'object') {
|
||||
const status = (error as { status?: number }).status
|
||||
// 429 = Too Many Requests, 503 = Service Unavailable (sometimes used for rate limiting)
|
||||
if (status === 429 || status === 503) return true
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
||||
/**
|
||||
* Execute a function with exponential backoff retry for rate limiting errors.
|
||||
* Only used for hosted key requests.
|
||||
*/
|
||||
async function executeWithRetry<T>(
|
||||
fn: () => Promise<T>,
|
||||
requestId: string,
|
||||
toolId: string,
|
||||
maxRetries = 3,
|
||||
baseDelayMs = 1000
|
||||
): Promise<T> {
|
||||
let lastError: unknown
|
||||
|
||||
for (let attempt = 0; attempt <= maxRetries; attempt++) {
|
||||
try {
|
||||
return await fn()
|
||||
} catch (error) {
|
||||
lastError = error
|
||||
|
||||
if (!isRateLimitError(error) || attempt === maxRetries) {
|
||||
throw error
|
||||
}
|
||||
|
||||
const delayMs = baseDelayMs * Math.pow(2, attempt)
|
||||
logger.warn(`[${requestId}] Rate limited for ${toolId}, retrying in ${delayMs}ms (attempt ${attempt + 1}/${maxRetries})`)
|
||||
await new Promise((resolve) => setTimeout(resolve, delayMs))
|
||||
}
|
||||
}
|
||||
|
||||
throw lastError
|
||||
}
|
||||
|
||||
/**
|
||||
* Calculate cost based on pricing model
|
||||
*/
|
||||
function calculateToolCost(
|
||||
pricing: ToolHostingPricing,
|
||||
params: Record<string, unknown>,
|
||||
response: Record<string, unknown>
|
||||
): number {
|
||||
switch (pricing.type) {
|
||||
case 'per_request':
|
||||
return pricing.cost
|
||||
|
||||
case 'per_unit': {
|
||||
const usage = pricing.getUsage(params, response)
|
||||
return usage * pricing.costPerUnit
|
||||
}
|
||||
|
||||
case 'per_result': {
|
||||
const resultCount = pricing.getResultCount(response)
|
||||
const billableResults = pricing.maxResults
|
||||
? Math.min(resultCount, pricing.maxResults)
|
||||
: resultCount
|
||||
return billableResults * pricing.costPerResult
|
||||
}
|
||||
|
||||
case 'per_second': {
|
||||
const duration = pricing.getDuration(response)
|
||||
const billableDuration = pricing.minimumSeconds
|
||||
? Math.max(duration, pricing.minimumSeconds)
|
||||
: duration
|
||||
return billableDuration * pricing.costPerSecond
|
||||
}
|
||||
|
||||
default: {
|
||||
const exhaustiveCheck: never = pricing
|
||||
throw new Error(`Unknown pricing type: ${(exhaustiveCheck as ToolHostingPricing).type}`)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Log usage for a tool that used a hosted API key
|
||||
*/
|
||||
async function logHostedToolUsage(
|
||||
tool: ToolConfig,
|
||||
params: Record<string, unknown>,
|
||||
response: Record<string, unknown>,
|
||||
executionContext: ExecutionContext | undefined,
|
||||
requestId: string
|
||||
): Promise<void> {
|
||||
if (!tool.hosting?.pricing || !executionContext?.userId) {
|
||||
return
|
||||
}
|
||||
|
||||
const cost = calculateToolCost(tool.hosting.pricing, params, response)
|
||||
|
||||
if (cost <= 0) return
|
||||
|
||||
try {
|
||||
await logFixedUsage({
|
||||
userId: executionContext.userId,
|
||||
source: 'workflow',
|
||||
description: `tool:${tool.id}`,
|
||||
cost,
|
||||
workspaceId: executionContext.workspaceId,
|
||||
workflowId: executionContext.workflowId,
|
||||
executionId: executionContext.executionId,
|
||||
})
|
||||
logger.debug(`[${requestId}] Logged hosted tool usage for ${tool.id}: $${cost}`)
|
||||
} catch (error) {
|
||||
logger.error(`[${requestId}] Failed to log hosted tool usage for ${tool.id}:`, error)
|
||||
// Don't throw - usage logging should not break the main flow
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Normalizes a tool ID by stripping resource ID suffix (UUID).
|
||||
* Workflow tools: 'workflow_executor_<uuid>' -> 'workflow_executor'
|
||||
@@ -477,14 +279,6 @@ export async function executeTool(
|
||||
throw new Error(`Tool not found: ${toolId}`)
|
||||
}
|
||||
|
||||
// Inject hosted API key if tool supports it and user didn't provide one
|
||||
const isUsingHostedKey = await injectHostedKeyIfNeeded(
|
||||
tool,
|
||||
contextParams,
|
||||
executionContext,
|
||||
requestId
|
||||
)
|
||||
|
||||
// If we have a credential parameter, fetch the access token
|
||||
if (contextParams.credential) {
|
||||
logger.info(
|
||||
@@ -593,11 +387,6 @@ export async function executeTool(
|
||||
// Process file outputs if execution context is available
|
||||
finalResult = await processFileOutputs(finalResult, tool, executionContext)
|
||||
|
||||
// Log usage for hosted key if execution was successful
|
||||
if (isUsingHostedKey && finalResult.success) {
|
||||
await logHostedToolUsage(tool, contextParams, finalResult.output, executionContext, requestId)
|
||||
}
|
||||
|
||||
// Add timing data to the result
|
||||
const endTime = new Date()
|
||||
const endTimeISO = endTime.toISOString()
|
||||
@@ -613,14 +402,7 @@ export async function executeTool(
|
||||
}
|
||||
|
||||
// Execute the tool request directly (internal routes use regular fetch, external use SSRF-protected fetch)
|
||||
// Wrap with retry logic for hosted keys to handle rate limiting due to higher usage
|
||||
const result = isUsingHostedKey
|
||||
? await executeWithRetry(
|
||||
() => executeToolRequest(toolId, tool, contextParams),
|
||||
requestId,
|
||||
toolId
|
||||
)
|
||||
: await executeToolRequest(toolId, tool, contextParams)
|
||||
const result = await executeToolRequest(toolId, tool, contextParams)
|
||||
|
||||
// Apply post-processing if available and not skipped
|
||||
let finalResult = result
|
||||
@@ -638,11 +420,6 @@ export async function executeTool(
|
||||
// Process file outputs if execution context is available
|
||||
finalResult = await processFileOutputs(finalResult, tool, executionContext)
|
||||
|
||||
// Log usage for hosted key if execution was successful
|
||||
if (isUsingHostedKey && finalResult.success) {
|
||||
await logHostedToolUsage(tool, contextParams, finalResult.output, executionContext, requestId)
|
||||
}
|
||||
|
||||
// Add timing data to the result
|
||||
const endTime = new Date()
|
||||
const endTimeISO = endTime.toISOString()
|
||||
|
||||
@@ -26,13 +26,6 @@ export const s3GetObjectTool: ToolConfig = {
|
||||
visibility: 'user-only',
|
||||
description: 'Your AWS Secret Access Key',
|
||||
},
|
||||
region: {
|
||||
type: 'string',
|
||||
required: false,
|
||||
visibility: 'user-only',
|
||||
description:
|
||||
'Optional region override when URL does not include region (e.g., us-east-1, eu-west-1)',
|
||||
},
|
||||
s3Uri: {
|
||||
type: 'string',
|
||||
required: true,
|
||||
@@ -44,7 +37,7 @@ export const s3GetObjectTool: ToolConfig = {
|
||||
request: {
|
||||
url: (params) => {
|
||||
try {
|
||||
const { bucketName, region, objectKey } = parseS3Uri(params.s3Uri, params.region)
|
||||
const { bucketName, region, objectKey } = parseS3Uri(params.s3Uri)
|
||||
|
||||
params.bucketName = bucketName
|
||||
params.region = region
|
||||
@@ -53,7 +46,7 @@ export const s3GetObjectTool: ToolConfig = {
|
||||
return `https://${bucketName}.s3.${region}.amazonaws.com/${encodeS3PathComponent(objectKey)}`
|
||||
} catch (_error) {
|
||||
throw new Error(
|
||||
'Invalid S3 Object URL. Use a valid S3 URL and optionally provide region if the URL omits it.'
|
||||
'Invalid S3 Object URL format. Expected format: https://bucket-name.s3.region.amazonaws.com/path/to/file'
|
||||
)
|
||||
}
|
||||
},
|
||||
@@ -62,7 +55,7 @@ export const s3GetObjectTool: ToolConfig = {
|
||||
try {
|
||||
// Parse S3 URI if not already parsed
|
||||
if (!params.bucketName || !params.region || !params.objectKey) {
|
||||
const { bucketName, region, objectKey } = parseS3Uri(params.s3Uri, params.region)
|
||||
const { bucketName, region, objectKey } = parseS3Uri(params.s3Uri)
|
||||
params.bucketName = bucketName
|
||||
params.region = region
|
||||
params.objectKey = objectKey
|
||||
@@ -109,7 +102,7 @@ export const s3GetObjectTool: ToolConfig = {
|
||||
transformResponse: async (response: Response, params) => {
|
||||
// Parse S3 URI if not already parsed
|
||||
if (!params.bucketName || !params.region || !params.objectKey) {
|
||||
const { bucketName, region, objectKey } = parseS3Uri(params.s3Uri, params.region)
|
||||
const { bucketName, region, objectKey } = parseS3Uri(params.s3Uri)
|
||||
params.bucketName = bucketName
|
||||
params.region = region
|
||||
params.objectKey = objectKey
|
||||
|
||||
@@ -20,10 +20,7 @@ export function getSignatureKey(
|
||||
return kSigning
|
||||
}
|
||||
|
||||
export function parseS3Uri(
|
||||
s3Uri: string,
|
||||
fallbackRegion?: string
|
||||
): {
|
||||
export function parseS3Uri(s3Uri: string): {
|
||||
bucketName: string
|
||||
region: string
|
||||
objectKey: string
|
||||
@@ -31,55 +28,10 @@ export function parseS3Uri(
|
||||
try {
|
||||
const url = new URL(s3Uri)
|
||||
const hostname = url.hostname
|
||||
const normalizedPath = url.pathname.startsWith('/') ? url.pathname.slice(1) : url.pathname
|
||||
|
||||
const virtualHostedDualstackMatch = hostname.match(
|
||||
/^(.+)\.s3\.dualstack\.([^.]+)\.amazonaws\.com(?:\.cn)?$/
|
||||
)
|
||||
const virtualHostedRegionalMatch = hostname.match(
|
||||
/^(.+)\.s3[.-]([^.]+)\.amazonaws\.com(?:\.cn)?$/
|
||||
)
|
||||
const virtualHostedGlobalMatch = hostname.match(/^(.+)\.s3\.amazonaws\.com(?:\.cn)?$/)
|
||||
|
||||
const pathStyleDualstackMatch = hostname.match(
|
||||
/^s3\.dualstack\.([^.]+)\.amazonaws\.com(?:\.cn)?$/
|
||||
)
|
||||
const pathStyleRegionalMatch = hostname.match(/^s3[.-]([^.]+)\.amazonaws\.com(?:\.cn)?$/)
|
||||
const pathStyleGlobalMatch = hostname.match(/^s3\.amazonaws\.com(?:\.cn)?$/)
|
||||
|
||||
const isPathStyleHost = Boolean(
|
||||
pathStyleDualstackMatch || pathStyleRegionalMatch || pathStyleGlobalMatch
|
||||
)
|
||||
|
||||
const firstSlashIndex = normalizedPath.indexOf('/')
|
||||
const pathStyleBucketName =
|
||||
firstSlashIndex === -1 ? normalizedPath : normalizedPath.slice(0, firstSlashIndex)
|
||||
const pathStyleObjectKey =
|
||||
firstSlashIndex === -1 ? '' : normalizedPath.slice(firstSlashIndex + 1)
|
||||
|
||||
const bucketName = isPathStyleHost
|
||||
? pathStyleBucketName
|
||||
: (virtualHostedDualstackMatch?.[1] ??
|
||||
virtualHostedRegionalMatch?.[1] ??
|
||||
virtualHostedGlobalMatch?.[1] ??
|
||||
'')
|
||||
|
||||
const rawObjectKey = isPathStyleHost ? pathStyleObjectKey : normalizedPath
|
||||
const objectKey = (() => {
|
||||
try {
|
||||
return decodeURIComponent(rawObjectKey)
|
||||
} catch {
|
||||
return rawObjectKey
|
||||
}
|
||||
})()
|
||||
|
||||
const normalizedFallbackRegion = fallbackRegion?.trim()
|
||||
const regionFromHost =
|
||||
virtualHostedDualstackMatch?.[2] ??
|
||||
virtualHostedRegionalMatch?.[2] ??
|
||||
pathStyleDualstackMatch?.[1] ??
|
||||
pathStyleRegionalMatch?.[1]
|
||||
const region = regionFromHost || normalizedFallbackRegion || 'us-east-1'
|
||||
const bucketName = hostname.split('.')[0]
|
||||
const regionMatch = hostname.match(/s3[.-]([^.]+)\.amazonaws\.com/)
|
||||
const region = regionMatch ? regionMatch[1] : 'us-east-1'
|
||||
const objectKey = url.pathname.startsWith('/') ? url.pathname.substring(1) : url.pathname
|
||||
|
||||
if (!bucketName || !objectKey) {
|
||||
throw new Error('Invalid S3 URI format')
|
||||
@@ -88,7 +40,7 @@ export function parseS3Uri(
|
||||
return { bucketName, region, objectKey }
|
||||
} catch (_error) {
|
||||
throw new Error(
|
||||
'Invalid S3 Object URL format. Expected S3 virtual-hosted or path-style URL with object key.'
|
||||
'Invalid S3 Object URL format. Expected format: https://bucket-name.s3.region.amazonaws.com/path/to/file'
|
||||
)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -48,15 +48,6 @@ export const searchTool: ToolConfig<SearchParams, SearchResponse> = {
|
||||
description: 'Serper API Key',
|
||||
},
|
||||
},
|
||||
hosting: {
|
||||
envKeys: ['SERPER_API_KEY'],
|
||||
apiKeyParam: 'apiKey',
|
||||
byokProviderId: 'serper',
|
||||
pricing: {
|
||||
type: 'per_request',
|
||||
cost: 0.001, // $0.001 per search (Serper pricing: ~$50/50k searches)
|
||||
},
|
||||
},
|
||||
|
||||
request: {
|
||||
url: (params) => `https://google.serper.dev/${params.type || 'search'}`,
|
||||
|
||||
@@ -127,13 +127,6 @@ export interface ToolConfig<P = any, R = any> {
|
||||
* Maps param IDs to their enrichment configuration.
|
||||
*/
|
||||
schemaEnrichment?: Record<string, SchemaEnrichmentConfig>
|
||||
|
||||
/**
|
||||
* Hosted API key configuration for this tool.
|
||||
* When configured, the tool can use Sim's hosted API keys if user doesn't provide their own.
|
||||
* Usage is billed according to the pricing config.
|
||||
*/
|
||||
hosting?: ToolHostingConfig
|
||||
}
|
||||
|
||||
export interface TableRow {
|
||||
@@ -177,64 +170,3 @@ export interface SchemaEnrichmentConfig {
|
||||
required?: string[]
|
||||
} | null>
|
||||
}
|
||||
|
||||
/**
|
||||
* Pricing models for hosted API key usage
|
||||
*/
|
||||
/** Flat fee per API call (e.g., Serper search) */
|
||||
export interface PerRequestPricing {
|
||||
type: 'per_request'
|
||||
/** Cost per request in dollars */
|
||||
cost: number
|
||||
}
|
||||
|
||||
/** Usage-based on input/output size (e.g., LLM tokens, TTS characters) */
|
||||
export interface PerUnitPricing {
|
||||
type: 'per_unit'
|
||||
/** Cost per unit in dollars */
|
||||
costPerUnit: number
|
||||
/** Unit of measurement */
|
||||
unit: 'token' | 'character' | 'byte' | 'kb' | 'mb'
|
||||
/** Extract usage count from params (before execution) or response (after execution) */
|
||||
getUsage: (params: Record<string, unknown>, response?: Record<string, unknown>) => number
|
||||
}
|
||||
|
||||
/** Based on result count (e.g., per search result, per email sent) */
|
||||
export interface PerResultPricing {
|
||||
type: 'per_result'
|
||||
/** Cost per result in dollars */
|
||||
costPerResult: number
|
||||
/** Maximum results to bill for (cap) */
|
||||
maxResults?: number
|
||||
/** Extract result count from response */
|
||||
getResultCount: (response: Record<string, unknown>) => number
|
||||
}
|
||||
|
||||
/** Billed by execution duration (e.g., browser sessions, video processing) */
|
||||
export interface PerSecondPricing {
|
||||
type: 'per_second'
|
||||
/** Cost per second in dollars */
|
||||
costPerSecond: number
|
||||
/** Minimum billable seconds */
|
||||
minimumSeconds?: number
|
||||
/** Extract duration from response (in seconds) */
|
||||
getDuration: (response: Record<string, unknown>) => number
|
||||
}
|
||||
|
||||
/** Union of all pricing models */
|
||||
export type ToolHostingPricing = PerRequestPricing | PerUnitPricing | PerResultPricing | PerSecondPricing
|
||||
|
||||
/**
|
||||
* Configuration for hosted API key support
|
||||
* When configured, the tool can use Sim's hosted API keys if user doesn't provide their own
|
||||
*/
|
||||
export interface ToolHostingConfig {
|
||||
/** Environment variable names to check for hosted keys (supports rotation with multiple keys) */
|
||||
envKeys: string[]
|
||||
/** The parameter name that receives the API key */
|
||||
apiKeyParam: string
|
||||
/** BYOK provider ID for workspace key lookup (e.g., 'serper') */
|
||||
byokProviderId?: string
|
||||
/** Pricing when using hosted key */
|
||||
pricing: ToolHostingPricing
|
||||
}
|
||||
|
||||
@@ -23,12 +23,7 @@ export const SYSTEM_SUBBLOCK_IDS: string[] = [
|
||||
* with default values from the trigger definition on load, which aren't present in
|
||||
* the deployed state, causing false positive change detection.
|
||||
*/
|
||||
export const TRIGGER_RUNTIME_SUBBLOCK_IDS: string[] = [
|
||||
'webhookId',
|
||||
'triggerPath',
|
||||
'triggerConfig',
|
||||
'triggerId',
|
||||
]
|
||||
export const TRIGGER_RUNTIME_SUBBLOCK_IDS: string[] = ['webhookId', 'triggerPath', 'triggerConfig']
|
||||
|
||||
/**
|
||||
* Maximum number of consecutive failures before a trigger (schedule/webhook) is auto-disabled.
|
||||
|
||||
Reference in New Issue
Block a user