mirror of
https://github.com/simstudioai/sim.git
synced 2026-02-13 16:05:09 -05:00
Compare commits
78 Commits
feat/strea
...
main
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
b45f3962fc | ||
|
|
7fbbc7ba7a | ||
|
|
a337aa7dfe | ||
|
|
022e84c4b1 | ||
|
|
602e371a7a | ||
|
|
9a06cae591 | ||
|
|
dce47a101c | ||
|
|
1130f8ddb2 | ||
|
|
ebc2ffa1c5 | ||
|
|
fc97ce007d | ||
|
|
6c006cdfec | ||
|
|
c380e59cb3 | ||
|
|
2944579d21 | ||
|
|
81dfeb0bb0 | ||
|
|
01577a18b4 | ||
|
|
07d50f8fe1 | ||
|
|
52aff4d60b | ||
|
|
3a3bddd6f8 | ||
|
|
639d50d6b9 | ||
|
|
cec74e09c2 | ||
|
|
d5a756c9f2 | ||
|
|
f3e994baf0 | ||
|
|
2f492cacc1 | ||
|
|
27973953f6 | ||
|
|
5792e7e5f9 | ||
|
|
50585273ce | ||
|
|
654cb2b407 | ||
|
|
6c66521d64 | ||
|
|
479cd347ad | ||
|
|
a3a99eda19 | ||
|
|
1a66d48add | ||
|
|
46822e91f3 | ||
|
|
2bb68335ee | ||
|
|
8528fbe2d2 | ||
|
|
31fdd2be13 | ||
|
|
028bc652c2 | ||
|
|
c6bf5cd58c | ||
|
|
11dc18a80d | ||
|
|
ab4e9dc72f | ||
|
|
1c58c35bd8 | ||
|
|
d63a5cb504 | ||
|
|
8bd5d41723 | ||
|
|
c12931bc50 | ||
|
|
e9c4251c1c | ||
|
|
cc2be33d6b | ||
|
|
45371e521e | ||
|
|
0ce0f98aa5 | ||
|
|
dff1c9d083 | ||
|
|
b09f683072 | ||
|
|
a8bb0db660 | ||
|
|
af82820a28 | ||
|
|
4372841797 | ||
|
|
5e8c843241 | ||
|
|
7bf3d73ee6 | ||
|
|
7ffc11a738 | ||
|
|
be578e2ed7 | ||
|
|
f415e5edc4 | ||
|
|
13a6e6c3fa | ||
|
|
f5ab7f21ae | ||
|
|
bfb6fffe38 | ||
|
|
4fbec0a43f | ||
|
|
585f5e365b | ||
|
|
3792bdd252 | ||
|
|
eb5d1f3e5b | ||
|
|
54ab82c8dd | ||
|
|
f895bf469b | ||
|
|
dd3209af06 | ||
|
|
b6ba3b50a7 | ||
|
|
b304233062 | ||
|
|
57e4b49bd6 | ||
|
|
e12dd204ed | ||
|
|
3d9d9cbc54 | ||
|
|
0f4ec962ad | ||
|
|
4827866f9a | ||
|
|
3e697d9ed9 | ||
|
|
4431a1a484 | ||
|
|
4d1a9a3f22 | ||
|
|
eb07a080fb |
@@ -1157,6 +1157,21 @@ export function AirweaveIcon(props: SVGProps<SVGSVGElement>) {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function GoogleBooksIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
|
return (
|
||||||
|
<svg {...props} xmlns='http://www.w3.org/2000/svg' viewBox='0 0 478.633 540.068'>
|
||||||
|
<path
|
||||||
|
fill='#1C51A4'
|
||||||
|
d='M449.059,218.231L245.519,99.538l-0.061,193.23c0.031,1.504-0.368,2.977-1.166,4.204c-0.798,1.258-1.565,1.995-2.915,2.547c-1.35,0.552-2.792,0.706-4.204,0.399c-1.412-0.307-2.7-1.043-3.713-2.117l-69.166-70.609l-69.381,70.179c-1.013,0.982-2.301,1.657-3.652,1.903c-1.381,0.246-2.792,0.092-4.081-0.491c-1.289-0.583-1.626-0.522-2.394-1.749c-0.767-1.197-1.197-2.608-1.197-4.081L85.031,6.007l-2.915-1.289C43.973-11.638,0,16.409,0,59.891v420.306c0,46.029,49.312,74.782,88.775,51.767l360.285-210.138C488.491,298.782,488.491,241.246,449.059,218.231z'
|
||||||
|
/>
|
||||||
|
<path
|
||||||
|
fill='#80D7FB'
|
||||||
|
d='M88.805,8.124c-2.179-1.289-4.419-2.363-6.659-3.345l0.123,288.663c0,1.442,0.43,2.854,1.197,4.081c0.767,1.197,1.872,2.148,3.161,2.731c1.289,0.583,2.7,0.736,4.081,0.491c1.381-0.246,2.639-0.921,3.652-1.903l69.749-69.688l69.811,69.749c1.013,1.074,2.301,1.81,3.713,2.117c1.412,0.307,2.884,0.153,4.204-0.399c1.319-0.552,2.455-1.565,3.253-2.792c0.798-1.258,1.197-2.731,1.166-4.204V99.998L88.805,8.124z'
|
||||||
|
/>
|
||||||
|
</svg>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
export function GoogleDocsIcon(props: SVGProps<SVGSVGElement>) {
|
export function GoogleDocsIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
return (
|
return (
|
||||||
<svg
|
<svg
|
||||||
|
|||||||
@@ -38,6 +38,7 @@ import {
|
|||||||
GithubIcon,
|
GithubIcon,
|
||||||
GitLabIcon,
|
GitLabIcon,
|
||||||
GmailIcon,
|
GmailIcon,
|
||||||
|
GoogleBooksIcon,
|
||||||
GoogleCalendarIcon,
|
GoogleCalendarIcon,
|
||||||
GoogleDocsIcon,
|
GoogleDocsIcon,
|
||||||
GoogleDriveIcon,
|
GoogleDriveIcon,
|
||||||
@@ -172,6 +173,7 @@ export const blockTypeToIconMap: Record<string, IconComponent> = {
|
|||||||
github_v2: GithubIcon,
|
github_v2: GithubIcon,
|
||||||
gitlab: GitLabIcon,
|
gitlab: GitLabIcon,
|
||||||
gmail_v2: GmailIcon,
|
gmail_v2: GmailIcon,
|
||||||
|
google_books: GoogleBooksIcon,
|
||||||
google_calendar_v2: GoogleCalendarIcon,
|
google_calendar_v2: GoogleCalendarIcon,
|
||||||
google_docs: GoogleDocsIcon,
|
google_docs: GoogleDocsIcon,
|
||||||
google_drive: GoogleDriveIcon,
|
google_drive: GoogleDriveIcon,
|
||||||
|
|||||||
@@ -41,9 +41,6 @@ Diese Tastenkombinationen wechseln zwischen den Panel-Tabs auf der rechten Seite
|
|||||||
|
|
||||||
| Tastenkombination | Aktion |
|
| Tastenkombination | Aktion |
|
||||||
|----------|--------|
|
|----------|--------|
|
||||||
| `C` | Copilot-Tab fokussieren |
|
|
||||||
| `T` | Toolbar-Tab fokussieren |
|
|
||||||
| `E` | Editor-Tab fokussieren |
|
|
||||||
| `Mod` + `F` | Toolbar-Suche fokussieren |
|
| `Mod` + `F` | Toolbar-Suche fokussieren |
|
||||||
|
|
||||||
## Globale Navigation
|
## Globale Navigation
|
||||||
|
|||||||
@@ -43,9 +43,6 @@ These shortcuts switch between panel tabs on the right side of the canvas.
|
|||||||
|
|
||||||
| Shortcut | Action |
|
| Shortcut | Action |
|
||||||
|----------|--------|
|
|----------|--------|
|
||||||
| `C` | Focus Copilot tab |
|
|
||||||
| `T` | Focus Toolbar tab |
|
|
||||||
| `E` | Focus Editor tab |
|
|
||||||
| `Mod` + `F` | Focus Toolbar search |
|
| `Mod` + `F` | Focus Toolbar search |
|
||||||
|
|
||||||
## Global Navigation
|
## Global Navigation
|
||||||
|
|||||||
@@ -399,6 +399,28 @@ Create a new custom property (metadata) on a Confluence page.
|
|||||||
| ↳ `authorId` | string | Account ID of the version author |
|
| ↳ `authorId` | string | Account ID of the version author |
|
||||||
| ↳ `createdAt` | string | ISO 8601 timestamp of version creation |
|
| ↳ `createdAt` | string | ISO 8601 timestamp of version creation |
|
||||||
|
|
||||||
|
### `confluence_delete_page_property`
|
||||||
|
|
||||||
|
Delete a content property from a Confluence page by its property ID.
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `domain` | string | Yes | Your Confluence domain \(e.g., yourcompany.atlassian.net\) |
|
||||||
|
| `pageId` | string | Yes | The ID of the page containing the property |
|
||||||
|
| `propertyId` | string | Yes | The ID of the property to delete |
|
||||||
|
| `cloudId` | string | No | Confluence Cloud ID for the instance. If not provided, it will be fetched using the domain. |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `ts` | string | ISO 8601 timestamp of the operation |
|
||||||
|
| `pageId` | string | ID of the page |
|
||||||
|
| `propertyId` | string | ID of the deleted property |
|
||||||
|
| `deleted` | boolean | Deletion status |
|
||||||
|
|
||||||
### `confluence_search`
|
### `confluence_search`
|
||||||
|
|
||||||
Search for content across Confluence pages, blog posts, and other content.
|
Search for content across Confluence pages, blog posts, and other content.
|
||||||
@@ -872,6 +894,90 @@ Add a label to a Confluence page for organization and categorization.
|
|||||||
| `labelName` | string | Name of the added label |
|
| `labelName` | string | Name of the added label |
|
||||||
| `labelId` | string | ID of the added label |
|
| `labelId` | string | ID of the added label |
|
||||||
|
|
||||||
|
### `confluence_delete_label`
|
||||||
|
|
||||||
|
Remove a label from a Confluence page.
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `domain` | string | Yes | Your Confluence domain \(e.g., yourcompany.atlassian.net\) |
|
||||||
|
| `pageId` | string | Yes | Confluence page ID to remove the label from |
|
||||||
|
| `labelName` | string | Yes | Name of the label to remove |
|
||||||
|
| `cloudId` | string | No | Confluence Cloud ID for the instance. If not provided, it will be fetched using the domain. |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `ts` | string | ISO 8601 timestamp of the operation |
|
||||||
|
| `pageId` | string | Page ID the label was removed from |
|
||||||
|
| `labelName` | string | Name of the removed label |
|
||||||
|
| `deleted` | boolean | Deletion status |
|
||||||
|
|
||||||
|
### `confluence_get_pages_by_label`
|
||||||
|
|
||||||
|
Retrieve all pages that have a specific label applied.
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `domain` | string | Yes | Your Confluence domain \(e.g., yourcompany.atlassian.net\) |
|
||||||
|
| `labelId` | string | Yes | The ID of the label to get pages for |
|
||||||
|
| `limit` | number | No | Maximum number of pages to return \(default: 50, max: 250\) |
|
||||||
|
| `cursor` | string | No | Pagination cursor from previous response |
|
||||||
|
| `cloudId` | string | No | Confluence Cloud ID for the instance. If not provided, it will be fetched using the domain. |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `ts` | string | ISO 8601 timestamp of the operation |
|
||||||
|
| `labelId` | string | ID of the label |
|
||||||
|
| `pages` | array | Array of pages with this label |
|
||||||
|
| ↳ `id` | string | Unique page identifier |
|
||||||
|
| ↳ `title` | string | Page title |
|
||||||
|
| ↳ `status` | string | Page status \(e.g., current, archived, trashed, draft\) |
|
||||||
|
| ↳ `spaceId` | string | ID of the space containing the page |
|
||||||
|
| ↳ `parentId` | string | ID of the parent page \(null if top-level\) |
|
||||||
|
| ↳ `authorId` | string | Account ID of the page author |
|
||||||
|
| ↳ `createdAt` | string | ISO 8601 timestamp when the page was created |
|
||||||
|
| ↳ `version` | object | Page version information |
|
||||||
|
| ↳ `number` | number | Version number |
|
||||||
|
| ↳ `message` | string | Version message |
|
||||||
|
| ↳ `minorEdit` | boolean | Whether this is a minor edit |
|
||||||
|
| ↳ `authorId` | string | Account ID of the version author |
|
||||||
|
| ↳ `createdAt` | string | ISO 8601 timestamp of version creation |
|
||||||
|
| `nextCursor` | string | Cursor for fetching the next page of results |
|
||||||
|
|
||||||
|
### `confluence_list_space_labels`
|
||||||
|
|
||||||
|
List all labels associated with a Confluence space.
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `domain` | string | Yes | Your Confluence domain \(e.g., yourcompany.atlassian.net\) |
|
||||||
|
| `spaceId` | string | Yes | The ID of the Confluence space to list labels from |
|
||||||
|
| `limit` | number | No | Maximum number of labels to return \(default: 25, max: 250\) |
|
||||||
|
| `cursor` | string | No | Pagination cursor from previous response |
|
||||||
|
| `cloudId` | string | No | Confluence Cloud ID for the instance. If not provided, it will be fetched using the domain. |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `ts` | string | ISO 8601 timestamp of the operation |
|
||||||
|
| `spaceId` | string | ID of the space |
|
||||||
|
| `labels` | array | Array of labels on the space |
|
||||||
|
| ↳ `id` | string | Unique label identifier |
|
||||||
|
| ↳ `name` | string | Label name |
|
||||||
|
| ↳ `prefix` | string | Label prefix/type \(e.g., global, my, team\) |
|
||||||
|
| `nextCursor` | string | Cursor for fetching the next page of results |
|
||||||
|
|
||||||
### `confluence_get_space`
|
### `confluence_get_space`
|
||||||
|
|
||||||
Get details about a specific Confluence space.
|
Get details about a specific Confluence space.
|
||||||
|
|||||||
96
apps/docs/content/docs/en/tools/google_books.mdx
Normal file
96
apps/docs/content/docs/en/tools/google_books.mdx
Normal file
@@ -0,0 +1,96 @@
|
|||||||
|
---
|
||||||
|
title: Google Books
|
||||||
|
description: Search and retrieve book information
|
||||||
|
---
|
||||||
|
|
||||||
|
import { BlockInfoCard } from "@/components/ui/block-info-card"
|
||||||
|
|
||||||
|
<BlockInfoCard
|
||||||
|
type="google_books"
|
||||||
|
color="#FFFFFF"
|
||||||
|
/>
|
||||||
|
|
||||||
|
## Usage Instructions
|
||||||
|
|
||||||
|
Search for books using the Google Books API. Find volumes by title, author, ISBN, or keywords, and retrieve detailed information about specific books including descriptions, ratings, and publication details.
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
## Tools
|
||||||
|
|
||||||
|
### `google_books_volume_search`
|
||||||
|
|
||||||
|
Search for books using the Google Books API
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `apiKey` | string | Yes | Google Books API key |
|
||||||
|
| `query` | string | Yes | Search query. Supports special keywords: intitle:, inauthor:, inpublisher:, subject:, isbn: |
|
||||||
|
| `filter` | string | No | Filter results by availability \(partial, full, free-ebooks, paid-ebooks, ebooks\) |
|
||||||
|
| `printType` | string | No | Restrict to print type \(all, books, magazines\) |
|
||||||
|
| `orderBy` | string | No | Sort order \(relevance, newest\) |
|
||||||
|
| `startIndex` | number | No | Index of the first result to return \(for pagination\) |
|
||||||
|
| `maxResults` | number | No | Maximum number of results to return \(1-40\) |
|
||||||
|
| `langRestrict` | string | No | Restrict results to a specific language \(ISO 639-1 code\) |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `totalItems` | number | Total number of matching results |
|
||||||
|
| `volumes` | array | List of matching volumes |
|
||||||
|
| ↳ `id` | string | Volume ID |
|
||||||
|
| ↳ `title` | string | Book title |
|
||||||
|
| ↳ `subtitle` | string | Book subtitle |
|
||||||
|
| ↳ `authors` | array | List of authors |
|
||||||
|
| ↳ `publisher` | string | Publisher name |
|
||||||
|
| ↳ `publishedDate` | string | Publication date |
|
||||||
|
| ↳ `description` | string | Book description |
|
||||||
|
| ↳ `pageCount` | number | Number of pages |
|
||||||
|
| ↳ `categories` | array | Book categories |
|
||||||
|
| ↳ `averageRating` | number | Average rating \(1-5\) |
|
||||||
|
| ↳ `ratingsCount` | number | Number of ratings |
|
||||||
|
| ↳ `language` | string | Language code |
|
||||||
|
| ↳ `previewLink` | string | Link to preview on Google Books |
|
||||||
|
| ↳ `infoLink` | string | Link to info page |
|
||||||
|
| ↳ `thumbnailUrl` | string | Book cover thumbnail URL |
|
||||||
|
| ↳ `isbn10` | string | ISBN-10 identifier |
|
||||||
|
| ↳ `isbn13` | string | ISBN-13 identifier |
|
||||||
|
|
||||||
|
### `google_books_volume_details`
|
||||||
|
|
||||||
|
Get detailed information about a specific book volume
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `apiKey` | string | Yes | Google Books API key |
|
||||||
|
| `volumeId` | string | Yes | The ID of the volume to retrieve |
|
||||||
|
| `projection` | string | No | Projection level \(full, lite\) |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `id` | string | Volume ID |
|
||||||
|
| `title` | string | Book title |
|
||||||
|
| `subtitle` | string | Book subtitle |
|
||||||
|
| `authors` | array | List of authors |
|
||||||
|
| `publisher` | string | Publisher name |
|
||||||
|
| `publishedDate` | string | Publication date |
|
||||||
|
| `description` | string | Book description |
|
||||||
|
| `pageCount` | number | Number of pages |
|
||||||
|
| `categories` | array | Book categories |
|
||||||
|
| `averageRating` | number | Average rating \(1-5\) |
|
||||||
|
| `ratingsCount` | number | Number of ratings |
|
||||||
|
| `language` | string | Language code |
|
||||||
|
| `previewLink` | string | Link to preview on Google Books |
|
||||||
|
| `infoLink` | string | Link to info page |
|
||||||
|
| `thumbnailUrl` | string | Book cover thumbnail URL |
|
||||||
|
| `isbn10` | string | ISBN-10 identifier |
|
||||||
|
| `isbn13` | string | ISBN-13 identifier |
|
||||||
|
|
||||||
|
|
||||||
@@ -33,6 +33,7 @@
|
|||||||
"github",
|
"github",
|
||||||
"gitlab",
|
"gitlab",
|
||||||
"gmail",
|
"gmail",
|
||||||
|
"google_books",
|
||||||
"google_calendar",
|
"google_calendar",
|
||||||
"google_docs",
|
"google_docs",
|
||||||
"google_drive",
|
"google_drive",
|
||||||
|
|||||||
@@ -42,9 +42,6 @@ Estos atajos cambian entre las pestañas del panel en el lado derecho del lienzo
|
|||||||
|
|
||||||
| Atajo | Acción |
|
| Atajo | Acción |
|
||||||
|----------|--------|
|
|----------|--------|
|
||||||
| `C` | Enfocar pestaña Copilot |
|
|
||||||
| `T` | Enfocar pestaña Barra de herramientas |
|
|
||||||
| `E` | Enfocar pestaña Editor |
|
|
||||||
| `Mod` + `F` | Enfocar búsqueda de Barra de herramientas |
|
| `Mod` + `F` | Enfocar búsqueda de Barra de herramientas |
|
||||||
|
|
||||||
## Navegación global
|
## Navegación global
|
||||||
|
|||||||
@@ -42,9 +42,6 @@ Ces raccourcis permettent de basculer entre les onglets du panneau sur le côté
|
|||||||
|
|
||||||
| Raccourci | Action |
|
| Raccourci | Action |
|
||||||
|----------|--------|
|
|----------|--------|
|
||||||
| `C` | Activer l'onglet Copilot |
|
|
||||||
| `T` | Activer l'onglet Barre d'outils |
|
|
||||||
| `E` | Activer l'onglet Éditeur |
|
|
||||||
| `Mod` + `F` | Activer la recherche dans la barre d'outils |
|
| `Mod` + `F` | Activer la recherche dans la barre d'outils |
|
||||||
|
|
||||||
## Navigation globale
|
## Navigation globale
|
||||||
|
|||||||
@@ -41,9 +41,6 @@ import { Callout } from 'fumadocs-ui/components/callout'
|
|||||||
|
|
||||||
| ショートカット | 操作 |
|
| ショートカット | 操作 |
|
||||||
|----------|--------|
|
|----------|--------|
|
||||||
| `C` | Copilotタブにフォーカス |
|
|
||||||
| `T` | Toolbarタブにフォーカス |
|
|
||||||
| `E` | Editorタブにフォーカス |
|
|
||||||
| `Mod` + `F` | Toolbar検索にフォーカス |
|
| `Mod` + `F` | Toolbar検索にフォーカス |
|
||||||
|
|
||||||
## グローバルナビゲーション
|
## グローバルナビゲーション
|
||||||
|
|||||||
@@ -41,9 +41,6 @@ import { Callout } from 'fumadocs-ui/components/callout'
|
|||||||
|
|
||||||
| 快捷键 | 操作 |
|
| 快捷键 | 操作 |
|
||||||
|----------|--------|
|
|----------|--------|
|
||||||
| `C` | 聚焦 Copilot 标签页 |
|
|
||||||
| `T` | 聚焦 Toolbar 标签页 |
|
|
||||||
| `E` | 聚焦 Editor 标签页 |
|
|
||||||
| `Mod` + `F` | 聚焦 Toolbar 搜索 |
|
| `Mod` + `F` | 聚焦 Toolbar 搜索 |
|
||||||
|
|
||||||
## 全局导航
|
## 全局导航
|
||||||
|
|||||||
@@ -13,6 +13,7 @@ BETTER_AUTH_URL=http://localhost:3000
|
|||||||
|
|
||||||
# NextJS (Required)
|
# NextJS (Required)
|
||||||
NEXT_PUBLIC_APP_URL=http://localhost:3000
|
NEXT_PUBLIC_APP_URL=http://localhost:3000
|
||||||
|
# INTERNAL_API_BASE_URL=http://sim-app.default.svc.cluster.local:3000 # Optional: internal URL for server-side /api self-calls; defaults to NEXT_PUBLIC_APP_URL
|
||||||
|
|
||||||
# Security (Required)
|
# Security (Required)
|
||||||
ENCRYPTION_KEY=your_encryption_key # Use `openssl rand -hex 32` to generate, used to encrypt environment variables
|
ENCRYPTION_KEY=your_encryption_key # Use `openssl rand -hex 32` to generate, used to encrypt environment variables
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
import type { Artifact, Message, PushNotificationConfig, Task, TaskState } from '@a2a-js/sdk'
|
import type { Artifact, Message, PushNotificationConfig, Task, TaskState } from '@a2a-js/sdk'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
import { generateInternalToken } from '@/lib/auth/internal'
|
import { generateInternalToken } from '@/lib/auth/internal'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
|
|
||||||
/** A2A v0.3 JSON-RPC method names */
|
/** A2A v0.3 JSON-RPC method names */
|
||||||
export const A2A_METHODS = {
|
export const A2A_METHODS = {
|
||||||
@@ -118,7 +118,7 @@ export interface ExecuteRequestResult {
|
|||||||
export async function buildExecuteRequest(
|
export async function buildExecuteRequest(
|
||||||
config: ExecuteRequestConfig
|
config: ExecuteRequestConfig
|
||||||
): Promise<ExecuteRequestResult> {
|
): Promise<ExecuteRequestResult> {
|
||||||
const url = `${getBaseUrl()}/api/workflows/${config.workflowId}/execute`
|
const url = `${getInternalApiBaseUrl()}/api/workflows/${config.workflowId}/execute`
|
||||||
const headers: Record<string, string> = { 'Content-Type': 'application/json' }
|
const headers: Record<string, string> = { 'Content-Type': 'application/json' }
|
||||||
let useInternalAuth = false
|
let useInternalAuth = false
|
||||||
|
|
||||||
|
|||||||
187
apps/sim/app/api/attribution/route.ts
Normal file
187
apps/sim/app/api/attribution/route.ts
Normal file
@@ -0,0 +1,187 @@
|
|||||||
|
/**
|
||||||
|
* POST /api/attribution
|
||||||
|
*
|
||||||
|
* Automatic UTM-based referral attribution.
|
||||||
|
*
|
||||||
|
* Reads the `sim_utm` cookie (set by proxy on auth pages), matches a campaign
|
||||||
|
* by UTM specificity, and atomically inserts an attribution record + applies
|
||||||
|
* bonus credits.
|
||||||
|
*
|
||||||
|
* Idempotent — the unique constraint on `userId` prevents double-attribution.
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { db } from '@sim/db'
|
||||||
|
import { referralAttribution, referralCampaigns, userStats } from '@sim/db/schema'
|
||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { eq } from 'drizzle-orm'
|
||||||
|
import { nanoid } from 'nanoid'
|
||||||
|
import { cookies } from 'next/headers'
|
||||||
|
import { NextResponse } from 'next/server'
|
||||||
|
import { z } from 'zod'
|
||||||
|
import { getSession } from '@/lib/auth'
|
||||||
|
import { applyBonusCredits } from '@/lib/billing/credits/bonus'
|
||||||
|
|
||||||
|
const logger = createLogger('AttributionAPI')
|
||||||
|
|
||||||
|
const COOKIE_NAME = 'sim_utm'
|
||||||
|
|
||||||
|
const UtmCookieSchema = z.object({
|
||||||
|
utm_source: z.string().optional(),
|
||||||
|
utm_medium: z.string().optional(),
|
||||||
|
utm_campaign: z.string().optional(),
|
||||||
|
utm_content: z.string().optional(),
|
||||||
|
referrer_url: z.string().optional(),
|
||||||
|
landing_page: z.string().optional(),
|
||||||
|
created_at: z.string().optional(),
|
||||||
|
})
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Finds the most specific active campaign matching the given UTM params.
|
||||||
|
* Null fields on a campaign act as wildcards. Ties broken by newest campaign.
|
||||||
|
*/
|
||||||
|
async function findMatchingCampaign(utmData: z.infer<typeof UtmCookieSchema>) {
|
||||||
|
const campaigns = await db
|
||||||
|
.select()
|
||||||
|
.from(referralCampaigns)
|
||||||
|
.where(eq(referralCampaigns.isActive, true))
|
||||||
|
|
||||||
|
let bestMatch: (typeof campaigns)[number] | null = null
|
||||||
|
let bestScore = -1
|
||||||
|
|
||||||
|
for (const campaign of campaigns) {
|
||||||
|
let score = 0
|
||||||
|
let mismatch = false
|
||||||
|
|
||||||
|
const fields = [
|
||||||
|
{ campaignVal: campaign.utmSource, utmVal: utmData.utm_source },
|
||||||
|
{ campaignVal: campaign.utmMedium, utmVal: utmData.utm_medium },
|
||||||
|
{ campaignVal: campaign.utmCampaign, utmVal: utmData.utm_campaign },
|
||||||
|
{ campaignVal: campaign.utmContent, utmVal: utmData.utm_content },
|
||||||
|
] as const
|
||||||
|
|
||||||
|
for (const { campaignVal, utmVal } of fields) {
|
||||||
|
if (campaignVal === null) continue
|
||||||
|
if (campaignVal === utmVal) {
|
||||||
|
score++
|
||||||
|
} else {
|
||||||
|
mismatch = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!mismatch && score > 0) {
|
||||||
|
if (
|
||||||
|
score > bestScore ||
|
||||||
|
(score === bestScore &&
|
||||||
|
bestMatch &&
|
||||||
|
campaign.createdAt.getTime() > bestMatch.createdAt.getTime())
|
||||||
|
) {
|
||||||
|
bestScore = score
|
||||||
|
bestMatch = campaign
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return bestMatch
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function POST() {
|
||||||
|
try {
|
||||||
|
const session = await getSession()
|
||||||
|
if (!session?.user?.id) {
|
||||||
|
return NextResponse.json({ error: 'Unauthorized' }, { status: 401 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const cookieStore = await cookies()
|
||||||
|
const utmCookie = cookieStore.get(COOKIE_NAME)
|
||||||
|
if (!utmCookie?.value) {
|
||||||
|
return NextResponse.json({ attributed: false, reason: 'no_utm_cookie' })
|
||||||
|
}
|
||||||
|
|
||||||
|
let utmData: z.infer<typeof UtmCookieSchema>
|
||||||
|
try {
|
||||||
|
let decoded: string
|
||||||
|
try {
|
||||||
|
decoded = decodeURIComponent(utmCookie.value)
|
||||||
|
} catch {
|
||||||
|
decoded = utmCookie.value
|
||||||
|
}
|
||||||
|
utmData = UtmCookieSchema.parse(JSON.parse(decoded))
|
||||||
|
} catch {
|
||||||
|
logger.warn('Failed to parse UTM cookie', { userId: session.user.id })
|
||||||
|
cookieStore.delete(COOKIE_NAME)
|
||||||
|
return NextResponse.json({ attributed: false, reason: 'invalid_cookie' })
|
||||||
|
}
|
||||||
|
|
||||||
|
const matchedCampaign = await findMatchingCampaign(utmData)
|
||||||
|
if (!matchedCampaign) {
|
||||||
|
cookieStore.delete(COOKIE_NAME)
|
||||||
|
return NextResponse.json({ attributed: false, reason: 'no_matching_campaign' })
|
||||||
|
}
|
||||||
|
|
||||||
|
const bonusAmount = Number(matchedCampaign.bonusCreditAmount)
|
||||||
|
|
||||||
|
let attributed = false
|
||||||
|
await db.transaction(async (tx) => {
|
||||||
|
const [existingStats] = await tx
|
||||||
|
.select({ id: userStats.id })
|
||||||
|
.from(userStats)
|
||||||
|
.where(eq(userStats.userId, session.user.id))
|
||||||
|
.limit(1)
|
||||||
|
|
||||||
|
if (!existingStats) {
|
||||||
|
await tx.insert(userStats).values({
|
||||||
|
id: nanoid(),
|
||||||
|
userId: session.user.id,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
const result = await tx
|
||||||
|
.insert(referralAttribution)
|
||||||
|
.values({
|
||||||
|
id: nanoid(),
|
||||||
|
userId: session.user.id,
|
||||||
|
campaignId: matchedCampaign.id,
|
||||||
|
utmSource: utmData.utm_source || null,
|
||||||
|
utmMedium: utmData.utm_medium || null,
|
||||||
|
utmCampaign: utmData.utm_campaign || null,
|
||||||
|
utmContent: utmData.utm_content || null,
|
||||||
|
referrerUrl: utmData.referrer_url || null,
|
||||||
|
landingPage: utmData.landing_page || null,
|
||||||
|
bonusCreditAmount: bonusAmount.toString(),
|
||||||
|
})
|
||||||
|
.onConflictDoNothing({ target: referralAttribution.userId })
|
||||||
|
.returning({ id: referralAttribution.id })
|
||||||
|
|
||||||
|
if (result.length > 0) {
|
||||||
|
await applyBonusCredits(session.user.id, bonusAmount, tx)
|
||||||
|
attributed = true
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
if (attributed) {
|
||||||
|
logger.info('Referral attribution created and bonus credits applied', {
|
||||||
|
userId: session.user.id,
|
||||||
|
campaignId: matchedCampaign.id,
|
||||||
|
campaignName: matchedCampaign.name,
|
||||||
|
utmSource: utmData.utm_source,
|
||||||
|
utmCampaign: utmData.utm_campaign,
|
||||||
|
utmContent: utmData.utm_content,
|
||||||
|
bonusAmount,
|
||||||
|
})
|
||||||
|
} else {
|
||||||
|
logger.info('User already attributed, skipping', { userId: session.user.id })
|
||||||
|
}
|
||||||
|
|
||||||
|
cookieStore.delete(COOKIE_NAME)
|
||||||
|
|
||||||
|
return NextResponse.json({
|
||||||
|
attributed,
|
||||||
|
bonusAmount: attributed ? bonusAmount : undefined,
|
||||||
|
reason: attributed ? undefined : 'already_attributed',
|
||||||
|
})
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Attribution error', { error })
|
||||||
|
return NextResponse.json({ error: 'Internal server error' }, { status: 500 })
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -4,20 +4,10 @@
|
|||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { loggerMock } from '@sim/testing'
|
import { databaseMock, loggerMock } from '@sim/testing'
|
||||||
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
||||||
|
|
||||||
vi.mock('@sim/db', () => ({
|
vi.mock('@sim/db', () => databaseMock)
|
||||||
db: {
|
|
||||||
select: vi.fn().mockReturnThis(),
|
|
||||||
from: vi.fn().mockReturnThis(),
|
|
||||||
where: vi.fn().mockReturnThis(),
|
|
||||||
limit: vi.fn().mockReturnValue([]),
|
|
||||||
update: vi.fn().mockReturnThis(),
|
|
||||||
set: vi.fn().mockReturnThis(),
|
|
||||||
orderBy: vi.fn().mockReturnThis(),
|
|
||||||
},
|
|
||||||
}))
|
|
||||||
|
|
||||||
vi.mock('@/lib/oauth/oauth', () => ({
|
vi.mock('@/lib/oauth/oauth', () => ({
|
||||||
refreshOAuthToken: vi.fn(),
|
refreshOAuthToken: vi.fn(),
|
||||||
@@ -34,13 +24,36 @@ import {
|
|||||||
refreshTokenIfNeeded,
|
refreshTokenIfNeeded,
|
||||||
} from '@/app/api/auth/oauth/utils'
|
} from '@/app/api/auth/oauth/utils'
|
||||||
|
|
||||||
const mockDbTyped = db as any
|
const mockDb = db as any
|
||||||
const mockRefreshOAuthToken = refreshOAuthToken as any
|
const mockRefreshOAuthToken = refreshOAuthToken as any
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a chainable mock for db.select() calls.
|
||||||
|
* Returns a nested chain: select() -> from() -> where() -> limit() / orderBy()
|
||||||
|
*/
|
||||||
|
function mockSelectChain(limitResult: unknown[]) {
|
||||||
|
const mockLimit = vi.fn().mockReturnValue(limitResult)
|
||||||
|
const mockOrderBy = vi.fn().mockReturnValue(limitResult)
|
||||||
|
const mockWhere = vi.fn().mockReturnValue({ limit: mockLimit, orderBy: mockOrderBy })
|
||||||
|
const mockFrom = vi.fn().mockReturnValue({ where: mockWhere })
|
||||||
|
mockDb.select.mockReturnValueOnce({ from: mockFrom })
|
||||||
|
return { mockFrom, mockWhere, mockLimit }
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a chainable mock for db.update() calls.
|
||||||
|
* Returns a nested chain: update() -> set() -> where()
|
||||||
|
*/
|
||||||
|
function mockUpdateChain() {
|
||||||
|
const mockWhere = vi.fn().mockResolvedValue({})
|
||||||
|
const mockSet = vi.fn().mockReturnValue({ where: mockWhere })
|
||||||
|
mockDb.update.mockReturnValueOnce({ set: mockSet })
|
||||||
|
return { mockSet, mockWhere }
|
||||||
|
}
|
||||||
|
|
||||||
describe('OAuth Utils', () => {
|
describe('OAuth Utils', () => {
|
||||||
beforeEach(() => {
|
beforeEach(() => {
|
||||||
vi.clearAllMocks()
|
vi.clearAllMocks()
|
||||||
mockDbTyped.limit.mockReturnValue([])
|
|
||||||
})
|
})
|
||||||
|
|
||||||
afterEach(() => {
|
afterEach(() => {
|
||||||
@@ -50,20 +63,20 @@ describe('OAuth Utils', () => {
|
|||||||
describe('getCredential', () => {
|
describe('getCredential', () => {
|
||||||
it('should return credential when found', async () => {
|
it('should return credential when found', async () => {
|
||||||
const mockCredential = { id: 'credential-id', userId: 'test-user-id' }
|
const mockCredential = { id: 'credential-id', userId: 'test-user-id' }
|
||||||
mockDbTyped.limit.mockReturnValueOnce([mockCredential])
|
const { mockFrom, mockWhere, mockLimit } = mockSelectChain([mockCredential])
|
||||||
|
|
||||||
const credential = await getCredential('request-id', 'credential-id', 'test-user-id')
|
const credential = await getCredential('request-id', 'credential-id', 'test-user-id')
|
||||||
|
|
||||||
expect(mockDbTyped.select).toHaveBeenCalled()
|
expect(mockDb.select).toHaveBeenCalled()
|
||||||
expect(mockDbTyped.from).toHaveBeenCalled()
|
expect(mockFrom).toHaveBeenCalled()
|
||||||
expect(mockDbTyped.where).toHaveBeenCalled()
|
expect(mockWhere).toHaveBeenCalled()
|
||||||
expect(mockDbTyped.limit).toHaveBeenCalledWith(1)
|
expect(mockLimit).toHaveBeenCalledWith(1)
|
||||||
|
|
||||||
expect(credential).toEqual(mockCredential)
|
expect(credential).toEqual(mockCredential)
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should return undefined when credential is not found', async () => {
|
it('should return undefined when credential is not found', async () => {
|
||||||
mockDbTyped.limit.mockReturnValueOnce([])
|
mockSelectChain([])
|
||||||
|
|
||||||
const credential = await getCredential('request-id', 'nonexistent-id', 'test-user-id')
|
const credential = await getCredential('request-id', 'nonexistent-id', 'test-user-id')
|
||||||
|
|
||||||
@@ -102,11 +115,12 @@ describe('OAuth Utils', () => {
|
|||||||
refreshToken: 'new-refresh-token',
|
refreshToken: 'new-refresh-token',
|
||||||
})
|
})
|
||||||
|
|
||||||
|
mockUpdateChain()
|
||||||
|
|
||||||
const result = await refreshTokenIfNeeded('request-id', mockCredential, 'credential-id')
|
const result = await refreshTokenIfNeeded('request-id', mockCredential, 'credential-id')
|
||||||
|
|
||||||
expect(mockRefreshOAuthToken).toHaveBeenCalledWith('google', 'refresh-token')
|
expect(mockRefreshOAuthToken).toHaveBeenCalledWith('google', 'refresh-token')
|
||||||
expect(mockDbTyped.update).toHaveBeenCalled()
|
expect(mockDb.update).toHaveBeenCalled()
|
||||||
expect(mockDbTyped.set).toHaveBeenCalled()
|
|
||||||
expect(result).toEqual({ accessToken: 'new-token', refreshed: true })
|
expect(result).toEqual({ accessToken: 'new-token', refreshed: true })
|
||||||
})
|
})
|
||||||
|
|
||||||
@@ -152,7 +166,7 @@ describe('OAuth Utils', () => {
|
|||||||
providerId: 'google',
|
providerId: 'google',
|
||||||
userId: 'test-user-id',
|
userId: 'test-user-id',
|
||||||
}
|
}
|
||||||
mockDbTyped.limit.mockReturnValueOnce([mockCredential])
|
mockSelectChain([mockCredential])
|
||||||
|
|
||||||
const token = await refreshAccessTokenIfNeeded('credential-id', 'test-user-id', 'request-id')
|
const token = await refreshAccessTokenIfNeeded('credential-id', 'test-user-id', 'request-id')
|
||||||
|
|
||||||
@@ -169,7 +183,8 @@ describe('OAuth Utils', () => {
|
|||||||
providerId: 'google',
|
providerId: 'google',
|
||||||
userId: 'test-user-id',
|
userId: 'test-user-id',
|
||||||
}
|
}
|
||||||
mockDbTyped.limit.mockReturnValueOnce([mockCredential])
|
mockSelectChain([mockCredential])
|
||||||
|
mockUpdateChain()
|
||||||
|
|
||||||
mockRefreshOAuthToken.mockResolvedValueOnce({
|
mockRefreshOAuthToken.mockResolvedValueOnce({
|
||||||
accessToken: 'new-token',
|
accessToken: 'new-token',
|
||||||
@@ -180,13 +195,12 @@ describe('OAuth Utils', () => {
|
|||||||
const token = await refreshAccessTokenIfNeeded('credential-id', 'test-user-id', 'request-id')
|
const token = await refreshAccessTokenIfNeeded('credential-id', 'test-user-id', 'request-id')
|
||||||
|
|
||||||
expect(mockRefreshOAuthToken).toHaveBeenCalledWith('google', 'refresh-token')
|
expect(mockRefreshOAuthToken).toHaveBeenCalledWith('google', 'refresh-token')
|
||||||
expect(mockDbTyped.update).toHaveBeenCalled()
|
expect(mockDb.update).toHaveBeenCalled()
|
||||||
expect(mockDbTyped.set).toHaveBeenCalled()
|
|
||||||
expect(token).toBe('new-token')
|
expect(token).toBe('new-token')
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should return null if credential not found', async () => {
|
it('should return null if credential not found', async () => {
|
||||||
mockDbTyped.limit.mockReturnValueOnce([])
|
mockSelectChain([])
|
||||||
|
|
||||||
const token = await refreshAccessTokenIfNeeded('nonexistent-id', 'test-user-id', 'request-id')
|
const token = await refreshAccessTokenIfNeeded('nonexistent-id', 'test-user-id', 'request-id')
|
||||||
|
|
||||||
@@ -202,7 +216,7 @@ describe('OAuth Utils', () => {
|
|||||||
providerId: 'google',
|
providerId: 'google',
|
||||||
userId: 'test-user-id',
|
userId: 'test-user-id',
|
||||||
}
|
}
|
||||||
mockDbTyped.limit.mockReturnValueOnce([mockCredential])
|
mockSelectChain([mockCredential])
|
||||||
|
|
||||||
mockRefreshOAuthToken.mockResolvedValueOnce(null)
|
mockRefreshOAuthToken.mockResolvedValueOnce(null)
|
||||||
|
|
||||||
|
|||||||
@@ -85,7 +85,7 @@ const ChatMessageSchema = z.object({
|
|||||||
chatId: z.string().optional(),
|
chatId: z.string().optional(),
|
||||||
workflowId: z.string().optional(),
|
workflowId: z.string().optional(),
|
||||||
workflowName: z.string().optional(),
|
workflowName: z.string().optional(),
|
||||||
model: z.string().optional().default('claude-opus-4-6'),
|
model: z.string().optional().default('claude-opus-4-5'),
|
||||||
mode: z.enum(COPILOT_REQUEST_MODES).optional().default('agent'),
|
mode: z.enum(COPILOT_REQUEST_MODES).optional().default('agent'),
|
||||||
prefetch: z.boolean().optional(),
|
prefetch: z.boolean().optional(),
|
||||||
createNewChat: z.boolean().optional().default(false),
|
createNewChat: z.boolean().optional().default(false),
|
||||||
@@ -113,6 +113,7 @@ const ChatMessageSchema = z.object({
|
|||||||
workflowId: z.string().optional(),
|
workflowId: z.string().optional(),
|
||||||
knowledgeId: z.string().optional(),
|
knowledgeId: z.string().optional(),
|
||||||
blockId: z.string().optional(),
|
blockId: z.string().optional(),
|
||||||
|
blockIds: z.array(z.string()).optional(),
|
||||||
templateId: z.string().optional(),
|
templateId: z.string().optional(),
|
||||||
executionId: z.string().optional(),
|
executionId: z.string().optional(),
|
||||||
// For workflow_block, provide both workflowId and blockId
|
// For workflow_block, provide both workflowId and blockId
|
||||||
@@ -159,6 +160,20 @@ export async function POST(req: NextRequest) {
|
|||||||
commands,
|
commands,
|
||||||
} = ChatMessageSchema.parse(body)
|
} = ChatMessageSchema.parse(body)
|
||||||
|
|
||||||
|
const normalizedContexts = Array.isArray(contexts)
|
||||||
|
? contexts.map((ctx) => {
|
||||||
|
if (ctx.kind !== 'blocks') return ctx
|
||||||
|
if (Array.isArray(ctx.blockIds) && ctx.blockIds.length > 0) return ctx
|
||||||
|
if (ctx.blockId) {
|
||||||
|
return {
|
||||||
|
...ctx,
|
||||||
|
blockIds: [ctx.blockId],
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return ctx
|
||||||
|
})
|
||||||
|
: contexts
|
||||||
|
|
||||||
// Resolve workflowId - if not provided, use first workflow or find by name
|
// Resolve workflowId - if not provided, use first workflow or find by name
|
||||||
const resolved = await resolveWorkflowIdForUser(
|
const resolved = await resolveWorkflowIdForUser(
|
||||||
authenticatedUserId,
|
authenticatedUserId,
|
||||||
@@ -176,10 +191,10 @@ export async function POST(req: NextRequest) {
|
|||||||
const userMessageIdToUse = userMessageId || crypto.randomUUID()
|
const userMessageIdToUse = userMessageId || crypto.randomUUID()
|
||||||
try {
|
try {
|
||||||
logger.info(`[${tracker.requestId}] Received chat POST`, {
|
logger.info(`[${tracker.requestId}] Received chat POST`, {
|
||||||
hasContexts: Array.isArray(contexts),
|
hasContexts: Array.isArray(normalizedContexts),
|
||||||
contextsCount: Array.isArray(contexts) ? contexts.length : 0,
|
contextsCount: Array.isArray(normalizedContexts) ? normalizedContexts.length : 0,
|
||||||
contextsPreview: Array.isArray(contexts)
|
contextsPreview: Array.isArray(normalizedContexts)
|
||||||
? contexts.map((c: any) => ({
|
? normalizedContexts.map((c: any) => ({
|
||||||
kind: c?.kind,
|
kind: c?.kind,
|
||||||
chatId: c?.chatId,
|
chatId: c?.chatId,
|
||||||
workflowId: c?.workflowId,
|
workflowId: c?.workflowId,
|
||||||
@@ -191,17 +206,25 @@ export async function POST(req: NextRequest) {
|
|||||||
} catch {}
|
} catch {}
|
||||||
// Preprocess contexts server-side
|
// Preprocess contexts server-side
|
||||||
let agentContexts: Array<{ type: string; content: string }> = []
|
let agentContexts: Array<{ type: string; content: string }> = []
|
||||||
if (Array.isArray(contexts) && contexts.length > 0) {
|
if (Array.isArray(normalizedContexts) && normalizedContexts.length > 0) {
|
||||||
try {
|
try {
|
||||||
const { processContextsServer } = await import('@/lib/copilot/process-contents')
|
const { processContextsServer } = await import('@/lib/copilot/process-contents')
|
||||||
const processed = await processContextsServer(contexts as any, authenticatedUserId, message)
|
const processed = await processContextsServer(
|
||||||
|
normalizedContexts as any,
|
||||||
|
authenticatedUserId,
|
||||||
|
message
|
||||||
|
)
|
||||||
agentContexts = processed
|
agentContexts = processed
|
||||||
logger.info(`[${tracker.requestId}] Contexts processed for request`, {
|
logger.info(`[${tracker.requestId}] Contexts processed for request`, {
|
||||||
processedCount: agentContexts.length,
|
processedCount: agentContexts.length,
|
||||||
kinds: agentContexts.map((c) => c.type),
|
kinds: agentContexts.map((c) => c.type),
|
||||||
lengthPreview: agentContexts.map((c) => c.content?.length ?? 0),
|
lengthPreview: agentContexts.map((c) => c.content?.length ?? 0),
|
||||||
})
|
})
|
||||||
if (Array.isArray(contexts) && contexts.length > 0 && agentContexts.length === 0) {
|
if (
|
||||||
|
Array.isArray(normalizedContexts) &&
|
||||||
|
normalizedContexts.length > 0 &&
|
||||||
|
agentContexts.length === 0
|
||||||
|
) {
|
||||||
logger.warn(
|
logger.warn(
|
||||||
`[${tracker.requestId}] Contexts provided but none processed. Check executionId for logs contexts.`
|
`[${tracker.requestId}] Contexts provided but none processed. Check executionId for logs contexts.`
|
||||||
)
|
)
|
||||||
@@ -215,7 +238,7 @@ export async function POST(req: NextRequest) {
|
|||||||
let currentChat: any = null
|
let currentChat: any = null
|
||||||
let conversationHistory: any[] = []
|
let conversationHistory: any[] = []
|
||||||
let actualChatId = chatId
|
let actualChatId = chatId
|
||||||
const selectedModel = model || 'claude-opus-4-6'
|
const selectedModel = model || 'claude-opus-4-5'
|
||||||
|
|
||||||
if (chatId || createNewChat) {
|
if (chatId || createNewChat) {
|
||||||
const chatResult = await resolveOrCreateChat({
|
const chatResult = await resolveOrCreateChat({
|
||||||
@@ -246,11 +269,13 @@ export async function POST(req: NextRequest) {
|
|||||||
mode,
|
mode,
|
||||||
model: selectedModel,
|
model: selectedModel,
|
||||||
provider,
|
provider,
|
||||||
|
conversationId: effectiveConversationId,
|
||||||
conversationHistory,
|
conversationHistory,
|
||||||
contexts: agentContexts,
|
contexts: agentContexts,
|
||||||
fileAttachments,
|
fileAttachments,
|
||||||
commands,
|
commands,
|
||||||
chatId: actualChatId,
|
chatId: actualChatId,
|
||||||
|
prefetch,
|
||||||
implicitFeedback,
|
implicitFeedback,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -432,10 +457,15 @@ export async function POST(req: NextRequest) {
|
|||||||
content: message,
|
content: message,
|
||||||
timestamp: new Date().toISOString(),
|
timestamp: new Date().toISOString(),
|
||||||
...(fileAttachments && fileAttachments.length > 0 && { fileAttachments }),
|
...(fileAttachments && fileAttachments.length > 0 && { fileAttachments }),
|
||||||
...(Array.isArray(contexts) && contexts.length > 0 && { contexts }),
|
...(Array.isArray(normalizedContexts) &&
|
||||||
...(Array.isArray(contexts) &&
|
normalizedContexts.length > 0 && {
|
||||||
contexts.length > 0 && {
|
contexts: normalizedContexts,
|
||||||
contentBlocks: [{ type: 'contexts', contexts: contexts as any, timestamp: Date.now() }],
|
}),
|
||||||
|
...(Array.isArray(normalizedContexts) &&
|
||||||
|
normalizedContexts.length > 0 && {
|
||||||
|
contentBlocks: [
|
||||||
|
{ type: 'contexts', contexts: normalizedContexts as any, timestamp: Date.now() },
|
||||||
|
],
|
||||||
}),
|
}),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -18,9 +18,9 @@ describe('Copilot Checkpoints Revert API Route', () => {
|
|||||||
setupCommonApiMocks()
|
setupCommonApiMocks()
|
||||||
mockCryptoUuid()
|
mockCryptoUuid()
|
||||||
|
|
||||||
// Mock getBaseUrl to return localhost for tests
|
|
||||||
vi.doMock('@/lib/core/utils/urls', () => ({
|
vi.doMock('@/lib/core/utils/urls', () => ({
|
||||||
getBaseUrl: vi.fn(() => 'http://localhost:3000'),
|
getBaseUrl: vi.fn(() => 'http://localhost:3000'),
|
||||||
|
getInternalApiBaseUrl: vi.fn(() => 'http://localhost:3000'),
|
||||||
getBaseDomain: vi.fn(() => 'localhost:3000'),
|
getBaseDomain: vi.fn(() => 'localhost:3000'),
|
||||||
getEmailDomain: vi.fn(() => 'localhost:3000'),
|
getEmailDomain: vi.fn(() => 'localhost:3000'),
|
||||||
}))
|
}))
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ import {
|
|||||||
createRequestTracker,
|
createRequestTracker,
|
||||||
createUnauthorizedResponse,
|
createUnauthorizedResponse,
|
||||||
} from '@/lib/copilot/request-helpers'
|
} from '@/lib/copilot/request-helpers'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { authorizeWorkflowByWorkspacePermission } from '@/lib/workflows/utils'
|
import { authorizeWorkflowByWorkspacePermission } from '@/lib/workflows/utils'
|
||||||
import { isUuidV4 } from '@/executor/constants'
|
import { isUuidV4 } from '@/executor/constants'
|
||||||
|
|
||||||
@@ -99,7 +99,7 @@ export async function POST(request: NextRequest) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const stateResponse = await fetch(
|
const stateResponse = await fetch(
|
||||||
`${getBaseUrl()}/api/workflows/${checkpoint.workflowId}/state`,
|
`${getInternalApiBaseUrl()}/api/workflows/${checkpoint.workflowId}/state`,
|
||||||
{
|
{
|
||||||
method: 'PUT',
|
method: 'PUT',
|
||||||
headers: {
|
headers: {
|
||||||
|
|||||||
@@ -4,16 +4,12 @@
|
|||||||
*
|
*
|
||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
import { createEnvMock, createMockLogger } from '@sim/testing'
|
import { createEnvMock, databaseMock, loggerMock } from '@sim/testing'
|
||||||
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
||||||
|
|
||||||
const loggerMock = vi.hoisted(() => ({
|
|
||||||
createLogger: () => createMockLogger(),
|
|
||||||
}))
|
|
||||||
|
|
||||||
vi.mock('drizzle-orm')
|
vi.mock('drizzle-orm')
|
||||||
vi.mock('@sim/logger', () => loggerMock)
|
vi.mock('@sim/logger', () => loggerMock)
|
||||||
vi.mock('@sim/db')
|
vi.mock('@sim/db', () => databaseMock)
|
||||||
vi.mock('@/lib/knowledge/documents/utils', () => ({
|
vi.mock('@/lib/knowledge/documents/utils', () => ({
|
||||||
retryWithExponentialBackoff: (fn: any) => fn(),
|
retryWithExponentialBackoff: (fn: any) => fn(),
|
||||||
}))
|
}))
|
||||||
|
|||||||
@@ -38,7 +38,7 @@ import {
|
|||||||
|
|
||||||
const logger = createLogger('CopilotMcpAPI')
|
const logger = createLogger('CopilotMcpAPI')
|
||||||
const mcpRateLimiter = new RateLimiter()
|
const mcpRateLimiter = new RateLimiter()
|
||||||
const DEFAULT_COPILOT_MODEL = 'claude-opus-4-6'
|
const DEFAULT_COPILOT_MODEL = 'claude-opus-4-5'
|
||||||
|
|
||||||
export const dynamic = 'force-dynamic'
|
export const dynamic = 'force-dynamic'
|
||||||
export const runtime = 'nodejs'
|
export const runtime = 'nodejs'
|
||||||
|
|||||||
@@ -72,6 +72,7 @@ describe('MCP Serve Route', () => {
|
|||||||
}))
|
}))
|
||||||
vi.doMock('@/lib/core/utils/urls', () => ({
|
vi.doMock('@/lib/core/utils/urls', () => ({
|
||||||
getBaseUrl: () => 'http://localhost:3000',
|
getBaseUrl: () => 'http://localhost:3000',
|
||||||
|
getInternalApiBaseUrl: () => 'http://localhost:3000',
|
||||||
}))
|
}))
|
||||||
vi.doMock('@/lib/core/execution-limits', () => ({
|
vi.doMock('@/lib/core/execution-limits', () => ({
|
||||||
getMaxExecutionTimeout: () => 10_000,
|
getMaxExecutionTimeout: () => 10_000,
|
||||||
|
|||||||
@@ -22,7 +22,7 @@ import { type NextRequest, NextResponse } from 'next/server'
|
|||||||
import { type AuthResult, checkHybridAuth } from '@/lib/auth/hybrid'
|
import { type AuthResult, checkHybridAuth } from '@/lib/auth/hybrid'
|
||||||
import { generateInternalToken } from '@/lib/auth/internal'
|
import { generateInternalToken } from '@/lib/auth/internal'
|
||||||
import { getMaxExecutionTimeout } from '@/lib/core/execution-limits'
|
import { getMaxExecutionTimeout } from '@/lib/core/execution-limits'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { getUserEntityPermissions } from '@/lib/workspaces/permissions/utils'
|
import { getUserEntityPermissions } from '@/lib/workspaces/permissions/utils'
|
||||||
|
|
||||||
const logger = createLogger('WorkflowMcpServeAPI')
|
const logger = createLogger('WorkflowMcpServeAPI')
|
||||||
@@ -285,7 +285,7 @@ async function handleToolsCall(
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
const executeUrl = `${getBaseUrl()}/api/workflows/${tool.workflowId}/execute`
|
const executeUrl = `${getInternalApiBaseUrl()}/api/workflows/${tool.workflowId}/execute`
|
||||||
const headers: Record<string, string> = { 'Content-Type': 'application/json' }
|
const headers: Record<string, string> = { 'Content-Type': 'application/json' }
|
||||||
|
|
||||||
if (publicServerOwnerId) {
|
if (publicServerOwnerId) {
|
||||||
|
|||||||
170
apps/sim/app/api/referral-code/redeem/route.ts
Normal file
170
apps/sim/app/api/referral-code/redeem/route.ts
Normal file
@@ -0,0 +1,170 @@
|
|||||||
|
/**
|
||||||
|
* POST /api/referral-code/redeem
|
||||||
|
*
|
||||||
|
* Redeem a referral/promo code to receive bonus credits.
|
||||||
|
*
|
||||||
|
* Body:
|
||||||
|
* - code: string — The referral code to redeem
|
||||||
|
*
|
||||||
|
* Response: { redeemed: boolean, bonusAmount?: number, error?: string }
|
||||||
|
*
|
||||||
|
* Constraints:
|
||||||
|
* - Enterprise users cannot redeem codes
|
||||||
|
* - One redemption per user, ever (unique constraint on userId)
|
||||||
|
* - One redemption per organization for team users (partial unique on organizationId)
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { db } from '@sim/db'
|
||||||
|
import { referralAttribution, referralCampaigns, userStats } from '@sim/db/schema'
|
||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { and, eq } from 'drizzle-orm'
|
||||||
|
import { nanoid } from 'nanoid'
|
||||||
|
import { NextResponse } from 'next/server'
|
||||||
|
import { z } from 'zod'
|
||||||
|
import { getSession } from '@/lib/auth'
|
||||||
|
import { getHighestPrioritySubscription } from '@/lib/billing/core/subscription'
|
||||||
|
import { applyBonusCredits } from '@/lib/billing/credits/bonus'
|
||||||
|
|
||||||
|
const logger = createLogger('ReferralCodeRedemption')
|
||||||
|
|
||||||
|
const RedeemCodeSchema = z.object({
|
||||||
|
code: z.string().min(1, 'Code is required'),
|
||||||
|
})
|
||||||
|
|
||||||
|
export async function POST(request: Request) {
|
||||||
|
try {
|
||||||
|
const session = await getSession()
|
||||||
|
if (!session?.user?.id) {
|
||||||
|
return NextResponse.json({ error: 'Unauthorized' }, { status: 401 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const body = await request.json()
|
||||||
|
const { code } = RedeemCodeSchema.parse(body)
|
||||||
|
|
||||||
|
const subscription = await getHighestPrioritySubscription(session.user.id)
|
||||||
|
|
||||||
|
if (subscription?.plan === 'enterprise') {
|
||||||
|
return NextResponse.json({
|
||||||
|
redeemed: false,
|
||||||
|
error: 'Enterprise accounts cannot redeem referral codes',
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
const isTeam = subscription?.plan === 'team'
|
||||||
|
const orgId = isTeam ? subscription.referenceId : null
|
||||||
|
|
||||||
|
const normalizedCode = code.trim().toUpperCase()
|
||||||
|
|
||||||
|
const [campaign] = await db
|
||||||
|
.select()
|
||||||
|
.from(referralCampaigns)
|
||||||
|
.where(and(eq(referralCampaigns.code, normalizedCode), eq(referralCampaigns.isActive, true)))
|
||||||
|
.limit(1)
|
||||||
|
|
||||||
|
if (!campaign) {
|
||||||
|
logger.info('Invalid code redemption attempt', {
|
||||||
|
userId: session.user.id,
|
||||||
|
code: normalizedCode,
|
||||||
|
})
|
||||||
|
return NextResponse.json({ error: 'Invalid or expired code' }, { status: 404 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const [existingUserAttribution] = await db
|
||||||
|
.select({ id: referralAttribution.id })
|
||||||
|
.from(referralAttribution)
|
||||||
|
.where(eq(referralAttribution.userId, session.user.id))
|
||||||
|
.limit(1)
|
||||||
|
|
||||||
|
if (existingUserAttribution) {
|
||||||
|
return NextResponse.json({
|
||||||
|
redeemed: false,
|
||||||
|
error: 'You have already redeemed a code',
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
if (orgId) {
|
||||||
|
const [existingOrgAttribution] = await db
|
||||||
|
.select({ id: referralAttribution.id })
|
||||||
|
.from(referralAttribution)
|
||||||
|
.where(eq(referralAttribution.organizationId, orgId))
|
||||||
|
.limit(1)
|
||||||
|
|
||||||
|
if (existingOrgAttribution) {
|
||||||
|
return NextResponse.json({
|
||||||
|
redeemed: false,
|
||||||
|
error: 'A code has already been redeemed for your organization',
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const bonusAmount = Number(campaign.bonusCreditAmount)
|
||||||
|
|
||||||
|
let redeemed = false
|
||||||
|
await db.transaction(async (tx) => {
|
||||||
|
const [existingStats] = await tx
|
||||||
|
.select({ id: userStats.id })
|
||||||
|
.from(userStats)
|
||||||
|
.where(eq(userStats.userId, session.user.id))
|
||||||
|
.limit(1)
|
||||||
|
|
||||||
|
if (!existingStats) {
|
||||||
|
await tx.insert(userStats).values({
|
||||||
|
id: nanoid(),
|
||||||
|
userId: session.user.id,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
const result = await tx
|
||||||
|
.insert(referralAttribution)
|
||||||
|
.values({
|
||||||
|
id: nanoid(),
|
||||||
|
userId: session.user.id,
|
||||||
|
organizationId: orgId,
|
||||||
|
campaignId: campaign.id,
|
||||||
|
utmSource: null,
|
||||||
|
utmMedium: null,
|
||||||
|
utmCampaign: null,
|
||||||
|
utmContent: null,
|
||||||
|
referrerUrl: null,
|
||||||
|
landingPage: null,
|
||||||
|
bonusCreditAmount: bonusAmount.toString(),
|
||||||
|
})
|
||||||
|
.onConflictDoNothing()
|
||||||
|
.returning({ id: referralAttribution.id })
|
||||||
|
|
||||||
|
if (result.length > 0) {
|
||||||
|
await applyBonusCredits(session.user.id, bonusAmount, tx)
|
||||||
|
redeemed = true
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
if (redeemed) {
|
||||||
|
logger.info('Referral code redeemed', {
|
||||||
|
userId: session.user.id,
|
||||||
|
organizationId: orgId,
|
||||||
|
code: normalizedCode,
|
||||||
|
campaignId: campaign.id,
|
||||||
|
campaignName: campaign.name,
|
||||||
|
bonusAmount,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!redeemed) {
|
||||||
|
return NextResponse.json({
|
||||||
|
redeemed: false,
|
||||||
|
error: 'You have already redeemed a code',
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
return NextResponse.json({
|
||||||
|
redeemed: true,
|
||||||
|
bonusAmount,
|
||||||
|
})
|
||||||
|
} catch (error) {
|
||||||
|
if (error instanceof z.ZodError) {
|
||||||
|
return NextResponse.json({ error: error.errors[0].message }, { status: 400 })
|
||||||
|
}
|
||||||
|
logger.error('Referral code redemption error', { error })
|
||||||
|
return NextResponse.json({ error: 'Internal server error' }, { status: 500 })
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -3,17 +3,14 @@
|
|||||||
*
|
*
|
||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
import { loggerMock } from '@sim/testing'
|
import { databaseMock, loggerMock } from '@sim/testing'
|
||||||
import { NextRequest } from 'next/server'
|
import { NextRequest } from 'next/server'
|
||||||
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
||||||
|
|
||||||
const { mockGetSession, mockAuthorizeWorkflowByWorkspacePermission, mockDbSelect, mockDbUpdate } =
|
const { mockGetSession, mockAuthorizeWorkflowByWorkspacePermission } = vi.hoisted(() => ({
|
||||||
vi.hoisted(() => ({
|
mockGetSession: vi.fn(),
|
||||||
mockGetSession: vi.fn(),
|
mockAuthorizeWorkflowByWorkspacePermission: vi.fn(),
|
||||||
mockAuthorizeWorkflowByWorkspacePermission: vi.fn(),
|
}))
|
||||||
mockDbSelect: vi.fn(),
|
|
||||||
mockDbUpdate: vi.fn(),
|
|
||||||
}))
|
|
||||||
|
|
||||||
vi.mock('@/lib/auth', () => ({
|
vi.mock('@/lib/auth', () => ({
|
||||||
getSession: mockGetSession,
|
getSession: mockGetSession,
|
||||||
@@ -23,12 +20,7 @@ vi.mock('@/lib/workflows/utils', () => ({
|
|||||||
authorizeWorkflowByWorkspacePermission: mockAuthorizeWorkflowByWorkspacePermission,
|
authorizeWorkflowByWorkspacePermission: mockAuthorizeWorkflowByWorkspacePermission,
|
||||||
}))
|
}))
|
||||||
|
|
||||||
vi.mock('@sim/db', () => ({
|
vi.mock('@sim/db', () => databaseMock)
|
||||||
db: {
|
|
||||||
select: mockDbSelect,
|
|
||||||
update: mockDbUpdate,
|
|
||||||
},
|
|
||||||
}))
|
|
||||||
|
|
||||||
vi.mock('@sim/db/schema', () => ({
|
vi.mock('@sim/db/schema', () => ({
|
||||||
workflow: { id: 'id', userId: 'userId', workspaceId: 'workspaceId' },
|
workflow: { id: 'id', userId: 'userId', workspaceId: 'workspaceId' },
|
||||||
@@ -59,6 +51,9 @@ function createParams(id: string): { params: Promise<{ id: string }> } {
|
|||||||
return { params: Promise.resolve({ id }) }
|
return { params: Promise.resolve({ id }) }
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const mockDbSelect = databaseMock.db.select as ReturnType<typeof vi.fn>
|
||||||
|
const mockDbUpdate = databaseMock.db.update as ReturnType<typeof vi.fn>
|
||||||
|
|
||||||
function mockDbChain(selectResults: unknown[][]) {
|
function mockDbChain(selectResults: unknown[][]) {
|
||||||
let selectCallIndex = 0
|
let selectCallIndex = 0
|
||||||
mockDbSelect.mockImplementation(() => ({
|
mockDbSelect.mockImplementation(() => ({
|
||||||
|
|||||||
@@ -3,17 +3,14 @@
|
|||||||
*
|
*
|
||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
import { loggerMock } from '@sim/testing'
|
import { databaseMock, loggerMock } from '@sim/testing'
|
||||||
import { NextRequest } from 'next/server'
|
import { NextRequest } from 'next/server'
|
||||||
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
||||||
|
|
||||||
const { mockGetSession, mockAuthorizeWorkflowByWorkspacePermission, mockDbSelect } = vi.hoisted(
|
const { mockGetSession, mockAuthorizeWorkflowByWorkspacePermission } = vi.hoisted(() => ({
|
||||||
() => ({
|
mockGetSession: vi.fn(),
|
||||||
mockGetSession: vi.fn(),
|
mockAuthorizeWorkflowByWorkspacePermission: vi.fn(),
|
||||||
mockAuthorizeWorkflowByWorkspacePermission: vi.fn(),
|
}))
|
||||||
mockDbSelect: vi.fn(),
|
|
||||||
})
|
|
||||||
)
|
|
||||||
|
|
||||||
vi.mock('@/lib/auth', () => ({
|
vi.mock('@/lib/auth', () => ({
|
||||||
getSession: mockGetSession,
|
getSession: mockGetSession,
|
||||||
@@ -23,11 +20,7 @@ vi.mock('@/lib/workflows/utils', () => ({
|
|||||||
authorizeWorkflowByWorkspacePermission: mockAuthorizeWorkflowByWorkspacePermission,
|
authorizeWorkflowByWorkspacePermission: mockAuthorizeWorkflowByWorkspacePermission,
|
||||||
}))
|
}))
|
||||||
|
|
||||||
vi.mock('@sim/db', () => ({
|
vi.mock('@sim/db', () => databaseMock)
|
||||||
db: {
|
|
||||||
select: mockDbSelect,
|
|
||||||
},
|
|
||||||
}))
|
|
||||||
|
|
||||||
vi.mock('@sim/db/schema', () => ({
|
vi.mock('@sim/db/schema', () => ({
|
||||||
workflow: { id: 'id', userId: 'userId', workspaceId: 'workspaceId' },
|
workflow: { id: 'id', userId: 'userId', workspaceId: 'workspaceId' },
|
||||||
@@ -62,6 +55,8 @@ function createRequest(url: string): NextRequest {
|
|||||||
return new NextRequest(new URL(url), { method: 'GET' })
|
return new NextRequest(new URL(url), { method: 'GET' })
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const mockDbSelect = databaseMock.db.select as ReturnType<typeof vi.fn>
|
||||||
|
|
||||||
function mockDbChain(results: any[]) {
|
function mockDbChain(results: any[]) {
|
||||||
let callIndex = 0
|
let callIndex = 0
|
||||||
mockDbSelect.mockImplementation(() => ({
|
mockDbSelect.mockImplementation(() => ({
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import { type NextRequest, NextResponse } from 'next/server'
|
|||||||
import { v4 as uuidv4 } from 'uuid'
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
import { getSession } from '@/lib/auth'
|
import { getSession } from '@/lib/auth'
|
||||||
import { generateRequestId } from '@/lib/core/utils/request'
|
import { generateRequestId } from '@/lib/core/utils/request'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import {
|
import {
|
||||||
type RegenerateStateInput,
|
type RegenerateStateInput,
|
||||||
regenerateWorkflowStateIds,
|
regenerateWorkflowStateIds,
|
||||||
@@ -115,15 +115,18 @@ export async function POST(request: NextRequest, { params }: { params: Promise<{
|
|||||||
// Step 3: Save the workflow state using the existing state endpoint (like imports do)
|
// Step 3: Save the workflow state using the existing state endpoint (like imports do)
|
||||||
// Ensure variables in state are remapped for the new workflow as well
|
// Ensure variables in state are remapped for the new workflow as well
|
||||||
const workflowStateWithVariables = { ...workflowState, variables: remappedVariables }
|
const workflowStateWithVariables = { ...workflowState, variables: remappedVariables }
|
||||||
const stateResponse = await fetch(`${getBaseUrl()}/api/workflows/${newWorkflowId}/state`, {
|
const stateResponse = await fetch(
|
||||||
method: 'PUT',
|
`${getInternalApiBaseUrl()}/api/workflows/${newWorkflowId}/state`,
|
||||||
headers: {
|
{
|
||||||
'Content-Type': 'application/json',
|
method: 'PUT',
|
||||||
// Forward the session cookie for authentication
|
headers: {
|
||||||
cookie: request.headers.get('cookie') || '',
|
'Content-Type': 'application/json',
|
||||||
},
|
// Forward the session cookie for authentication
|
||||||
body: JSON.stringify(workflowStateWithVariables),
|
cookie: request.headers.get('cookie') || '',
|
||||||
})
|
},
|
||||||
|
body: JSON.stringify(workflowStateWithVariables),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
if (!stateResponse.ok) {
|
if (!stateResponse.ok) {
|
||||||
logger.error(`[${requestId}] Failed to save workflow state for template use`)
|
logger.error(`[${requestId}] Failed to save workflow state for template use`)
|
||||||
|
|||||||
@@ -191,3 +191,84 @@ export async function GET(request: NextRequest) {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Delete a label from a page
|
||||||
|
export async function DELETE(request: NextRequest) {
|
||||||
|
try {
|
||||||
|
const auth = await checkSessionOrInternalAuth(request)
|
||||||
|
if (!auth.success || !auth.userId) {
|
||||||
|
return NextResponse.json({ error: auth.error || 'Unauthorized' }, { status: 401 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const {
|
||||||
|
domain,
|
||||||
|
accessToken,
|
||||||
|
cloudId: providedCloudId,
|
||||||
|
pageId,
|
||||||
|
labelName,
|
||||||
|
} = await request.json()
|
||||||
|
|
||||||
|
if (!domain) {
|
||||||
|
return NextResponse.json({ error: 'Domain is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!accessToken) {
|
||||||
|
return NextResponse.json({ error: 'Access token is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!pageId) {
|
||||||
|
return NextResponse.json({ error: 'Page ID is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!labelName) {
|
||||||
|
return NextResponse.json({ error: 'Label name is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const pageIdValidation = validateAlphanumericId(pageId, 'pageId', 255)
|
||||||
|
if (!pageIdValidation.isValid) {
|
||||||
|
return NextResponse.json({ error: pageIdValidation.error }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const cloudId = providedCloudId || (await getConfluenceCloudId(domain, accessToken))
|
||||||
|
|
||||||
|
const cloudIdValidation = validateJiraCloudId(cloudId, 'cloudId')
|
||||||
|
if (!cloudIdValidation.isValid) {
|
||||||
|
return NextResponse.json({ error: cloudIdValidation.error }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const encodedLabel = encodeURIComponent(labelName.trim())
|
||||||
|
const url = `https://api.atlassian.com/ex/confluence/${cloudId}/wiki/rest/api/content/${pageId}/label?name=${encodedLabel}`
|
||||||
|
|
||||||
|
const response = await fetch(url, {
|
||||||
|
method: 'DELETE',
|
||||||
|
headers: {
|
||||||
|
Accept: 'application/json',
|
||||||
|
Authorization: `Bearer ${accessToken}`,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
if (!response.ok) {
|
||||||
|
const errorData = await response.json().catch(() => null)
|
||||||
|
logger.error('Confluence API error response:', {
|
||||||
|
status: response.status,
|
||||||
|
statusText: response.statusText,
|
||||||
|
error: JSON.stringify(errorData, null, 2),
|
||||||
|
})
|
||||||
|
const errorMessage =
|
||||||
|
errorData?.message || `Failed to delete Confluence label (${response.status})`
|
||||||
|
return NextResponse.json({ error: errorMessage }, { status: response.status })
|
||||||
|
}
|
||||||
|
|
||||||
|
return NextResponse.json({
|
||||||
|
pageId,
|
||||||
|
labelName,
|
||||||
|
deleted: true,
|
||||||
|
})
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Error deleting Confluence label:', error)
|
||||||
|
return NextResponse.json(
|
||||||
|
{ error: (error as Error).message || 'Internal server error' },
|
||||||
|
{ status: 500 }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
103
apps/sim/app/api/tools/confluence/pages-by-label/route.ts
Normal file
103
apps/sim/app/api/tools/confluence/pages-by-label/route.ts
Normal file
@@ -0,0 +1,103 @@
|
|||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { type NextRequest, NextResponse } from 'next/server'
|
||||||
|
import { checkSessionOrInternalAuth } from '@/lib/auth/hybrid'
|
||||||
|
import { validateAlphanumericId, validateJiraCloudId } from '@/lib/core/security/input-validation'
|
||||||
|
import { getConfluenceCloudId } from '@/tools/confluence/utils'
|
||||||
|
|
||||||
|
const logger = createLogger('ConfluencePagesByLabelAPI')
|
||||||
|
|
||||||
|
export const dynamic = 'force-dynamic'
|
||||||
|
|
||||||
|
export async function GET(request: NextRequest) {
|
||||||
|
try {
|
||||||
|
const auth = await checkSessionOrInternalAuth(request)
|
||||||
|
if (!auth.success || !auth.userId) {
|
||||||
|
return NextResponse.json({ error: auth.error || 'Unauthorized' }, { status: 401 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const { searchParams } = new URL(request.url)
|
||||||
|
const domain = searchParams.get('domain')
|
||||||
|
const accessToken = searchParams.get('accessToken')
|
||||||
|
const labelId = searchParams.get('labelId')
|
||||||
|
const providedCloudId = searchParams.get('cloudId')
|
||||||
|
const limit = searchParams.get('limit') || '50'
|
||||||
|
const cursor = searchParams.get('cursor')
|
||||||
|
|
||||||
|
if (!domain) {
|
||||||
|
return NextResponse.json({ error: 'Domain is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!accessToken) {
|
||||||
|
return NextResponse.json({ error: 'Access token is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!labelId) {
|
||||||
|
return NextResponse.json({ error: 'Label ID is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const labelIdValidation = validateAlphanumericId(labelId, 'labelId', 255)
|
||||||
|
if (!labelIdValidation.isValid) {
|
||||||
|
return NextResponse.json({ error: labelIdValidation.error }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const cloudId = providedCloudId || (await getConfluenceCloudId(domain, accessToken))
|
||||||
|
|
||||||
|
const cloudIdValidation = validateJiraCloudId(cloudId, 'cloudId')
|
||||||
|
if (!cloudIdValidation.isValid) {
|
||||||
|
return NextResponse.json({ error: cloudIdValidation.error }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const queryParams = new URLSearchParams()
|
||||||
|
queryParams.append('limit', String(Math.min(Number(limit), 250)))
|
||||||
|
if (cursor) {
|
||||||
|
queryParams.append('cursor', cursor)
|
||||||
|
}
|
||||||
|
const url = `https://api.atlassian.com/ex/confluence/${cloudId}/wiki/api/v2/labels/${labelId}/pages?${queryParams.toString()}`
|
||||||
|
|
||||||
|
const response = await fetch(url, {
|
||||||
|
method: 'GET',
|
||||||
|
headers: {
|
||||||
|
Accept: 'application/json',
|
||||||
|
Authorization: `Bearer ${accessToken}`,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
if (!response.ok) {
|
||||||
|
const errorData = await response.json().catch(() => null)
|
||||||
|
logger.error('Confluence API error response:', {
|
||||||
|
status: response.status,
|
||||||
|
statusText: response.statusText,
|
||||||
|
error: JSON.stringify(errorData, null, 2),
|
||||||
|
})
|
||||||
|
const errorMessage = errorData?.message || `Failed to get pages by label (${response.status})`
|
||||||
|
return NextResponse.json({ error: errorMessage }, { status: response.status })
|
||||||
|
}
|
||||||
|
|
||||||
|
const data = await response.json()
|
||||||
|
|
||||||
|
const pages = (data.results || []).map((page: any) => ({
|
||||||
|
id: page.id,
|
||||||
|
title: page.title,
|
||||||
|
status: page.status ?? null,
|
||||||
|
spaceId: page.spaceId ?? null,
|
||||||
|
parentId: page.parentId ?? null,
|
||||||
|
authorId: page.authorId ?? null,
|
||||||
|
createdAt: page.createdAt ?? null,
|
||||||
|
version: page.version ?? null,
|
||||||
|
}))
|
||||||
|
|
||||||
|
return NextResponse.json({
|
||||||
|
pages,
|
||||||
|
labelId,
|
||||||
|
nextCursor: data._links?.next
|
||||||
|
? new URL(data._links.next, 'https://placeholder').searchParams.get('cursor')
|
||||||
|
: null,
|
||||||
|
})
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Error getting pages by label:', error)
|
||||||
|
return NextResponse.json(
|
||||||
|
{ error: (error as Error).message || 'Internal server error' },
|
||||||
|
{ status: 500 }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
98
apps/sim/app/api/tools/confluence/space-labels/route.ts
Normal file
98
apps/sim/app/api/tools/confluence/space-labels/route.ts
Normal file
@@ -0,0 +1,98 @@
|
|||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { type NextRequest, NextResponse } from 'next/server'
|
||||||
|
import { checkSessionOrInternalAuth } from '@/lib/auth/hybrid'
|
||||||
|
import { validateAlphanumericId, validateJiraCloudId } from '@/lib/core/security/input-validation'
|
||||||
|
import { getConfluenceCloudId } from '@/tools/confluence/utils'
|
||||||
|
|
||||||
|
const logger = createLogger('ConfluenceSpaceLabelsAPI')
|
||||||
|
|
||||||
|
export const dynamic = 'force-dynamic'
|
||||||
|
|
||||||
|
export async function GET(request: NextRequest) {
|
||||||
|
try {
|
||||||
|
const auth = await checkSessionOrInternalAuth(request)
|
||||||
|
if (!auth.success || !auth.userId) {
|
||||||
|
return NextResponse.json({ error: auth.error || 'Unauthorized' }, { status: 401 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const { searchParams } = new URL(request.url)
|
||||||
|
const domain = searchParams.get('domain')
|
||||||
|
const accessToken = searchParams.get('accessToken')
|
||||||
|
const spaceId = searchParams.get('spaceId')
|
||||||
|
const providedCloudId = searchParams.get('cloudId')
|
||||||
|
const limit = searchParams.get('limit') || '25'
|
||||||
|
const cursor = searchParams.get('cursor')
|
||||||
|
|
||||||
|
if (!domain) {
|
||||||
|
return NextResponse.json({ error: 'Domain is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!accessToken) {
|
||||||
|
return NextResponse.json({ error: 'Access token is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!spaceId) {
|
||||||
|
return NextResponse.json({ error: 'Space ID is required' }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const spaceIdValidation = validateAlphanumericId(spaceId, 'spaceId', 255)
|
||||||
|
if (!spaceIdValidation.isValid) {
|
||||||
|
return NextResponse.json({ error: spaceIdValidation.error }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const cloudId = providedCloudId || (await getConfluenceCloudId(domain, accessToken))
|
||||||
|
|
||||||
|
const cloudIdValidation = validateJiraCloudId(cloudId, 'cloudId')
|
||||||
|
if (!cloudIdValidation.isValid) {
|
||||||
|
return NextResponse.json({ error: cloudIdValidation.error }, { status: 400 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const queryParams = new URLSearchParams()
|
||||||
|
queryParams.append('limit', String(Math.min(Number(limit), 250)))
|
||||||
|
if (cursor) {
|
||||||
|
queryParams.append('cursor', cursor)
|
||||||
|
}
|
||||||
|
const url = `https://api.atlassian.com/ex/confluence/${cloudId}/wiki/api/v2/spaces/${spaceId}/labels?${queryParams.toString()}`
|
||||||
|
|
||||||
|
const response = await fetch(url, {
|
||||||
|
method: 'GET',
|
||||||
|
headers: {
|
||||||
|
Accept: 'application/json',
|
||||||
|
Authorization: `Bearer ${accessToken}`,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
if (!response.ok) {
|
||||||
|
const errorData = await response.json().catch(() => null)
|
||||||
|
logger.error('Confluence API error response:', {
|
||||||
|
status: response.status,
|
||||||
|
statusText: response.statusText,
|
||||||
|
error: JSON.stringify(errorData, null, 2),
|
||||||
|
})
|
||||||
|
const errorMessage = errorData?.message || `Failed to list space labels (${response.status})`
|
||||||
|
return NextResponse.json({ error: errorMessage }, { status: response.status })
|
||||||
|
}
|
||||||
|
|
||||||
|
const data = await response.json()
|
||||||
|
|
||||||
|
const labels = (data.results || []).map((label: any) => ({
|
||||||
|
id: label.id,
|
||||||
|
name: label.name,
|
||||||
|
prefix: label.prefix || 'global',
|
||||||
|
}))
|
||||||
|
|
||||||
|
return NextResponse.json({
|
||||||
|
labels,
|
||||||
|
spaceId,
|
||||||
|
nextCursor: data._links?.next
|
||||||
|
? new URL(data._links.next, 'https://placeholder').searchParams.get('cursor')
|
||||||
|
: null,
|
||||||
|
})
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Error listing space labels:', error)
|
||||||
|
return NextResponse.json(
|
||||||
|
{ error: (error as Error).message || 'Internal server error' },
|
||||||
|
{ status: 500 }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -66,6 +66,12 @@
|
|||||||
* Credits:
|
* Credits:
|
||||||
* POST /api/v1/admin/credits - Issue credits to user (by userId or email)
|
* POST /api/v1/admin/credits - Issue credits to user (by userId or email)
|
||||||
*
|
*
|
||||||
|
* Referral Campaigns:
|
||||||
|
* GET /api/v1/admin/referral-campaigns - List campaigns (?active=true/false)
|
||||||
|
* POST /api/v1/admin/referral-campaigns - Create campaign
|
||||||
|
* GET /api/v1/admin/referral-campaigns/:id - Get campaign details
|
||||||
|
* PATCH /api/v1/admin/referral-campaigns/:id - Update campaign fields
|
||||||
|
*
|
||||||
* Access Control (Permission Groups):
|
* Access Control (Permission Groups):
|
||||||
* GET /api/v1/admin/access-control - List permission groups (?organizationId=X)
|
* GET /api/v1/admin/access-control - List permission groups (?organizationId=X)
|
||||||
* DELETE /api/v1/admin/access-control - Delete permission groups for org (?organizationId=X)
|
* DELETE /api/v1/admin/access-control - Delete permission groups for org (?organizationId=X)
|
||||||
@@ -97,6 +103,7 @@ export type {
|
|||||||
AdminOrganization,
|
AdminOrganization,
|
||||||
AdminOrganizationBillingSummary,
|
AdminOrganizationBillingSummary,
|
||||||
AdminOrganizationDetail,
|
AdminOrganizationDetail,
|
||||||
|
AdminReferralCampaign,
|
||||||
AdminSeatAnalytics,
|
AdminSeatAnalytics,
|
||||||
AdminSingleResponse,
|
AdminSingleResponse,
|
||||||
AdminSubscription,
|
AdminSubscription,
|
||||||
@@ -111,6 +118,7 @@ export type {
|
|||||||
AdminWorkspaceMember,
|
AdminWorkspaceMember,
|
||||||
DbMember,
|
DbMember,
|
||||||
DbOrganization,
|
DbOrganization,
|
||||||
|
DbReferralCampaign,
|
||||||
DbSubscription,
|
DbSubscription,
|
||||||
DbUser,
|
DbUser,
|
||||||
DbUserStats,
|
DbUserStats,
|
||||||
@@ -139,6 +147,7 @@ export {
|
|||||||
parseWorkflowVariables,
|
parseWorkflowVariables,
|
||||||
toAdminFolder,
|
toAdminFolder,
|
||||||
toAdminOrganization,
|
toAdminOrganization,
|
||||||
|
toAdminReferralCampaign,
|
||||||
toAdminSubscription,
|
toAdminSubscription,
|
||||||
toAdminUser,
|
toAdminUser,
|
||||||
toAdminWorkflow,
|
toAdminWorkflow,
|
||||||
|
|||||||
142
apps/sim/app/api/v1/admin/referral-campaigns/[id]/route.ts
Normal file
142
apps/sim/app/api/v1/admin/referral-campaigns/[id]/route.ts
Normal file
@@ -0,0 +1,142 @@
|
|||||||
|
/**
|
||||||
|
* GET /api/v1/admin/referral-campaigns/:id
|
||||||
|
*
|
||||||
|
* Get a single referral campaign by ID.
|
||||||
|
*
|
||||||
|
* PATCH /api/v1/admin/referral-campaigns/:id
|
||||||
|
*
|
||||||
|
* Update campaign fields. All fields are optional.
|
||||||
|
*
|
||||||
|
* Body:
|
||||||
|
* - name: string (non-empty) - Campaign name
|
||||||
|
* - bonusCreditAmount: number (> 0) - Bonus credits in dollars
|
||||||
|
* - isActive: boolean - Enable/disable the campaign
|
||||||
|
* - code: string | null (min 6 chars, auto-uppercased, null to remove) - Redeemable code
|
||||||
|
* - utmSource: string | null - UTM source match (null = wildcard)
|
||||||
|
* - utmMedium: string | null - UTM medium match (null = wildcard)
|
||||||
|
* - utmCampaign: string | null - UTM campaign match (null = wildcard)
|
||||||
|
* - utmContent: string | null - UTM content match (null = wildcard)
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { db } from '@sim/db'
|
||||||
|
import { referralCampaigns } from '@sim/db/schema'
|
||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { eq } from 'drizzle-orm'
|
||||||
|
import { getBaseUrl } from '@/lib/core/utils/urls'
|
||||||
|
import { withAdminAuthParams } from '@/app/api/v1/admin/middleware'
|
||||||
|
import {
|
||||||
|
badRequestResponse,
|
||||||
|
internalErrorResponse,
|
||||||
|
notFoundResponse,
|
||||||
|
singleResponse,
|
||||||
|
} from '@/app/api/v1/admin/responses'
|
||||||
|
import { toAdminReferralCampaign } from '@/app/api/v1/admin/types'
|
||||||
|
|
||||||
|
const logger = createLogger('AdminReferralCampaignDetailAPI')
|
||||||
|
|
||||||
|
interface RouteParams {
|
||||||
|
id: string
|
||||||
|
}
|
||||||
|
|
||||||
|
export const GET = withAdminAuthParams<RouteParams>(async (_, context) => {
|
||||||
|
try {
|
||||||
|
const { id: campaignId } = await context.params
|
||||||
|
|
||||||
|
const [campaign] = await db
|
||||||
|
.select()
|
||||||
|
.from(referralCampaigns)
|
||||||
|
.where(eq(referralCampaigns.id, campaignId))
|
||||||
|
.limit(1)
|
||||||
|
|
||||||
|
if (!campaign) {
|
||||||
|
return notFoundResponse('Campaign')
|
||||||
|
}
|
||||||
|
|
||||||
|
logger.info(`Admin API: Retrieved referral campaign ${campaignId}`)
|
||||||
|
|
||||||
|
return singleResponse(toAdminReferralCampaign(campaign, getBaseUrl()))
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Admin API: Failed to get referral campaign', { error })
|
||||||
|
return internalErrorResponse('Failed to get referral campaign')
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
export const PATCH = withAdminAuthParams<RouteParams>(async (request, context) => {
|
||||||
|
try {
|
||||||
|
const { id: campaignId } = await context.params
|
||||||
|
const body = await request.json()
|
||||||
|
|
||||||
|
const [existing] = await db
|
||||||
|
.select()
|
||||||
|
.from(referralCampaigns)
|
||||||
|
.where(eq(referralCampaigns.id, campaignId))
|
||||||
|
.limit(1)
|
||||||
|
|
||||||
|
if (!existing) {
|
||||||
|
return notFoundResponse('Campaign')
|
||||||
|
}
|
||||||
|
|
||||||
|
const updateData: Record<string, unknown> = { updatedAt: new Date() }
|
||||||
|
|
||||||
|
if (body.name !== undefined) {
|
||||||
|
if (typeof body.name !== 'string' || body.name.trim().length === 0) {
|
||||||
|
return badRequestResponse('name must be a non-empty string')
|
||||||
|
}
|
||||||
|
updateData.name = body.name.trim()
|
||||||
|
}
|
||||||
|
|
||||||
|
if (body.bonusCreditAmount !== undefined) {
|
||||||
|
if (
|
||||||
|
typeof body.bonusCreditAmount !== 'number' ||
|
||||||
|
!Number.isFinite(body.bonusCreditAmount) ||
|
||||||
|
body.bonusCreditAmount <= 0
|
||||||
|
) {
|
||||||
|
return badRequestResponse('bonusCreditAmount must be a positive number')
|
||||||
|
}
|
||||||
|
updateData.bonusCreditAmount = body.bonusCreditAmount.toString()
|
||||||
|
}
|
||||||
|
|
||||||
|
if (body.isActive !== undefined) {
|
||||||
|
if (typeof body.isActive !== 'boolean') {
|
||||||
|
return badRequestResponse('isActive must be a boolean')
|
||||||
|
}
|
||||||
|
updateData.isActive = body.isActive
|
||||||
|
}
|
||||||
|
|
||||||
|
if (body.code !== undefined) {
|
||||||
|
if (body.code !== null) {
|
||||||
|
if (typeof body.code !== 'string') {
|
||||||
|
return badRequestResponse('code must be a string or null')
|
||||||
|
}
|
||||||
|
if (body.code.trim().length < 6) {
|
||||||
|
return badRequestResponse('code must be at least 6 characters')
|
||||||
|
}
|
||||||
|
}
|
||||||
|
updateData.code = body.code ? body.code.trim().toUpperCase() : null
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const field of ['utmSource', 'utmMedium', 'utmCampaign', 'utmContent'] as const) {
|
||||||
|
if (body[field] !== undefined) {
|
||||||
|
if (body[field] !== null && typeof body[field] !== 'string') {
|
||||||
|
return badRequestResponse(`${field} must be a string or null`)
|
||||||
|
}
|
||||||
|
updateData[field] = body[field] || null
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const [updated] = await db
|
||||||
|
.update(referralCampaigns)
|
||||||
|
.set(updateData)
|
||||||
|
.where(eq(referralCampaigns.id, campaignId))
|
||||||
|
.returning()
|
||||||
|
|
||||||
|
logger.info(`Admin API: Updated referral campaign ${campaignId}`, {
|
||||||
|
fields: Object.keys(updateData).filter((k) => k !== 'updatedAt'),
|
||||||
|
})
|
||||||
|
|
||||||
|
return singleResponse(toAdminReferralCampaign(updated, getBaseUrl()))
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Admin API: Failed to update referral campaign', { error })
|
||||||
|
return internalErrorResponse('Failed to update referral campaign')
|
||||||
|
}
|
||||||
|
})
|
||||||
140
apps/sim/app/api/v1/admin/referral-campaigns/route.ts
Normal file
140
apps/sim/app/api/v1/admin/referral-campaigns/route.ts
Normal file
@@ -0,0 +1,140 @@
|
|||||||
|
/**
|
||||||
|
* GET /api/v1/admin/referral-campaigns
|
||||||
|
*
|
||||||
|
* List referral campaigns with optional filtering and pagination.
|
||||||
|
*
|
||||||
|
* Query Parameters:
|
||||||
|
* - active: string (optional) - Filter by active status ('true' or 'false')
|
||||||
|
* - limit: number (default: 50, max: 250)
|
||||||
|
* - offset: number (default: 0)
|
||||||
|
*
|
||||||
|
* POST /api/v1/admin/referral-campaigns
|
||||||
|
*
|
||||||
|
* Create a new referral campaign.
|
||||||
|
*
|
||||||
|
* Body:
|
||||||
|
* - name: string (required) - Campaign name
|
||||||
|
* - bonusCreditAmount: number (required, > 0) - Bonus credits in dollars
|
||||||
|
* - code: string | null (optional, min 6 chars, auto-uppercased) - Redeemable code
|
||||||
|
* - utmSource: string | null (optional) - UTM source match (null = wildcard)
|
||||||
|
* - utmMedium: string | null (optional) - UTM medium match (null = wildcard)
|
||||||
|
* - utmCampaign: string | null (optional) - UTM campaign match (null = wildcard)
|
||||||
|
* - utmContent: string | null (optional) - UTM content match (null = wildcard)
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { db } from '@sim/db'
|
||||||
|
import { referralCampaigns } from '@sim/db/schema'
|
||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { count, eq, type SQL } from 'drizzle-orm'
|
||||||
|
import { nanoid } from 'nanoid'
|
||||||
|
import { getBaseUrl } from '@/lib/core/utils/urls'
|
||||||
|
import { withAdminAuth } from '@/app/api/v1/admin/middleware'
|
||||||
|
import {
|
||||||
|
badRequestResponse,
|
||||||
|
internalErrorResponse,
|
||||||
|
listResponse,
|
||||||
|
singleResponse,
|
||||||
|
} from '@/app/api/v1/admin/responses'
|
||||||
|
import {
|
||||||
|
type AdminReferralCampaign,
|
||||||
|
createPaginationMeta,
|
||||||
|
parsePaginationParams,
|
||||||
|
toAdminReferralCampaign,
|
||||||
|
} from '@/app/api/v1/admin/types'
|
||||||
|
|
||||||
|
const logger = createLogger('AdminReferralCampaignsAPI')
|
||||||
|
|
||||||
|
export const GET = withAdminAuth(async (request) => {
|
||||||
|
const url = new URL(request.url)
|
||||||
|
const { limit, offset } = parsePaginationParams(url)
|
||||||
|
const activeFilter = url.searchParams.get('active')
|
||||||
|
|
||||||
|
try {
|
||||||
|
const conditions: SQL<unknown>[] = []
|
||||||
|
if (activeFilter === 'true') {
|
||||||
|
conditions.push(eq(referralCampaigns.isActive, true))
|
||||||
|
} else if (activeFilter === 'false') {
|
||||||
|
conditions.push(eq(referralCampaigns.isActive, false))
|
||||||
|
}
|
||||||
|
|
||||||
|
const whereClause = conditions.length > 0 ? conditions[0] : undefined
|
||||||
|
const baseUrl = getBaseUrl()
|
||||||
|
|
||||||
|
const [countResult, campaigns] = await Promise.all([
|
||||||
|
db.select({ total: count() }).from(referralCampaigns).where(whereClause),
|
||||||
|
db
|
||||||
|
.select()
|
||||||
|
.from(referralCampaigns)
|
||||||
|
.where(whereClause)
|
||||||
|
.orderBy(referralCampaigns.createdAt)
|
||||||
|
.limit(limit)
|
||||||
|
.offset(offset),
|
||||||
|
])
|
||||||
|
|
||||||
|
const total = countResult[0].total
|
||||||
|
const data: AdminReferralCampaign[] = campaigns.map((c) => toAdminReferralCampaign(c, baseUrl))
|
||||||
|
const pagination = createPaginationMeta(total, limit, offset)
|
||||||
|
|
||||||
|
logger.info(`Admin API: Listed ${data.length} referral campaigns (total: ${total})`)
|
||||||
|
|
||||||
|
return listResponse(data, pagination)
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Admin API: Failed to list referral campaigns', { error })
|
||||||
|
return internalErrorResponse('Failed to list referral campaigns')
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
export const POST = withAdminAuth(async (request) => {
|
||||||
|
try {
|
||||||
|
const body = await request.json()
|
||||||
|
const { name, code, utmSource, utmMedium, utmCampaign, utmContent, bonusCreditAmount } = body
|
||||||
|
|
||||||
|
if (!name || typeof name !== 'string') {
|
||||||
|
return badRequestResponse('name is required and must be a string')
|
||||||
|
}
|
||||||
|
|
||||||
|
if (
|
||||||
|
typeof bonusCreditAmount !== 'number' ||
|
||||||
|
!Number.isFinite(bonusCreditAmount) ||
|
||||||
|
bonusCreditAmount <= 0
|
||||||
|
) {
|
||||||
|
return badRequestResponse('bonusCreditAmount must be a positive number')
|
||||||
|
}
|
||||||
|
|
||||||
|
if (code !== undefined && code !== null) {
|
||||||
|
if (typeof code !== 'string') {
|
||||||
|
return badRequestResponse('code must be a string or null')
|
||||||
|
}
|
||||||
|
if (code.trim().length < 6) {
|
||||||
|
return badRequestResponse('code must be at least 6 characters')
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const id = nanoid()
|
||||||
|
|
||||||
|
const [campaign] = await db
|
||||||
|
.insert(referralCampaigns)
|
||||||
|
.values({
|
||||||
|
id,
|
||||||
|
name,
|
||||||
|
code: code ? code.trim().toUpperCase() : null,
|
||||||
|
utmSource: utmSource || null,
|
||||||
|
utmMedium: utmMedium || null,
|
||||||
|
utmCampaign: utmCampaign || null,
|
||||||
|
utmContent: utmContent || null,
|
||||||
|
bonusCreditAmount: bonusCreditAmount.toString(),
|
||||||
|
})
|
||||||
|
.returning()
|
||||||
|
|
||||||
|
logger.info(`Admin API: Created referral campaign ${id}`, {
|
||||||
|
name,
|
||||||
|
code: campaign.code,
|
||||||
|
bonusCreditAmount,
|
||||||
|
})
|
||||||
|
|
||||||
|
return singleResponse(toAdminReferralCampaign(campaign, getBaseUrl()))
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Admin API: Failed to create referral campaign', { error })
|
||||||
|
return internalErrorResponse('Failed to create referral campaign')
|
||||||
|
}
|
||||||
|
})
|
||||||
@@ -8,6 +8,7 @@
|
|||||||
import type {
|
import type {
|
||||||
member,
|
member,
|
||||||
organization,
|
organization,
|
||||||
|
referralCampaigns,
|
||||||
subscription,
|
subscription,
|
||||||
user,
|
user,
|
||||||
userStats,
|
userStats,
|
||||||
@@ -31,6 +32,7 @@ export type DbOrganization = InferSelectModel<typeof organization>
|
|||||||
export type DbSubscription = InferSelectModel<typeof subscription>
|
export type DbSubscription = InferSelectModel<typeof subscription>
|
||||||
export type DbMember = InferSelectModel<typeof member>
|
export type DbMember = InferSelectModel<typeof member>
|
||||||
export type DbUserStats = InferSelectModel<typeof userStats>
|
export type DbUserStats = InferSelectModel<typeof userStats>
|
||||||
|
export type DbReferralCampaign = InferSelectModel<typeof referralCampaigns>
|
||||||
|
|
||||||
// =============================================================================
|
// =============================================================================
|
||||||
// Pagination
|
// Pagination
|
||||||
@@ -646,3 +648,49 @@ export interface AdminDeployResult {
|
|||||||
export interface AdminUndeployResult {
|
export interface AdminUndeployResult {
|
||||||
isDeployed: boolean
|
isDeployed: boolean
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// =============================================================================
|
||||||
|
// Referral Campaign Types
|
||||||
|
// =============================================================================
|
||||||
|
|
||||||
|
export interface AdminReferralCampaign {
|
||||||
|
id: string
|
||||||
|
name: string
|
||||||
|
code: string | null
|
||||||
|
utmSource: string | null
|
||||||
|
utmMedium: string | null
|
||||||
|
utmCampaign: string | null
|
||||||
|
utmContent: string | null
|
||||||
|
bonusCreditAmount: string
|
||||||
|
isActive: boolean
|
||||||
|
signupUrl: string | null
|
||||||
|
createdAt: string
|
||||||
|
updatedAt: string
|
||||||
|
}
|
||||||
|
|
||||||
|
export function toAdminReferralCampaign(
|
||||||
|
dbCampaign: DbReferralCampaign,
|
||||||
|
baseUrl: string
|
||||||
|
): AdminReferralCampaign {
|
||||||
|
const utmParams = new URLSearchParams()
|
||||||
|
if (dbCampaign.utmSource) utmParams.set('utm_source', dbCampaign.utmSource)
|
||||||
|
if (dbCampaign.utmMedium) utmParams.set('utm_medium', dbCampaign.utmMedium)
|
||||||
|
if (dbCampaign.utmCampaign) utmParams.set('utm_campaign', dbCampaign.utmCampaign)
|
||||||
|
if (dbCampaign.utmContent) utmParams.set('utm_content', dbCampaign.utmContent)
|
||||||
|
const query = utmParams.toString()
|
||||||
|
|
||||||
|
return {
|
||||||
|
id: dbCampaign.id,
|
||||||
|
name: dbCampaign.name,
|
||||||
|
code: dbCampaign.code,
|
||||||
|
utmSource: dbCampaign.utmSource,
|
||||||
|
utmMedium: dbCampaign.utmMedium,
|
||||||
|
utmCampaign: dbCampaign.utmCampaign,
|
||||||
|
utmContent: dbCampaign.utmContent,
|
||||||
|
bonusCreditAmount: dbCampaign.bonusCreditAmount,
|
||||||
|
isActive: dbCampaign.isActive,
|
||||||
|
signupUrl: query ? `${baseUrl}/signup?${query}` : null,
|
||||||
|
createdAt: dbCampaign.createdAt.toISOString(),
|
||||||
|
updatedAt: dbCampaign.updatedAt.toISOString(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import { resolveWorkflowIdForUser } from '@/lib/workflows/utils'
|
|||||||
import { authenticateV1Request } from '@/app/api/v1/auth'
|
import { authenticateV1Request } from '@/app/api/v1/auth'
|
||||||
|
|
||||||
const logger = createLogger('CopilotHeadlessAPI')
|
const logger = createLogger('CopilotHeadlessAPI')
|
||||||
const DEFAULT_COPILOT_MODEL = 'claude-opus-4-6'
|
const DEFAULT_COPILOT_MODEL = 'claude-opus-4-5'
|
||||||
|
|
||||||
const RequestSchema = z.object({
|
const RequestSchema = z.object({
|
||||||
message: z.string().min(1, 'message is required'),
|
message: z.string().min(1, 'message is required'),
|
||||||
|
|||||||
@@ -29,7 +29,7 @@ const patchBodySchema = z
|
|||||||
description: z
|
description: z
|
||||||
.string()
|
.string()
|
||||||
.trim()
|
.trim()
|
||||||
.max(500, 'Description must be 500 characters or less')
|
.max(2000, 'Description must be 2000 characters or less')
|
||||||
.nullable()
|
.nullable()
|
||||||
.optional(),
|
.optional(),
|
||||||
isActive: z.literal(true).optional(), // Set to true to activate this version
|
isActive: z.literal(true).optional(), // Set to true to activate this version
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ import {
|
|||||||
import { generateRequestId } from '@/lib/core/utils/request'
|
import { generateRequestId } from '@/lib/core/utils/request'
|
||||||
import { SSE_HEADERS } from '@/lib/core/utils/sse'
|
import { SSE_HEADERS } from '@/lib/core/utils/sse'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { markExecutionCancelled } from '@/lib/execution/cancellation'
|
import { createExecutionEventWriter, setExecutionMeta } from '@/lib/execution/event-buffer'
|
||||||
import { processInputFileFields } from '@/lib/execution/files'
|
import { processInputFileFields } from '@/lib/execution/files'
|
||||||
import { preprocessExecution } from '@/lib/execution/preprocessing'
|
import { preprocessExecution } from '@/lib/execution/preprocessing'
|
||||||
import { LoggingSession } from '@/lib/logs/execution/logging-session'
|
import { LoggingSession } from '@/lib/logs/execution/logging-session'
|
||||||
@@ -700,15 +700,27 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
const timeoutController = createTimeoutAbortController(preprocessResult.executionTimeout?.sync)
|
const timeoutController = createTimeoutAbortController(preprocessResult.executionTimeout?.sync)
|
||||||
let isStreamClosed = false
|
let isStreamClosed = false
|
||||||
|
|
||||||
|
const eventWriter = createExecutionEventWriter(executionId)
|
||||||
|
setExecutionMeta(executionId, {
|
||||||
|
status: 'active',
|
||||||
|
userId: actorUserId,
|
||||||
|
workflowId,
|
||||||
|
}).catch(() => {})
|
||||||
|
|
||||||
const stream = new ReadableStream<Uint8Array>({
|
const stream = new ReadableStream<Uint8Array>({
|
||||||
async start(controller) {
|
async start(controller) {
|
||||||
const sendEvent = (event: ExecutionEvent) => {
|
let finalMetaStatus: 'complete' | 'error' | 'cancelled' | null = null
|
||||||
if (isStreamClosed) return
|
|
||||||
|
|
||||||
try {
|
const sendEvent = (event: ExecutionEvent) => {
|
||||||
controller.enqueue(encodeSSEEvent(event))
|
if (!isStreamClosed) {
|
||||||
} catch {
|
try {
|
||||||
isStreamClosed = true
|
controller.enqueue(encodeSSEEvent(event))
|
||||||
|
} catch {
|
||||||
|
isStreamClosed = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (event.type !== 'stream:chunk' && event.type !== 'stream:done') {
|
||||||
|
eventWriter.write(event).catch(() => {})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -829,14 +841,12 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
|
|
||||||
const reader = streamingExec.stream.getReader()
|
const reader = streamingExec.stream.getReader()
|
||||||
const decoder = new TextDecoder()
|
const decoder = new TextDecoder()
|
||||||
let chunkCount = 0
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
while (true) {
|
while (true) {
|
||||||
const { done, value } = await reader.read()
|
const { done, value } = await reader.read()
|
||||||
if (done) break
|
if (done) break
|
||||||
|
|
||||||
chunkCount++
|
|
||||||
const chunk = decoder.decode(value, { stream: true })
|
const chunk = decoder.decode(value, { stream: true })
|
||||||
sendEvent({
|
sendEvent({
|
||||||
type: 'stream:chunk',
|
type: 'stream:chunk',
|
||||||
@@ -951,6 +961,7 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
duration: result.metadata?.duration || 0,
|
duration: result.metadata?.duration || 0,
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
finalMetaStatus = 'error'
|
||||||
} else {
|
} else {
|
||||||
logger.info(`[${requestId}] Workflow execution was cancelled`)
|
logger.info(`[${requestId}] Workflow execution was cancelled`)
|
||||||
|
|
||||||
@@ -963,6 +974,7 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
duration: result.metadata?.duration || 0,
|
duration: result.metadata?.duration || 0,
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
finalMetaStatus = 'cancelled'
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
@@ -986,6 +998,7 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
endTime: result.metadata?.endTime || new Date().toISOString(),
|
endTime: result.metadata?.endTime || new Date().toISOString(),
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
finalMetaStatus = 'complete'
|
||||||
} catch (error: unknown) {
|
} catch (error: unknown) {
|
||||||
const isTimeout = isTimeoutError(error) || timeoutController.isTimedOut()
|
const isTimeout = isTimeoutError(error) || timeoutController.isTimedOut()
|
||||||
const errorMessage = isTimeout
|
const errorMessage = isTimeout
|
||||||
@@ -1017,7 +1030,18 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
duration: executionResult?.metadata?.duration || 0,
|
duration: executionResult?.metadata?.duration || 0,
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
finalMetaStatus = 'error'
|
||||||
} finally {
|
} finally {
|
||||||
|
try {
|
||||||
|
await eventWriter.close()
|
||||||
|
} catch (closeError) {
|
||||||
|
logger.warn(`[${requestId}] Failed to close event writer`, {
|
||||||
|
error: closeError instanceof Error ? closeError.message : String(closeError),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
if (finalMetaStatus) {
|
||||||
|
setExecutionMeta(executionId, { status: finalMetaStatus }).catch(() => {})
|
||||||
|
}
|
||||||
timeoutController.cleanup()
|
timeoutController.cleanup()
|
||||||
if (executionId) {
|
if (executionId) {
|
||||||
await cleanupExecutionBase64Cache(executionId)
|
await cleanupExecutionBase64Cache(executionId)
|
||||||
@@ -1032,10 +1056,7 @@ export async function POST(req: NextRequest, { params }: { params: Promise<{ id:
|
|||||||
},
|
},
|
||||||
cancel() {
|
cancel() {
|
||||||
isStreamClosed = true
|
isStreamClosed = true
|
||||||
timeoutController.cleanup()
|
logger.info(`[${requestId}] Client disconnected from SSE stream`)
|
||||||
logger.info(`[${requestId}] Client aborted SSE stream, signalling cancellation`)
|
|
||||||
timeoutController.abort()
|
|
||||||
markExecutionCancelled(executionId).catch(() => {})
|
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|||||||
@@ -0,0 +1,170 @@
|
|||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { type NextRequest, NextResponse } from 'next/server'
|
||||||
|
import { checkHybridAuth } from '@/lib/auth/hybrid'
|
||||||
|
import { SSE_HEADERS } from '@/lib/core/utils/sse'
|
||||||
|
import {
|
||||||
|
type ExecutionStreamStatus,
|
||||||
|
getExecutionMeta,
|
||||||
|
readExecutionEvents,
|
||||||
|
} from '@/lib/execution/event-buffer'
|
||||||
|
import { formatSSEEvent } from '@/lib/workflows/executor/execution-events'
|
||||||
|
import { authorizeWorkflowByWorkspacePermission } from '@/lib/workflows/utils'
|
||||||
|
|
||||||
|
const logger = createLogger('ExecutionStreamReconnectAPI')
|
||||||
|
|
||||||
|
const POLL_INTERVAL_MS = 500
|
||||||
|
const MAX_POLL_DURATION_MS = 10 * 60 * 1000 // 10 minutes
|
||||||
|
|
||||||
|
function isTerminalStatus(status: ExecutionStreamStatus): boolean {
|
||||||
|
return status === 'complete' || status === 'error' || status === 'cancelled'
|
||||||
|
}
|
||||||
|
|
||||||
|
export const runtime = 'nodejs'
|
||||||
|
export const dynamic = 'force-dynamic'
|
||||||
|
|
||||||
|
export async function GET(
|
||||||
|
req: NextRequest,
|
||||||
|
{ params }: { params: Promise<{ id: string; executionId: string }> }
|
||||||
|
) {
|
||||||
|
const { id: workflowId, executionId } = await params
|
||||||
|
|
||||||
|
try {
|
||||||
|
const auth = await checkHybridAuth(req, { requireWorkflowId: false })
|
||||||
|
if (!auth.success || !auth.userId) {
|
||||||
|
return NextResponse.json({ error: auth.error || 'Unauthorized' }, { status: 401 })
|
||||||
|
}
|
||||||
|
|
||||||
|
const workflowAuthorization = await authorizeWorkflowByWorkspacePermission({
|
||||||
|
workflowId,
|
||||||
|
userId: auth.userId,
|
||||||
|
action: 'read',
|
||||||
|
})
|
||||||
|
if (!workflowAuthorization.allowed) {
|
||||||
|
return NextResponse.json(
|
||||||
|
{ error: workflowAuthorization.message || 'Access denied' },
|
||||||
|
{ status: workflowAuthorization.status }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
const meta = await getExecutionMeta(executionId)
|
||||||
|
if (!meta) {
|
||||||
|
return NextResponse.json({ error: 'Execution buffer not found or expired' }, { status: 404 })
|
||||||
|
}
|
||||||
|
|
||||||
|
if (meta.workflowId && meta.workflowId !== workflowId) {
|
||||||
|
return NextResponse.json(
|
||||||
|
{ error: 'Execution does not belong to this workflow' },
|
||||||
|
{ status: 403 }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
const fromParam = req.nextUrl.searchParams.get('from')
|
||||||
|
const parsed = fromParam ? Number.parseInt(fromParam, 10) : 0
|
||||||
|
const fromEventId = Number.isFinite(parsed) && parsed >= 0 ? parsed : 0
|
||||||
|
|
||||||
|
logger.info('Reconnection stream requested', {
|
||||||
|
workflowId,
|
||||||
|
executionId,
|
||||||
|
fromEventId,
|
||||||
|
metaStatus: meta.status,
|
||||||
|
})
|
||||||
|
|
||||||
|
const encoder = new TextEncoder()
|
||||||
|
|
||||||
|
let closed = false
|
||||||
|
|
||||||
|
const stream = new ReadableStream<Uint8Array>({
|
||||||
|
async start(controller) {
|
||||||
|
let lastEventId = fromEventId
|
||||||
|
const pollDeadline = Date.now() + MAX_POLL_DURATION_MS
|
||||||
|
|
||||||
|
const enqueue = (text: string) => {
|
||||||
|
if (closed) return
|
||||||
|
try {
|
||||||
|
controller.enqueue(encoder.encode(text))
|
||||||
|
} catch {
|
||||||
|
closed = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
const events = await readExecutionEvents(executionId, lastEventId)
|
||||||
|
for (const entry of events) {
|
||||||
|
if (closed) return
|
||||||
|
enqueue(formatSSEEvent(entry.event))
|
||||||
|
lastEventId = entry.eventId
|
||||||
|
}
|
||||||
|
|
||||||
|
const currentMeta = await getExecutionMeta(executionId)
|
||||||
|
if (!currentMeta || isTerminalStatus(currentMeta.status)) {
|
||||||
|
enqueue('data: [DONE]\n\n')
|
||||||
|
if (!closed) controller.close()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
while (!closed && Date.now() < pollDeadline) {
|
||||||
|
await new Promise((resolve) => setTimeout(resolve, POLL_INTERVAL_MS))
|
||||||
|
if (closed) return
|
||||||
|
|
||||||
|
const newEvents = await readExecutionEvents(executionId, lastEventId)
|
||||||
|
for (const entry of newEvents) {
|
||||||
|
if (closed) return
|
||||||
|
enqueue(formatSSEEvent(entry.event))
|
||||||
|
lastEventId = entry.eventId
|
||||||
|
}
|
||||||
|
|
||||||
|
const polledMeta = await getExecutionMeta(executionId)
|
||||||
|
if (!polledMeta || isTerminalStatus(polledMeta.status)) {
|
||||||
|
const finalEvents = await readExecutionEvents(executionId, lastEventId)
|
||||||
|
for (const entry of finalEvents) {
|
||||||
|
if (closed) return
|
||||||
|
enqueue(formatSSEEvent(entry.event))
|
||||||
|
lastEventId = entry.eventId
|
||||||
|
}
|
||||||
|
enqueue('data: [DONE]\n\n')
|
||||||
|
if (!closed) controller.close()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!closed) {
|
||||||
|
logger.warn('Reconnection stream poll deadline reached', { executionId })
|
||||||
|
enqueue('data: [DONE]\n\n')
|
||||||
|
controller.close()
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Error in reconnection stream', {
|
||||||
|
executionId,
|
||||||
|
error: error instanceof Error ? error.message : String(error),
|
||||||
|
})
|
||||||
|
if (!closed) {
|
||||||
|
try {
|
||||||
|
controller.close()
|
||||||
|
} catch {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
cancel() {
|
||||||
|
closed = true
|
||||||
|
logger.info('Client disconnected from reconnection stream', { executionId })
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
return new NextResponse(stream, {
|
||||||
|
headers: {
|
||||||
|
...SSE_HEADERS,
|
||||||
|
'X-Execution-Id': executionId,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
} catch (error: any) {
|
||||||
|
logger.error('Failed to start reconnection stream', {
|
||||||
|
workflowId,
|
||||||
|
executionId,
|
||||||
|
error: error.message,
|
||||||
|
})
|
||||||
|
return NextResponse.json(
|
||||||
|
{ error: error.message || 'Failed to start reconnection stream' },
|
||||||
|
{ status: 500 }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -5,7 +5,7 @@
|
|||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { loggerMock } from '@sim/testing'
|
import { loggerMock, setupGlobalFetchMock } from '@sim/testing'
|
||||||
import { NextRequest } from 'next/server'
|
import { NextRequest } from 'next/server'
|
||||||
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
import { afterEach, beforeEach, describe, expect, it, vi } from 'vitest'
|
||||||
|
|
||||||
@@ -284,9 +284,7 @@ describe('Workflow By ID API Route', () => {
|
|||||||
where: vi.fn().mockResolvedValue([{ id: 'workflow-123' }]),
|
where: vi.fn().mockResolvedValue([{ id: 'workflow-123' }]),
|
||||||
})
|
})
|
||||||
|
|
||||||
global.fetch = vi.fn().mockResolvedValue({
|
setupGlobalFetchMock({ ok: true })
|
||||||
ok: true,
|
|
||||||
})
|
|
||||||
|
|
||||||
const req = new NextRequest('http://localhost:3000/api/workflows/workflow-123', {
|
const req = new NextRequest('http://localhost:3000/api/workflows/workflow-123', {
|
||||||
method: 'DELETE',
|
method: 'DELETE',
|
||||||
@@ -331,9 +329,7 @@ describe('Workflow By ID API Route', () => {
|
|||||||
where: vi.fn().mockResolvedValue([{ id: 'workflow-123' }]),
|
where: vi.fn().mockResolvedValue([{ id: 'workflow-123' }]),
|
||||||
})
|
})
|
||||||
|
|
||||||
global.fetch = vi.fn().mockResolvedValue({
|
setupGlobalFetchMock({ ok: true })
|
||||||
ok: true,
|
|
||||||
})
|
|
||||||
|
|
||||||
const req = new NextRequest('http://localhost:3000/api/workflows/workflow-123', {
|
const req = new NextRequest('http://localhost:3000/api/workflows/workflow-123', {
|
||||||
method: 'DELETE',
|
method: 'DELETE',
|
||||||
|
|||||||
@@ -38,6 +38,7 @@ export async function GET(request: NextRequest, { params }: { params: Promise<{
|
|||||||
return NextResponse.json({ error: 'Unauthorized' }, { status: 401 })
|
return NextResponse.json({ error: 'Unauthorized' }, { status: 401 })
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const isInternalCall = auth.authType === 'internal_jwt'
|
||||||
const userId = auth.userId || null
|
const userId = auth.userId || null
|
||||||
|
|
||||||
let workflowData = await getWorkflowById(workflowId)
|
let workflowData = await getWorkflowById(workflowId)
|
||||||
@@ -47,29 +48,32 @@ export async function GET(request: NextRequest, { params }: { params: Promise<{
|
|||||||
return NextResponse.json({ error: 'Workflow not found' }, { status: 404 })
|
return NextResponse.json({ error: 'Workflow not found' }, { status: 404 })
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if user has access to this workflow
|
if (isInternalCall && !userId) {
|
||||||
if (!userId) {
|
// Internal system calls (e.g. workflow-in-workflow executor) may not carry a userId.
|
||||||
|
// These are already authenticated via internal JWT; allow read access.
|
||||||
|
logger.info(`[${requestId}] Internal API call for workflow ${workflowId}`)
|
||||||
|
} else if (!userId) {
|
||||||
logger.warn(`[${requestId}] Unauthorized access attempt for workflow ${workflowId}`)
|
logger.warn(`[${requestId}] Unauthorized access attempt for workflow ${workflowId}`)
|
||||||
return NextResponse.json({ error: 'Unauthorized' }, { status: 401 })
|
return NextResponse.json({ error: 'Unauthorized' }, { status: 401 })
|
||||||
}
|
} else {
|
||||||
|
const authorization = await authorizeWorkflowByWorkspacePermission({
|
||||||
|
workflowId,
|
||||||
|
userId,
|
||||||
|
action: 'read',
|
||||||
|
})
|
||||||
|
if (!authorization.workflow) {
|
||||||
|
logger.warn(`[${requestId}] Workflow ${workflowId} not found`)
|
||||||
|
return NextResponse.json({ error: 'Workflow not found' }, { status: 404 })
|
||||||
|
}
|
||||||
|
|
||||||
const authorization = await authorizeWorkflowByWorkspacePermission({
|
workflowData = authorization.workflow
|
||||||
workflowId,
|
if (!authorization.allowed) {
|
||||||
userId,
|
logger.warn(`[${requestId}] User ${userId} denied access to workflow ${workflowId}`)
|
||||||
action: 'read',
|
return NextResponse.json(
|
||||||
})
|
{ error: authorization.message || 'Access denied' },
|
||||||
if (!authorization.workflow) {
|
{ status: authorization.status }
|
||||||
logger.warn(`[${requestId}] Workflow ${workflowId} not found`)
|
)
|
||||||
return NextResponse.json({ error: 'Workflow not found' }, { status: 404 })
|
}
|
||||||
}
|
|
||||||
|
|
||||||
workflowData = authorization.workflow
|
|
||||||
if (!authorization.allowed) {
|
|
||||||
logger.warn(`[${requestId}] User ${userId} denied access to workflow ${workflowId}`)
|
|
||||||
return NextResponse.json(
|
|
||||||
{ error: authorization.message || 'Access denied' },
|
|
||||||
{ status: authorization.status }
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.debug(`[${requestId}] Attempting to load workflow ${workflowId} from normalized tables`)
|
logger.debug(`[${requestId}] Attempting to load workflow ${workflowId} from normalized tables`)
|
||||||
|
|||||||
@@ -13,9 +13,6 @@ export type CommandId =
|
|||||||
| 'goto-logs'
|
| 'goto-logs'
|
||||||
| 'open-search'
|
| 'open-search'
|
||||||
| 'run-workflow'
|
| 'run-workflow'
|
||||||
| 'focus-copilot-tab'
|
|
||||||
| 'focus-toolbar-tab'
|
|
||||||
| 'focus-editor-tab'
|
|
||||||
| 'clear-terminal-console'
|
| 'clear-terminal-console'
|
||||||
| 'focus-toolbar-search'
|
| 'focus-toolbar-search'
|
||||||
| 'clear-notifications'
|
| 'clear-notifications'
|
||||||
@@ -75,21 +72,6 @@ export const COMMAND_DEFINITIONS: Record<CommandId, CommandDefinition> = {
|
|||||||
shortcut: 'Mod+Enter',
|
shortcut: 'Mod+Enter',
|
||||||
allowInEditable: false,
|
allowInEditable: false,
|
||||||
},
|
},
|
||||||
'focus-copilot-tab': {
|
|
||||||
id: 'focus-copilot-tab',
|
|
||||||
shortcut: 'C',
|
|
||||||
allowInEditable: false,
|
|
||||||
},
|
|
||||||
'focus-toolbar-tab': {
|
|
||||||
id: 'focus-toolbar-tab',
|
|
||||||
shortcut: 'T',
|
|
||||||
allowInEditable: false,
|
|
||||||
},
|
|
||||||
'focus-editor-tab': {
|
|
||||||
id: 'focus-editor-tab',
|
|
||||||
shortcut: 'E',
|
|
||||||
allowInEditable: false,
|
|
||||||
},
|
|
||||||
'clear-terminal-console': {
|
'clear-terminal-console': {
|
||||||
id: 'clear-terminal-console',
|
id: 'clear-terminal-console',
|
||||||
shortcut: 'Mod+D',
|
shortcut: 'Mod+D',
|
||||||
|
|||||||
@@ -108,7 +108,7 @@ const SmoothThinkingText = memo(
|
|||||||
return (
|
return (
|
||||||
<div
|
<div
|
||||||
ref={textRef}
|
ref={textRef}
|
||||||
className='[&_*]:!text-[var(--text-muted)] [&_*]:!text-[12px] [&_*]:!leading-[1.4] [&_p]:!m-0 [&_p]:!mb-1 [&_h1]:!text-[12px] [&_h1]:!font-semibold [&_h1]:!m-0 [&_h1]:!mb-1 [&_h2]:!text-[12px] [&_h2]:!font-semibold [&_h2]:!m-0 [&_h2]:!mb-1 [&_h3]:!text-[12px] [&_h3]:!font-semibold [&_h3]:!m-0 [&_h3]:!mb-1 [&_code]:!text-[11px] [&_ul]:!pl-5 [&_ul]:!my-1 [&_ol]:!pl-8 [&_ol]:!my-1 [&_li]:!my-0.5 [&_li]:!py-0 font-season text-[12px] text-[var(--text-muted)]'
|
className='[&_*]:!text-[var(--text-muted)] [&_*]:!text-[12px] [&_*]:!leading-[1.4] [&_p]:!m-0 [&_p]:!mb-1 [&_h1]:!text-[12px] [&_h1]:!font-semibold [&_h1]:!m-0 [&_h1]:!mb-1 [&_h2]:!text-[12px] [&_h2]:!font-semibold [&_h2]:!m-0 [&_h2]:!mb-1 [&_h3]:!text-[12px] [&_h3]:!font-semibold [&_h3]:!m-0 [&_h3]:!mb-1 [&_code]:!text-[11px] [&_ul]:!pl-5 [&_ul]:!my-1 [&_ol]:!pl-6 [&_ol]:!my-1 [&_li]:!my-0.5 [&_li]:!py-0 font-season text-[12px] text-[var(--text-muted)]'
|
||||||
>
|
>
|
||||||
<CopilotMarkdownRenderer content={displayedContent} />
|
<CopilotMarkdownRenderer content={displayedContent} />
|
||||||
</div>
|
</div>
|
||||||
@@ -355,7 +355,7 @@ export function ThinkingBlock({
|
|||||||
isExpanded ? 'mt-1.5 max-h-[150px] opacity-100' : 'max-h-0 opacity-0'
|
isExpanded ? 'mt-1.5 max-h-[150px] opacity-100' : 'max-h-0 opacity-0'
|
||||||
)}
|
)}
|
||||||
>
|
>
|
||||||
<div className='[&_*]:!text-[var(--text-muted)] [&_*]:!text-[12px] [&_*]:!leading-[1.4] [&_p]:!m-0 [&_p]:!mb-1 [&_h1]:!text-[12px] [&_h1]:!font-semibold [&_h1]:!m-0 [&_h1]:!mb-1 [&_h2]:!text-[12px] [&_h2]:!font-semibold [&_h2]:!m-0 [&_h2]:!mb-1 [&_h3]:!text-[12px] [&_h3]:!font-semibold [&_h3]:!m-0 [&_h3]:!mb-1 [&_code]:!text-[11px] [&_ul]:!pl-5 [&_ul]:!my-1 [&_ol]:!pl-8 [&_ol]:!my-1 [&_li]:!my-0.5 [&_li]:!py-0 font-season text-[12px] text-[var(--text-muted)]'>
|
<div className='[&_*]:!text-[var(--text-muted)] [&_*]:!text-[12px] [&_*]:!leading-[1.4] [&_p]:!m-0 [&_p]:!mb-1 [&_h1]:!text-[12px] [&_h1]:!font-semibold [&_h1]:!m-0 [&_h1]:!mb-1 [&_h2]:!text-[12px] [&_h2]:!font-semibold [&_h2]:!m-0 [&_h2]:!mb-1 [&_h3]:!text-[12px] [&_h3]:!font-semibold [&_h3]:!m-0 [&_h3]:!mb-1 [&_code]:!text-[11px] [&_ul]:!pl-5 [&_ul]:!my-1 [&_ol]:!pl-6 [&_ol]:!my-1 [&_li]:!my-0.5 [&_li]:!py-0 font-season text-[12px] text-[var(--text-muted)]'>
|
||||||
<CopilotMarkdownRenderer content={cleanContent} />
|
<CopilotMarkdownRenderer content={cleanContent} />
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@@ -113,7 +113,7 @@ export function VersionDescriptionModal({
|
|||||||
className='min-h-[120px] resize-none'
|
className='min-h-[120px] resize-none'
|
||||||
value={description}
|
value={description}
|
||||||
onChange={(e) => setDescription(e.target.value)}
|
onChange={(e) => setDescription(e.target.value)}
|
||||||
maxLength={500}
|
maxLength={2000}
|
||||||
disabled={isGenerating}
|
disabled={isGenerating}
|
||||||
/>
|
/>
|
||||||
<div className='flex items-center justify-between'>
|
<div className='flex items-center justify-between'>
|
||||||
@@ -123,7 +123,7 @@ export function VersionDescriptionModal({
|
|||||||
</p>
|
</p>
|
||||||
)}
|
)}
|
||||||
{!updateMutation.error && !generateMutation.error && <div />}
|
{!updateMutation.error && !generateMutation.error && <div />}
|
||||||
<p className='text-[11px] text-[var(--text-tertiary)]'>{description.length}/500</p>
|
<p className='text-[11px] text-[var(--text-tertiary)]'>{description.length}/2000</p>
|
||||||
</div>
|
</div>
|
||||||
</ModalBody>
|
</ModalBody>
|
||||||
<ModalFooter>
|
<ModalFooter>
|
||||||
|
|||||||
@@ -57,6 +57,21 @@ export function useChangeDetection({
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (block.triggerMode) {
|
||||||
|
const triggerConfigValue = blockSubValues?.triggerConfig
|
||||||
|
if (
|
||||||
|
triggerConfigValue &&
|
||||||
|
typeof triggerConfigValue === 'object' &&
|
||||||
|
!subBlocks.triggerConfig
|
||||||
|
) {
|
||||||
|
subBlocks.triggerConfig = {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: triggerConfigValue,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
blocksWithSubBlocks[blockId] = {
|
blocksWithSubBlocks[blockId] = {
|
||||||
...block,
|
...block,
|
||||||
subBlocks,
|
subBlocks,
|
||||||
|
|||||||
@@ -1,7 +1,10 @@
|
|||||||
import { useCallback, useState } from 'react'
|
import { useCallback, useState } from 'react'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { runPreDeployChecks } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/deploy/hooks/use-predeploy-checks'
|
||||||
import { useNotificationStore } from '@/stores/notifications'
|
import { useNotificationStore } from '@/stores/notifications'
|
||||||
import { useWorkflowRegistry } from '@/stores/workflows/registry/store'
|
import { useWorkflowRegistry } from '@/stores/workflows/registry/store'
|
||||||
|
import { mergeSubblockState } from '@/stores/workflows/utils'
|
||||||
|
import { useWorkflowStore } from '@/stores/workflows/workflow/store'
|
||||||
|
|
||||||
const logger = createLogger('useDeployment')
|
const logger = createLogger('useDeployment')
|
||||||
|
|
||||||
@@ -35,6 +38,24 @@ export function useDeployment({
|
|||||||
return { success: true, shouldOpenModal: true }
|
return { success: true, shouldOpenModal: true }
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const { blocks, edges, loops, parallels } = useWorkflowStore.getState()
|
||||||
|
const liveBlocks = mergeSubblockState(blocks, workflowId)
|
||||||
|
const checkResult = runPreDeployChecks({
|
||||||
|
blocks: liveBlocks,
|
||||||
|
edges,
|
||||||
|
loops,
|
||||||
|
parallels,
|
||||||
|
workflowId,
|
||||||
|
})
|
||||||
|
if (!checkResult.passed) {
|
||||||
|
addNotification({
|
||||||
|
level: 'error',
|
||||||
|
message: checkResult.error || 'Pre-deploy validation failed',
|
||||||
|
workflowId,
|
||||||
|
})
|
||||||
|
return { success: false, shouldOpenModal: false }
|
||||||
|
}
|
||||||
|
|
||||||
setIsDeploying(true)
|
setIsDeploying(true)
|
||||||
try {
|
try {
|
||||||
const response = await fetch(`/api/workflows/${workflowId}/deploy`, {
|
const response = await fetch(`/api/workflows/${workflowId}/deploy`, {
|
||||||
|
|||||||
@@ -4,6 +4,7 @@ import { Button, Combobox } from '@/components/emcn/components'
|
|||||||
import {
|
import {
|
||||||
getCanonicalScopesForProvider,
|
getCanonicalScopesForProvider,
|
||||||
getProviderIdFromServiceId,
|
getProviderIdFromServiceId,
|
||||||
|
getServiceConfigByProviderId,
|
||||||
OAUTH_PROVIDERS,
|
OAUTH_PROVIDERS,
|
||||||
type OAuthProvider,
|
type OAuthProvider,
|
||||||
type OAuthService,
|
type OAuthService,
|
||||||
@@ -26,6 +27,11 @@ const getProviderIcon = (providerName: OAuthProvider) => {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const getProviderName = (providerName: OAuthProvider) => {
|
const getProviderName = (providerName: OAuthProvider) => {
|
||||||
|
const serviceConfig = getServiceConfigByProviderId(providerName)
|
||||||
|
if (serviceConfig) {
|
||||||
|
return serviceConfig.name
|
||||||
|
}
|
||||||
|
|
||||||
const { baseProvider } = parseProvider(providerName)
|
const { baseProvider } = parseProvider(providerName)
|
||||||
const baseProviderConfig = OAUTH_PROVIDERS[baseProvider]
|
const baseProviderConfig = OAUTH_PROVIDERS[baseProvider]
|
||||||
|
|
||||||
@@ -54,7 +60,7 @@ export function ToolCredentialSelector({
|
|||||||
onChange,
|
onChange,
|
||||||
provider,
|
provider,
|
||||||
requiredScopes = [],
|
requiredScopes = [],
|
||||||
label = 'Select account',
|
label,
|
||||||
serviceId,
|
serviceId,
|
||||||
disabled = false,
|
disabled = false,
|
||||||
}: ToolCredentialSelectorProps) {
|
}: ToolCredentialSelectorProps) {
|
||||||
@@ -64,6 +70,7 @@ export function ToolCredentialSelector({
|
|||||||
const { activeWorkflowId } = useWorkflowRegistry()
|
const { activeWorkflowId } = useWorkflowRegistry()
|
||||||
|
|
||||||
const selectedId = value || ''
|
const selectedId = value || ''
|
||||||
|
const effectiveLabel = label || `Select ${getProviderName(provider)} account`
|
||||||
|
|
||||||
const effectiveProviderId = useMemo(() => getProviderIdFromServiceId(serviceId), [serviceId])
|
const effectiveProviderId = useMemo(() => getProviderIdFromServiceId(serviceId), [serviceId])
|
||||||
|
|
||||||
@@ -203,7 +210,7 @@ export function ToolCredentialSelector({
|
|||||||
selectedValue={selectedId}
|
selectedValue={selectedId}
|
||||||
onChange={handleComboboxChange}
|
onChange={handleComboboxChange}
|
||||||
onOpenChange={handleOpenChange}
|
onOpenChange={handleOpenChange}
|
||||||
placeholder={label}
|
placeholder={effectiveLabel}
|
||||||
disabled={disabled}
|
disabled={disabled}
|
||||||
editable={true}
|
editable={true}
|
||||||
filterOptions={!isForeign}
|
filterOptions={!isForeign}
|
||||||
@@ -0,0 +1,186 @@
|
|||||||
|
'use client'
|
||||||
|
|
||||||
|
import type React from 'react'
|
||||||
|
import { useRef, useState } from 'react'
|
||||||
|
import { ArrowLeftRight, ArrowUp } from 'lucide-react'
|
||||||
|
import { Button, Input, Label, Tooltip } from '@/components/emcn'
|
||||||
|
import { cn } from '@/lib/core/utils/cn'
|
||||||
|
import type { WandControlHandlers } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/sub-block'
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Props for a generic parameter with label component
|
||||||
|
*/
|
||||||
|
export interface ParameterWithLabelProps {
|
||||||
|
paramId: string
|
||||||
|
title: string
|
||||||
|
isRequired: boolean
|
||||||
|
visibility: string
|
||||||
|
wandConfig?: {
|
||||||
|
enabled: boolean
|
||||||
|
prompt?: string
|
||||||
|
placeholder?: string
|
||||||
|
}
|
||||||
|
canonicalToggle?: {
|
||||||
|
mode: 'basic' | 'advanced'
|
||||||
|
disabled?: boolean
|
||||||
|
onToggle?: () => void
|
||||||
|
}
|
||||||
|
disabled: boolean
|
||||||
|
isPreview: boolean
|
||||||
|
children: (wandControlRef: React.MutableRefObject<WandControlHandlers | null>) => React.ReactNode
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Generic wrapper component for parameters that manages wand state and renders label + input
|
||||||
|
*/
|
||||||
|
export function ParameterWithLabel({
|
||||||
|
paramId,
|
||||||
|
title,
|
||||||
|
isRequired,
|
||||||
|
visibility,
|
||||||
|
wandConfig,
|
||||||
|
canonicalToggle,
|
||||||
|
disabled,
|
||||||
|
isPreview,
|
||||||
|
children,
|
||||||
|
}: ParameterWithLabelProps) {
|
||||||
|
const [isSearchActive, setIsSearchActive] = useState(false)
|
||||||
|
const [searchQuery, setSearchQuery] = useState('')
|
||||||
|
const searchInputRef = useRef<HTMLInputElement>(null)
|
||||||
|
const wandControlRef = useRef<WandControlHandlers | null>(null)
|
||||||
|
|
||||||
|
const isWandEnabled = wandConfig?.enabled ?? false
|
||||||
|
const showWand = isWandEnabled && !isPreview && !disabled
|
||||||
|
|
||||||
|
const handleSearchClick = (): void => {
|
||||||
|
setIsSearchActive(true)
|
||||||
|
setTimeout(() => {
|
||||||
|
searchInputRef.current?.focus()
|
||||||
|
}, 0)
|
||||||
|
}
|
||||||
|
|
||||||
|
const handleSearchBlur = (): void => {
|
||||||
|
if (!searchQuery.trim() && !wandControlRef.current?.isWandStreaming) {
|
||||||
|
setIsSearchActive(false)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const handleSearchChange = (value: string): void => {
|
||||||
|
setSearchQuery(value)
|
||||||
|
}
|
||||||
|
|
||||||
|
const handleSearchSubmit = (): void => {
|
||||||
|
if (searchQuery.trim() && wandControlRef.current) {
|
||||||
|
wandControlRef.current.onWandTrigger(searchQuery)
|
||||||
|
setSearchQuery('')
|
||||||
|
setIsSearchActive(false)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const handleSearchCancel = (): void => {
|
||||||
|
setSearchQuery('')
|
||||||
|
setIsSearchActive(false)
|
||||||
|
}
|
||||||
|
|
||||||
|
const isStreaming = wandControlRef.current?.isWandStreaming ?? false
|
||||||
|
|
||||||
|
return (
|
||||||
|
<div key={paramId} className='relative min-w-0 space-y-[6px]'>
|
||||||
|
<div className='flex items-center justify-between gap-[6px] pl-[2px]'>
|
||||||
|
<Label className='flex items-baseline gap-[6px] whitespace-nowrap font-medium text-[13px] text-[var(--text-primary)]'>
|
||||||
|
{title}
|
||||||
|
{isRequired && visibility === 'user-only' && <span className='ml-0.5'>*</span>}
|
||||||
|
</Label>
|
||||||
|
<div className='flex min-w-0 flex-1 items-center justify-end gap-[6px]'>
|
||||||
|
{showWand &&
|
||||||
|
(!isSearchActive ? (
|
||||||
|
<Button
|
||||||
|
variant='active'
|
||||||
|
className='-my-1 h-5 px-2 py-0 text-[11px]'
|
||||||
|
onClick={handleSearchClick}
|
||||||
|
>
|
||||||
|
Generate
|
||||||
|
</Button>
|
||||||
|
) : (
|
||||||
|
<div className='-my-1 flex min-w-[120px] max-w-[280px] flex-1 items-center gap-[4px]'>
|
||||||
|
<Input
|
||||||
|
ref={searchInputRef}
|
||||||
|
value={isStreaming ? 'Generating...' : searchQuery}
|
||||||
|
onChange={(e: React.ChangeEvent<HTMLInputElement>) =>
|
||||||
|
handleSearchChange(e.target.value)
|
||||||
|
}
|
||||||
|
onBlur={(e: React.FocusEvent<HTMLInputElement>) => {
|
||||||
|
const relatedTarget = e.relatedTarget as HTMLElement | null
|
||||||
|
if (relatedTarget?.closest('button')) return
|
||||||
|
handleSearchBlur()
|
||||||
|
}}
|
||||||
|
onKeyDown={(e: React.KeyboardEvent<HTMLInputElement>) => {
|
||||||
|
if (e.key === 'Enter' && searchQuery.trim() && !isStreaming) {
|
||||||
|
handleSearchSubmit()
|
||||||
|
} else if (e.key === 'Escape') {
|
||||||
|
handleSearchCancel()
|
||||||
|
}
|
||||||
|
}}
|
||||||
|
disabled={isStreaming}
|
||||||
|
className={cn(
|
||||||
|
'h-5 min-w-[80px] flex-1 text-[11px]',
|
||||||
|
isStreaming && 'text-muted-foreground'
|
||||||
|
)}
|
||||||
|
placeholder='Generate with AI...'
|
||||||
|
/>
|
||||||
|
<Button
|
||||||
|
variant='tertiary'
|
||||||
|
disabled={!searchQuery.trim() || isStreaming}
|
||||||
|
onMouseDown={(e: React.MouseEvent) => {
|
||||||
|
e.preventDefault()
|
||||||
|
e.stopPropagation()
|
||||||
|
}}
|
||||||
|
onClick={(e: React.MouseEvent) => {
|
||||||
|
e.stopPropagation()
|
||||||
|
handleSearchSubmit()
|
||||||
|
}}
|
||||||
|
className='h-[20px] w-[20px] flex-shrink-0 p-0'
|
||||||
|
>
|
||||||
|
<ArrowUp className='h-[12px] w-[12px]' />
|
||||||
|
</Button>
|
||||||
|
</div>
|
||||||
|
))}
|
||||||
|
{canonicalToggle && !isPreview && (
|
||||||
|
<Tooltip.Root>
|
||||||
|
<Tooltip.Trigger asChild>
|
||||||
|
<button
|
||||||
|
type='button'
|
||||||
|
className='flex h-[12px] w-[12px] flex-shrink-0 items-center justify-center bg-transparent p-0 disabled:cursor-not-allowed disabled:opacity-50'
|
||||||
|
onClick={canonicalToggle.onToggle}
|
||||||
|
disabled={canonicalToggle.disabled || disabled}
|
||||||
|
aria-label={
|
||||||
|
canonicalToggle.mode === 'advanced'
|
||||||
|
? 'Switch to selector'
|
||||||
|
: 'Switch to manual ID'
|
||||||
|
}
|
||||||
|
>
|
||||||
|
<ArrowLeftRight
|
||||||
|
className={cn(
|
||||||
|
'!h-[12px] !w-[12px]',
|
||||||
|
canonicalToggle.mode === 'advanced'
|
||||||
|
? 'text-[var(--text-primary)]'
|
||||||
|
: 'text-[var(--text-secondary)]'
|
||||||
|
)}
|
||||||
|
/>
|
||||||
|
</button>
|
||||||
|
</Tooltip.Trigger>
|
||||||
|
<Tooltip.Content side='top'>
|
||||||
|
<p>
|
||||||
|
{canonicalToggle.mode === 'advanced'
|
||||||
|
? 'Switch to selector'
|
||||||
|
: 'Switch to manual ID'}
|
||||||
|
</p>
|
||||||
|
</Tooltip.Content>
|
||||||
|
</Tooltip.Root>
|
||||||
|
)}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
<div className='relative w-full min-w-0'>{children(wandControlRef)}</div>
|
||||||
|
</div>
|
||||||
|
)
|
||||||
|
}
|
||||||
@@ -0,0 +1,114 @@
|
|||||||
|
'use client'
|
||||||
|
|
||||||
|
import { useEffect, useRef } from 'react'
|
||||||
|
import { useSubBlockValue } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/hooks/use-sub-block-value'
|
||||||
|
import { SubBlock } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/sub-block'
|
||||||
|
import type { SubBlockConfig as BlockSubBlockConfig } from '@/blocks/types'
|
||||||
|
|
||||||
|
interface ToolSubBlockRendererProps {
|
||||||
|
blockId: string
|
||||||
|
subBlockId: string
|
||||||
|
toolIndex: number
|
||||||
|
subBlock: BlockSubBlockConfig
|
||||||
|
effectiveParamId: string
|
||||||
|
toolParams: Record<string, string> | undefined
|
||||||
|
onParamChange: (toolIndex: number, paramId: string, value: string) => void
|
||||||
|
disabled: boolean
|
||||||
|
canonicalToggle?: {
|
||||||
|
mode: 'basic' | 'advanced'
|
||||||
|
disabled?: boolean
|
||||||
|
onToggle?: () => void
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* SubBlock types whose store values are objects/arrays/non-strings.
|
||||||
|
* tool.params stores strings (via JSON.stringify), so when syncing
|
||||||
|
* back to the store we parse them to restore the native shape.
|
||||||
|
*/
|
||||||
|
const OBJECT_SUBBLOCK_TYPES = new Set(['file-upload', 'table', 'grouped-checkbox-list'])
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Bridges the subblock store with StoredTool.params via a synthetic store key,
|
||||||
|
* then delegates all rendering to SubBlock for full parity.
|
||||||
|
*/
|
||||||
|
export function ToolSubBlockRenderer({
|
||||||
|
blockId,
|
||||||
|
subBlockId,
|
||||||
|
toolIndex,
|
||||||
|
subBlock,
|
||||||
|
effectiveParamId,
|
||||||
|
toolParams,
|
||||||
|
onParamChange,
|
||||||
|
disabled,
|
||||||
|
canonicalToggle,
|
||||||
|
}: ToolSubBlockRendererProps) {
|
||||||
|
const syntheticId = `${subBlockId}-tool-${toolIndex}-${effectiveParamId}`
|
||||||
|
const [storeValue, setStoreValue] = useSubBlockValue(blockId, syntheticId)
|
||||||
|
|
||||||
|
const toolParamValue = toolParams?.[effectiveParamId] ?? ''
|
||||||
|
const isObjectType = OBJECT_SUBBLOCK_TYPES.has(subBlock.type)
|
||||||
|
|
||||||
|
const lastPushedToStoreRef = useRef<string | null>(null)
|
||||||
|
const lastPushedToParamsRef = useRef<string | null>(null)
|
||||||
|
|
||||||
|
useEffect(() => {
|
||||||
|
if (!toolParamValue && lastPushedToStoreRef.current === null) {
|
||||||
|
lastPushedToStoreRef.current = toolParamValue
|
||||||
|
lastPushedToParamsRef.current = toolParamValue
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if (toolParamValue !== lastPushedToStoreRef.current) {
|
||||||
|
lastPushedToStoreRef.current = toolParamValue
|
||||||
|
lastPushedToParamsRef.current = toolParamValue
|
||||||
|
|
||||||
|
if (isObjectType && typeof toolParamValue === 'string' && toolParamValue) {
|
||||||
|
try {
|
||||||
|
const parsed = JSON.parse(toolParamValue)
|
||||||
|
if (typeof parsed === 'object' && parsed !== null) {
|
||||||
|
setStoreValue(parsed)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
} catch {
|
||||||
|
// Not valid JSON — fall through to set as string
|
||||||
|
}
|
||||||
|
}
|
||||||
|
setStoreValue(toolParamValue)
|
||||||
|
}
|
||||||
|
}, [toolParamValue, setStoreValue, isObjectType])
|
||||||
|
|
||||||
|
useEffect(() => {
|
||||||
|
if (storeValue == null && lastPushedToParamsRef.current === null) return
|
||||||
|
const stringValue =
|
||||||
|
storeValue == null
|
||||||
|
? ''
|
||||||
|
: typeof storeValue === 'string'
|
||||||
|
? storeValue
|
||||||
|
: JSON.stringify(storeValue)
|
||||||
|
if (stringValue !== lastPushedToParamsRef.current) {
|
||||||
|
lastPushedToParamsRef.current = stringValue
|
||||||
|
lastPushedToStoreRef.current = stringValue
|
||||||
|
onParamChange(toolIndex, effectiveParamId, stringValue)
|
||||||
|
}
|
||||||
|
}, [storeValue, toolIndex, effectiveParamId, onParamChange])
|
||||||
|
|
||||||
|
const visibility = subBlock.paramVisibility ?? 'user-or-llm'
|
||||||
|
const isOptionalForUser = visibility !== 'user-only'
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
...subBlock,
|
||||||
|
id: syntheticId,
|
||||||
|
...(isOptionalForUser && { required: false }),
|
||||||
|
}
|
||||||
|
|
||||||
|
return (
|
||||||
|
<SubBlock
|
||||||
|
blockId={blockId}
|
||||||
|
config={config}
|
||||||
|
isPreview={false}
|
||||||
|
disabled={disabled}
|
||||||
|
canonicalToggle={canonicalToggle}
|
||||||
|
dependencyContext={toolParams}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
@@ -2,37 +2,12 @@
|
|||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
import { describe, expect, it } from 'vitest'
|
import { describe, expect, it } from 'vitest'
|
||||||
|
import type { StoredTool } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/components/tool-input/types'
|
||||||
interface StoredTool {
|
import {
|
||||||
type: string
|
isCustomToolAlreadySelected,
|
||||||
title?: string
|
isMcpToolAlreadySelected,
|
||||||
toolId?: string
|
isWorkflowAlreadySelected,
|
||||||
params?: Record<string, string>
|
} from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/components/tool-input/utils'
|
||||||
customToolId?: string
|
|
||||||
schema?: any
|
|
||||||
code?: string
|
|
||||||
operation?: string
|
|
||||||
usageControl?: 'auto' | 'force' | 'none'
|
|
||||||
}
|
|
||||||
|
|
||||||
const isMcpToolAlreadySelected = (selectedTools: StoredTool[], mcpToolId: string): boolean => {
|
|
||||||
return selectedTools.some((tool) => tool.type === 'mcp' && tool.toolId === mcpToolId)
|
|
||||||
}
|
|
||||||
|
|
||||||
const isCustomToolAlreadySelected = (
|
|
||||||
selectedTools: StoredTool[],
|
|
||||||
customToolId: string
|
|
||||||
): boolean => {
|
|
||||||
return selectedTools.some(
|
|
||||||
(tool) => tool.type === 'custom-tool' && tool.customToolId === customToolId
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
const isWorkflowAlreadySelected = (selectedTools: StoredTool[], workflowId: string): boolean => {
|
|
||||||
return selectedTools.some(
|
|
||||||
(tool) => tool.type === 'workflow_input' && tool.params?.workflowId === workflowId
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
describe('isMcpToolAlreadySelected', () => {
|
describe('isMcpToolAlreadySelected', () => {
|
||||||
describe('basic functionality', () => {
|
describe('basic functionality', () => {
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -0,0 +1,31 @@
|
|||||||
|
/**
|
||||||
|
* Represents a tool selected and configured in the workflow
|
||||||
|
*
|
||||||
|
* @remarks
|
||||||
|
* For custom tools (new format), we only store: type, customToolId, usageControl, isExpanded.
|
||||||
|
* Everything else (title, schema, code) is loaded dynamically from the database.
|
||||||
|
* Legacy custom tools with inline schema/code are still supported for backwards compatibility.
|
||||||
|
*/
|
||||||
|
export interface StoredTool {
|
||||||
|
/** Block type identifier */
|
||||||
|
type: string
|
||||||
|
/** Display title for the tool (optional for new custom tool format) */
|
||||||
|
title?: string
|
||||||
|
/** Direct tool ID for execution (optional for new custom tool format) */
|
||||||
|
toolId?: string
|
||||||
|
/** Parameter values configured by the user (optional for new custom tool format) */
|
||||||
|
params?: Record<string, string>
|
||||||
|
/** Whether the tool details are expanded in UI */
|
||||||
|
isExpanded?: boolean
|
||||||
|
/** Database ID for custom tools (new format - reference only) */
|
||||||
|
customToolId?: string
|
||||||
|
/** Tool schema for custom tools (legacy format - inline JSON schema) */
|
||||||
|
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
||||||
|
schema?: Record<string, any>
|
||||||
|
/** Implementation code for custom tools (legacy format - inline) */
|
||||||
|
code?: string
|
||||||
|
/** Selected operation for multi-operation tools */
|
||||||
|
operation?: string
|
||||||
|
/** Tool usage control mode for LLM */
|
||||||
|
usageControl?: 'auto' | 'force' | 'none'
|
||||||
|
}
|
||||||
@@ -0,0 +1,32 @@
|
|||||||
|
import type { StoredTool } from '@/app/workspace/[workspaceId]/w/[workflowId]/components/panel/components/editor/components/sub-block/components/tool-input/types'
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Checks if an MCP tool is already selected.
|
||||||
|
*/
|
||||||
|
export function isMcpToolAlreadySelected(selectedTools: StoredTool[], mcpToolId: string): boolean {
|
||||||
|
return selectedTools.some((tool) => tool.type === 'mcp' && tool.toolId === mcpToolId)
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Checks if a custom tool is already selected.
|
||||||
|
*/
|
||||||
|
export function isCustomToolAlreadySelected(
|
||||||
|
selectedTools: StoredTool[],
|
||||||
|
customToolId: string
|
||||||
|
): boolean {
|
||||||
|
return selectedTools.some(
|
||||||
|
(tool) => tool.type === 'custom-tool' && tool.customToolId === customToolId
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Checks if a workflow is already selected.
|
||||||
|
*/
|
||||||
|
export function isWorkflowAlreadySelected(
|
||||||
|
selectedTools: StoredTool[],
|
||||||
|
workflowId: string
|
||||||
|
): boolean {
|
||||||
|
return selectedTools.some(
|
||||||
|
(tool) => tool.type === 'workflow_input' && tool.params?.workflowId === workflowId
|
||||||
|
)
|
||||||
|
}
|
||||||
@@ -3,7 +3,6 @@ import { isEqual } from 'lodash'
|
|||||||
import { AlertTriangle, ArrowLeftRight, ArrowUp, Check, Clipboard } from 'lucide-react'
|
import { AlertTriangle, ArrowLeftRight, ArrowUp, Check, Clipboard } from 'lucide-react'
|
||||||
import { Button, Input, Label, Tooltip } from '@/components/emcn/components'
|
import { Button, Input, Label, Tooltip } from '@/components/emcn/components'
|
||||||
import { cn } from '@/lib/core/utils/cn'
|
import { cn } from '@/lib/core/utils/cn'
|
||||||
import type { FieldDiffStatus } from '@/lib/workflows/diff/types'
|
|
||||||
import {
|
import {
|
||||||
CheckboxList,
|
CheckboxList,
|
||||||
Code,
|
Code,
|
||||||
@@ -69,13 +68,15 @@ interface SubBlockProps {
|
|||||||
isPreview?: boolean
|
isPreview?: boolean
|
||||||
subBlockValues?: Record<string, any>
|
subBlockValues?: Record<string, any>
|
||||||
disabled?: boolean
|
disabled?: boolean
|
||||||
fieldDiffStatus?: FieldDiffStatus
|
|
||||||
allowExpandInPreview?: boolean
|
allowExpandInPreview?: boolean
|
||||||
canonicalToggle?: {
|
canonicalToggle?: {
|
||||||
mode: 'basic' | 'advanced'
|
mode: 'basic' | 'advanced'
|
||||||
disabled?: boolean
|
disabled?: boolean
|
||||||
onToggle?: () => void
|
onToggle?: () => void
|
||||||
}
|
}
|
||||||
|
labelSuffix?: React.ReactNode
|
||||||
|
/** Provides sibling values for dependency resolution in non-preview contexts (e.g. tool-input) */
|
||||||
|
dependencyContext?: Record<string, unknown>
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -162,16 +163,14 @@ const getPreviewValue = (
|
|||||||
/**
|
/**
|
||||||
* Renders the label with optional validation and description tooltips.
|
* Renders the label with optional validation and description tooltips.
|
||||||
*
|
*
|
||||||
* @remarks
|
|
||||||
* Handles JSON validation indicators for code blocks and required field markers.
|
|
||||||
* Includes inline AI generate button when wand is enabled.
|
|
||||||
*
|
|
||||||
* @param config - The sub-block configuration defining the label content
|
* @param config - The sub-block configuration defining the label content
|
||||||
* @param isValidJson - Whether the JSON content is valid (for code blocks)
|
* @param isValidJson - Whether the JSON content is valid (for code blocks)
|
||||||
* @param subBlockValues - Current values of all subblocks for evaluating conditional requirements
|
* @param subBlockValues - Current values of all subblocks for evaluating conditional requirements
|
||||||
* @param wandState - Optional state and handlers for the AI wand feature
|
* @param wandState - State and handlers for the inline AI generate feature
|
||||||
* @param canonicalToggle - Optional canonical toggle metadata and handlers
|
* @param canonicalToggle - Metadata and handlers for the basic/advanced mode toggle
|
||||||
* @param canonicalToggleIsDisabled - Whether the canonical toggle is disabled
|
* @param canonicalToggleIsDisabled - Whether the canonical toggle is disabled (includes dependsOn gating)
|
||||||
|
* @param copyState - State and handler for the copy-to-clipboard button
|
||||||
|
* @param labelSuffix - Additional content rendered after the label text
|
||||||
* @returns The label JSX element, or `null` for switch types or when no title is defined
|
* @returns The label JSX element, or `null` for switch types or when no title is defined
|
||||||
*/
|
*/
|
||||||
const renderLabel = (
|
const renderLabel = (
|
||||||
@@ -202,7 +201,8 @@ const renderLabel = (
|
|||||||
showCopyButton: boolean
|
showCopyButton: boolean
|
||||||
copied: boolean
|
copied: boolean
|
||||||
onCopy: () => void
|
onCopy: () => void
|
||||||
}
|
},
|
||||||
|
labelSuffix?: React.ReactNode
|
||||||
): JSX.Element | null => {
|
): JSX.Element | null => {
|
||||||
if (config.type === 'switch') return null
|
if (config.type === 'switch') return null
|
||||||
if (!config.title) return null
|
if (!config.title) return null
|
||||||
@@ -215,9 +215,10 @@ const renderLabel = (
|
|||||||
|
|
||||||
return (
|
return (
|
||||||
<div className='flex items-center justify-between gap-[6px] pl-[2px]'>
|
<div className='flex items-center justify-between gap-[6px] pl-[2px]'>
|
||||||
<Label className='flex items-center gap-[6px] whitespace-nowrap'>
|
<Label className='flex items-baseline gap-[6px] whitespace-nowrap'>
|
||||||
{config.title}
|
{config.title}
|
||||||
{required && <span className='ml-0.5'>*</span>}
|
{required && <span className='ml-0.5'>*</span>}
|
||||||
|
{labelSuffix}
|
||||||
{config.type === 'code' &&
|
{config.type === 'code' &&
|
||||||
config.language === 'json' &&
|
config.language === 'json' &&
|
||||||
!isValidJson &&
|
!isValidJson &&
|
||||||
@@ -383,28 +384,25 @@ const arePropsEqual = (prevProps: SubBlockProps, nextProps: SubBlockProps): bool
|
|||||||
prevProps.isPreview === nextProps.isPreview &&
|
prevProps.isPreview === nextProps.isPreview &&
|
||||||
valueEqual &&
|
valueEqual &&
|
||||||
prevProps.disabled === nextProps.disabled &&
|
prevProps.disabled === nextProps.disabled &&
|
||||||
prevProps.fieldDiffStatus === nextProps.fieldDiffStatus &&
|
|
||||||
prevProps.allowExpandInPreview === nextProps.allowExpandInPreview &&
|
prevProps.allowExpandInPreview === nextProps.allowExpandInPreview &&
|
||||||
canonicalToggleEqual
|
canonicalToggleEqual &&
|
||||||
|
prevProps.labelSuffix === nextProps.labelSuffix &&
|
||||||
|
prevProps.dependencyContext === nextProps.dependencyContext
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Renders a single workflow sub-block input based on config.type.
|
* Renders a single workflow sub-block input based on config.type.
|
||||||
*
|
*
|
||||||
* @remarks
|
|
||||||
* Supports multiple input types including short-input, long-input, dropdown,
|
|
||||||
* combobox, slider, table, code, switch, tool-input, and many more.
|
|
||||||
* Handles preview mode, disabled states, and AI wand generation.
|
|
||||||
*
|
|
||||||
* @param blockId - The parent block identifier
|
* @param blockId - The parent block identifier
|
||||||
* @param config - Configuration defining the input type and properties
|
* @param config - Configuration defining the input type and properties
|
||||||
* @param isPreview - Whether to render in preview mode
|
* @param isPreview - Whether to render in preview mode
|
||||||
* @param subBlockValues - Current values of all subblocks
|
* @param subBlockValues - Current values of all subblocks
|
||||||
* @param disabled - Whether the input is disabled
|
* @param disabled - Whether the input is disabled
|
||||||
* @param fieldDiffStatus - Optional diff status for visual indicators
|
|
||||||
* @param allowExpandInPreview - Whether to allow expanding in preview mode
|
* @param allowExpandInPreview - Whether to allow expanding in preview mode
|
||||||
* @returns The rendered sub-block input component
|
* @param canonicalToggle - Metadata and handlers for the basic/advanced mode toggle
|
||||||
|
* @param labelSuffix - Additional content rendered after the label text
|
||||||
|
* @param dependencyContext - Sibling values for dependency resolution in non-preview contexts (e.g. tool-input)
|
||||||
*/
|
*/
|
||||||
function SubBlockComponent({
|
function SubBlockComponent({
|
||||||
blockId,
|
blockId,
|
||||||
@@ -412,9 +410,10 @@ function SubBlockComponent({
|
|||||||
isPreview = false,
|
isPreview = false,
|
||||||
subBlockValues,
|
subBlockValues,
|
||||||
disabled = false,
|
disabled = false,
|
||||||
fieldDiffStatus,
|
|
||||||
allowExpandInPreview,
|
allowExpandInPreview,
|
||||||
canonicalToggle,
|
canonicalToggle,
|
||||||
|
labelSuffix,
|
||||||
|
dependencyContext,
|
||||||
}: SubBlockProps): JSX.Element {
|
}: SubBlockProps): JSX.Element {
|
||||||
const [isValidJson, setIsValidJson] = useState(true)
|
const [isValidJson, setIsValidJson] = useState(true)
|
||||||
const [isSearchActive, setIsSearchActive] = useState(false)
|
const [isSearchActive, setIsSearchActive] = useState(false)
|
||||||
@@ -423,7 +422,6 @@ function SubBlockComponent({
|
|||||||
const searchInputRef = useRef<HTMLInputElement>(null)
|
const searchInputRef = useRef<HTMLInputElement>(null)
|
||||||
const wandControlRef = useRef<WandControlHandlers | null>(null)
|
const wandControlRef = useRef<WandControlHandlers | null>(null)
|
||||||
|
|
||||||
// Use webhook management hook when config has useWebhookUrl enabled
|
|
||||||
const webhookManagement = useWebhookManagement({
|
const webhookManagement = useWebhookManagement({
|
||||||
blockId,
|
blockId,
|
||||||
triggerId: undefined,
|
triggerId: undefined,
|
||||||
@@ -510,10 +508,12 @@ function SubBlockComponent({
|
|||||||
| null
|
| null
|
||||||
| undefined
|
| undefined
|
||||||
|
|
||||||
|
const contextValues = dependencyContext ?? (isPreview ? subBlockValues : undefined)
|
||||||
|
|
||||||
const { finalDisabled: gatedDisabled } = useDependsOnGate(blockId, config, {
|
const { finalDisabled: gatedDisabled } = useDependsOnGate(blockId, config, {
|
||||||
disabled,
|
disabled,
|
||||||
isPreview,
|
isPreview,
|
||||||
previewContextValues: isPreview ? subBlockValues : undefined,
|
previewContextValues: contextValues,
|
||||||
})
|
})
|
||||||
|
|
||||||
const isDisabled = gatedDisabled
|
const isDisabled = gatedDisabled
|
||||||
@@ -797,7 +797,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue}
|
previewValue={previewValue}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -809,7 +809,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue}
|
previewValue={previewValue}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -821,7 +821,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue}
|
previewValue={previewValue}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -833,7 +833,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue}
|
previewValue={previewValue}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -845,7 +845,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue}
|
previewValue={previewValue}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -868,7 +868,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue as any}
|
previewValue={previewValue as any}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -880,7 +880,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue as any}
|
previewValue={previewValue as any}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -892,7 +892,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue as any}
|
previewValue={previewValue as any}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -917,7 +917,7 @@ function SubBlockComponent({
|
|||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue as any}
|
previewValue={previewValue as any}
|
||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -953,7 +953,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue}
|
previewValue={previewValue}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -987,7 +987,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue as any}
|
previewValue={previewValue as any}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -999,7 +999,7 @@ function SubBlockComponent({
|
|||||||
disabled={isDisabled}
|
disabled={isDisabled}
|
||||||
isPreview={isPreview}
|
isPreview={isPreview}
|
||||||
previewValue={previewValue}
|
previewValue={previewValue}
|
||||||
previewContextValues={isPreview ? subBlockValues : undefined}
|
previewContextValues={contextValues}
|
||||||
/>
|
/>
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -1059,7 +1059,8 @@ function SubBlockComponent({
|
|||||||
showCopyButton: Boolean(config.showCopyButton && config.useWebhookUrl),
|
showCopyButton: Boolean(config.showCopyButton && config.useWebhookUrl),
|
||||||
copied,
|
copied,
|
||||||
onCopy: handleCopy,
|
onCopy: handleCopy,
|
||||||
}
|
},
|
||||||
|
labelSuffix
|
||||||
)}
|
)}
|
||||||
{renderInput()}
|
{renderInput()}
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@@ -571,7 +571,6 @@ export function Editor() {
|
|||||||
isPreview={false}
|
isPreview={false}
|
||||||
subBlockValues={subBlockState}
|
subBlockValues={subBlockState}
|
||||||
disabled={!canEditBlock}
|
disabled={!canEditBlock}
|
||||||
fieldDiffStatus={undefined}
|
|
||||||
allowExpandInPreview={false}
|
allowExpandInPreview={false}
|
||||||
canonicalToggle={
|
canonicalToggle={
|
||||||
isCanonicalSwap && canonicalMode && canonicalId
|
isCanonicalSwap && canonicalMode && canonicalId
|
||||||
@@ -635,7 +634,6 @@ export function Editor() {
|
|||||||
isPreview={false}
|
isPreview={false}
|
||||||
subBlockValues={subBlockState}
|
subBlockValues={subBlockState}
|
||||||
disabled={!canEditBlock}
|
disabled={!canEditBlock}
|
||||||
fieldDiffStatus={undefined}
|
|
||||||
allowExpandInPreview={false}
|
allowExpandInPreview={false}
|
||||||
/>
|
/>
|
||||||
{index < advancedOnlySubBlocks.length - 1 && (
|
{index < advancedOnlySubBlocks.length - 1 && (
|
||||||
|
|||||||
@@ -340,13 +340,7 @@ export const Panel = memo(function Panel() {
|
|||||||
* Register global keyboard shortcuts using the central commands registry.
|
* Register global keyboard shortcuts using the central commands registry.
|
||||||
*
|
*
|
||||||
* - Mod+Enter: Run / cancel workflow (matches the Run button behavior)
|
* - Mod+Enter: Run / cancel workflow (matches the Run button behavior)
|
||||||
* - C: Focus Copilot tab
|
|
||||||
* - T: Focus Toolbar tab
|
|
||||||
* - E: Focus Editor tab
|
|
||||||
* - Mod+F: Focus Toolbar tab and search input
|
* - Mod+F: Focus Toolbar tab and search input
|
||||||
*
|
|
||||||
* The tab-switching commands are disabled inside editable elements so typing
|
|
||||||
* in inputs or textareas is not interrupted.
|
|
||||||
*/
|
*/
|
||||||
useRegisterGlobalCommands(() =>
|
useRegisterGlobalCommands(() =>
|
||||||
createCommands([
|
createCommands([
|
||||||
@@ -363,33 +357,6 @@ export const Panel = memo(function Panel() {
|
|||||||
allowInEditable: false,
|
allowInEditable: false,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
|
||||||
id: 'focus-copilot-tab',
|
|
||||||
handler: () => {
|
|
||||||
setActiveTab('copilot')
|
|
||||||
},
|
|
||||||
overrides: {
|
|
||||||
allowInEditable: false,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
id: 'focus-toolbar-tab',
|
|
||||||
handler: () => {
|
|
||||||
setActiveTab('toolbar')
|
|
||||||
},
|
|
||||||
overrides: {
|
|
||||||
allowInEditable: false,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
id: 'focus-editor-tab',
|
|
||||||
handler: () => {
|
|
||||||
setActiveTab('editor')
|
|
||||||
},
|
|
||||||
overrides: {
|
|
||||||
allowInEditable: false,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
id: 'focus-toolbar-search',
|
id: 'focus-toolbar-search',
|
||||||
handler: () => {
|
handler: () => {
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { useCallback, useRef, useState } from 'react'
|
import { useCallback, useEffect, useRef, useState } from 'react'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { useQueryClient } from '@tanstack/react-query'
|
import { useQueryClient } from '@tanstack/react-query'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
@@ -46,7 +46,13 @@ import { useWorkflowStore } from '@/stores/workflows/workflow/store'
|
|||||||
|
|
||||||
const logger = createLogger('useWorkflowExecution')
|
const logger = createLogger('useWorkflowExecution')
|
||||||
|
|
||||||
// Debug state validation result
|
/**
|
||||||
|
* Module-level Set tracking which workflows have an active reconnection effect.
|
||||||
|
* Prevents multiple hook instances (from different components) from starting
|
||||||
|
* concurrent reconnection streams for the same workflow during the same mount cycle.
|
||||||
|
*/
|
||||||
|
const activeReconnections = new Set<string>()
|
||||||
|
|
||||||
interface DebugValidationResult {
|
interface DebugValidationResult {
|
||||||
isValid: boolean
|
isValid: boolean
|
||||||
error?: string
|
error?: string
|
||||||
@@ -54,7 +60,7 @@ interface DebugValidationResult {
|
|||||||
|
|
||||||
interface BlockEventHandlerConfig {
|
interface BlockEventHandlerConfig {
|
||||||
workflowId?: string
|
workflowId?: string
|
||||||
executionId?: string
|
executionIdRef: { current: string }
|
||||||
workflowEdges: Array<{ id: string; target: string; sourceHandle?: string | null }>
|
workflowEdges: Array<{ id: string; target: string; sourceHandle?: string | null }>
|
||||||
activeBlocksSet: Set<string>
|
activeBlocksSet: Set<string>
|
||||||
accumulatedBlockLogs: BlockLog[]
|
accumulatedBlockLogs: BlockLog[]
|
||||||
@@ -108,12 +114,15 @@ export function useWorkflowExecution() {
|
|||||||
const queryClient = useQueryClient()
|
const queryClient = useQueryClient()
|
||||||
const currentWorkflow = useCurrentWorkflow()
|
const currentWorkflow = useCurrentWorkflow()
|
||||||
const { activeWorkflowId, workflows } = useWorkflowRegistry()
|
const { activeWorkflowId, workflows } = useWorkflowRegistry()
|
||||||
const { toggleConsole, addConsole, updateConsole, cancelRunningEntries } =
|
const { toggleConsole, addConsole, updateConsole, cancelRunningEntries, clearExecutionEntries } =
|
||||||
useTerminalConsoleStore()
|
useTerminalConsoleStore()
|
||||||
|
const hasHydrated = useTerminalConsoleStore((s) => s._hasHydrated)
|
||||||
const { getAllVariables } = useEnvironmentStore()
|
const { getAllVariables } = useEnvironmentStore()
|
||||||
const { getVariablesByWorkflowId, variables } = useVariablesStore()
|
const { getVariablesByWorkflowId, variables } = useVariablesStore()
|
||||||
const { isExecuting, isDebugging, pendingBlocks, executor, debugContext } =
|
const { isExecuting, isDebugging, pendingBlocks, executor, debugContext } =
|
||||||
useCurrentWorkflowExecution()
|
useCurrentWorkflowExecution()
|
||||||
|
const setCurrentExecutionId = useExecutionStore((s) => s.setCurrentExecutionId)
|
||||||
|
const getCurrentExecutionId = useExecutionStore((s) => s.getCurrentExecutionId)
|
||||||
const setIsExecuting = useExecutionStore((s) => s.setIsExecuting)
|
const setIsExecuting = useExecutionStore((s) => s.setIsExecuting)
|
||||||
const setIsDebugging = useExecutionStore((s) => s.setIsDebugging)
|
const setIsDebugging = useExecutionStore((s) => s.setIsDebugging)
|
||||||
const setPendingBlocks = useExecutionStore((s) => s.setPendingBlocks)
|
const setPendingBlocks = useExecutionStore((s) => s.setPendingBlocks)
|
||||||
@@ -297,7 +306,7 @@ export function useWorkflowExecution() {
|
|||||||
(config: BlockEventHandlerConfig) => {
|
(config: BlockEventHandlerConfig) => {
|
||||||
const {
|
const {
|
||||||
workflowId,
|
workflowId,
|
||||||
executionId,
|
executionIdRef,
|
||||||
workflowEdges,
|
workflowEdges,
|
||||||
activeBlocksSet,
|
activeBlocksSet,
|
||||||
accumulatedBlockLogs,
|
accumulatedBlockLogs,
|
||||||
@@ -308,6 +317,14 @@ export function useWorkflowExecution() {
|
|||||||
onBlockCompleteCallback,
|
onBlockCompleteCallback,
|
||||||
} = config
|
} = config
|
||||||
|
|
||||||
|
/** Returns true if this execution was cancelled or superseded by another run. */
|
||||||
|
const isStaleExecution = () =>
|
||||||
|
!!(
|
||||||
|
workflowId &&
|
||||||
|
executionIdRef.current &&
|
||||||
|
useExecutionStore.getState().getCurrentExecutionId(workflowId) !== executionIdRef.current
|
||||||
|
)
|
||||||
|
|
||||||
const updateActiveBlocks = (blockId: string, isActive: boolean) => {
|
const updateActiveBlocks = (blockId: string, isActive: boolean) => {
|
||||||
if (!workflowId) return
|
if (!workflowId) return
|
||||||
if (isActive) {
|
if (isActive) {
|
||||||
@@ -360,7 +377,7 @@ export function useWorkflowExecution() {
|
|||||||
endedAt: data.endedAt,
|
endedAt: data.endedAt,
|
||||||
workflowId,
|
workflowId,
|
||||||
blockId: data.blockId,
|
blockId: data.blockId,
|
||||||
executionId,
|
executionId: executionIdRef.current,
|
||||||
blockName: data.blockName || 'Unknown Block',
|
blockName: data.blockName || 'Unknown Block',
|
||||||
blockType: data.blockType || 'unknown',
|
blockType: data.blockType || 'unknown',
|
||||||
iterationCurrent: data.iterationCurrent,
|
iterationCurrent: data.iterationCurrent,
|
||||||
@@ -383,7 +400,7 @@ export function useWorkflowExecution() {
|
|||||||
endedAt: data.endedAt,
|
endedAt: data.endedAt,
|
||||||
workflowId,
|
workflowId,
|
||||||
blockId: data.blockId,
|
blockId: data.blockId,
|
||||||
executionId,
|
executionId: executionIdRef.current,
|
||||||
blockName: data.blockName || 'Unknown Block',
|
blockName: data.blockName || 'Unknown Block',
|
||||||
blockType: data.blockType || 'unknown',
|
blockType: data.blockType || 'unknown',
|
||||||
iterationCurrent: data.iterationCurrent,
|
iterationCurrent: data.iterationCurrent,
|
||||||
@@ -410,7 +427,7 @@ export function useWorkflowExecution() {
|
|||||||
iterationType: data.iterationType,
|
iterationType: data.iterationType,
|
||||||
iterationContainerId: data.iterationContainerId,
|
iterationContainerId: data.iterationContainerId,
|
||||||
},
|
},
|
||||||
executionId
|
executionIdRef.current
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -432,11 +449,12 @@ export function useWorkflowExecution() {
|
|||||||
iterationType: data.iterationType,
|
iterationType: data.iterationType,
|
||||||
iterationContainerId: data.iterationContainerId,
|
iterationContainerId: data.iterationContainerId,
|
||||||
},
|
},
|
||||||
executionId
|
executionIdRef.current
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
const onBlockStarted = (data: BlockStartedData) => {
|
const onBlockStarted = (data: BlockStartedData) => {
|
||||||
|
if (isStaleExecution()) return
|
||||||
updateActiveBlocks(data.blockId, true)
|
updateActiveBlocks(data.blockId, true)
|
||||||
markIncomingEdges(data.blockId)
|
markIncomingEdges(data.blockId)
|
||||||
|
|
||||||
@@ -453,7 +471,7 @@ export function useWorkflowExecution() {
|
|||||||
endedAt: undefined,
|
endedAt: undefined,
|
||||||
workflowId,
|
workflowId,
|
||||||
blockId: data.blockId,
|
blockId: data.blockId,
|
||||||
executionId,
|
executionId: executionIdRef.current,
|
||||||
blockName: data.blockName || 'Unknown Block',
|
blockName: data.blockName || 'Unknown Block',
|
||||||
blockType: data.blockType || 'unknown',
|
blockType: data.blockType || 'unknown',
|
||||||
isRunning: true,
|
isRunning: true,
|
||||||
@@ -465,6 +483,7 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const onBlockCompleted = (data: BlockCompletedData) => {
|
const onBlockCompleted = (data: BlockCompletedData) => {
|
||||||
|
if (isStaleExecution()) return
|
||||||
updateActiveBlocks(data.blockId, false)
|
updateActiveBlocks(data.blockId, false)
|
||||||
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'success')
|
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'success')
|
||||||
|
|
||||||
@@ -495,6 +514,7 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const onBlockError = (data: BlockErrorData) => {
|
const onBlockError = (data: BlockErrorData) => {
|
||||||
|
if (isStaleExecution()) return
|
||||||
updateActiveBlocks(data.blockId, false)
|
updateActiveBlocks(data.blockId, false)
|
||||||
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'error')
|
if (workflowId) setBlockRunStatus(workflowId, data.blockId, 'error')
|
||||||
|
|
||||||
@@ -902,10 +922,6 @@ export function useWorkflowExecution() {
|
|||||||
|
|
||||||
// Update block logs with actual stream completion times
|
// Update block logs with actual stream completion times
|
||||||
if (result.logs && streamCompletionTimes.size > 0) {
|
if (result.logs && streamCompletionTimes.size > 0) {
|
||||||
const streamCompletionEndTime = new Date(
|
|
||||||
Math.max(...Array.from(streamCompletionTimes.values()))
|
|
||||||
).toISOString()
|
|
||||||
|
|
||||||
result.logs.forEach((log: BlockLog) => {
|
result.logs.forEach((log: BlockLog) => {
|
||||||
if (streamCompletionTimes.has(log.blockId)) {
|
if (streamCompletionTimes.has(log.blockId)) {
|
||||||
const completionTime = streamCompletionTimes.get(log.blockId)!
|
const completionTime = streamCompletionTimes.get(log.blockId)!
|
||||||
@@ -987,7 +1003,6 @@ export function useWorkflowExecution() {
|
|||||||
return { success: true, stream }
|
return { success: true, stream }
|
||||||
}
|
}
|
||||||
|
|
||||||
// For manual (non-chat) execution
|
|
||||||
const manualExecutionId = uuidv4()
|
const manualExecutionId = uuidv4()
|
||||||
try {
|
try {
|
||||||
const result = await executeWorkflow(
|
const result = await executeWorkflow(
|
||||||
@@ -1002,29 +1017,10 @@ export function useWorkflowExecution() {
|
|||||||
if (result.metadata.pendingBlocks) {
|
if (result.metadata.pendingBlocks) {
|
||||||
setPendingBlocks(activeWorkflowId, result.metadata.pendingBlocks)
|
setPendingBlocks(activeWorkflowId, result.metadata.pendingBlocks)
|
||||||
}
|
}
|
||||||
} else if (result && 'success' in result) {
|
|
||||||
setExecutionResult(result)
|
|
||||||
// Reset execution state after successful non-debug execution
|
|
||||||
setIsExecuting(activeWorkflowId, false)
|
|
||||||
setIsDebugging(activeWorkflowId, false)
|
|
||||||
setActiveBlocks(activeWorkflowId, new Set())
|
|
||||||
|
|
||||||
if (isChatExecution) {
|
|
||||||
if (!result.metadata) {
|
|
||||||
result.metadata = { duration: 0, startTime: new Date().toISOString() }
|
|
||||||
}
|
|
||||||
;(result.metadata as any).source = 'chat'
|
|
||||||
}
|
|
||||||
|
|
||||||
// Invalidate subscription queries to update usage
|
|
||||||
setTimeout(() => {
|
|
||||||
queryClient.invalidateQueries({ queryKey: subscriptionKeys.all })
|
|
||||||
}, 1000)
|
|
||||||
}
|
}
|
||||||
return result
|
return result
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
const errorResult = handleExecutionError(error, { executionId: manualExecutionId })
|
const errorResult = handleExecutionError(error, { executionId: manualExecutionId })
|
||||||
// Note: Error logs are already persisted server-side via execution-core.ts
|
|
||||||
return errorResult
|
return errorResult
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@@ -1275,7 +1271,7 @@ export function useWorkflowExecution() {
|
|||||||
if (activeWorkflowId) {
|
if (activeWorkflowId) {
|
||||||
logger.info('Using server-side executor')
|
logger.info('Using server-side executor')
|
||||||
|
|
||||||
const executionId = uuidv4()
|
const executionIdRef = { current: '' }
|
||||||
|
|
||||||
let executionResult: ExecutionResult = {
|
let executionResult: ExecutionResult = {
|
||||||
success: false,
|
success: false,
|
||||||
@@ -1293,7 +1289,7 @@ export function useWorkflowExecution() {
|
|||||||
try {
|
try {
|
||||||
const blockHandlers = buildBlockEventHandlers({
|
const blockHandlers = buildBlockEventHandlers({
|
||||||
workflowId: activeWorkflowId,
|
workflowId: activeWorkflowId,
|
||||||
executionId,
|
executionIdRef,
|
||||||
workflowEdges,
|
workflowEdges,
|
||||||
activeBlocksSet,
|
activeBlocksSet,
|
||||||
accumulatedBlockLogs,
|
accumulatedBlockLogs,
|
||||||
@@ -1326,6 +1322,10 @@ export function useWorkflowExecution() {
|
|||||||
loops: clientWorkflowState.loops,
|
loops: clientWorkflowState.loops,
|
||||||
parallels: clientWorkflowState.parallels,
|
parallels: clientWorkflowState.parallels,
|
||||||
},
|
},
|
||||||
|
onExecutionId: (id) => {
|
||||||
|
executionIdRef.current = id
|
||||||
|
setCurrentExecutionId(activeWorkflowId, id)
|
||||||
|
},
|
||||||
callbacks: {
|
callbacks: {
|
||||||
onExecutionStarted: (data) => {
|
onExecutionStarted: (data) => {
|
||||||
logger.info('Server execution started:', data)
|
logger.info('Server execution started:', data)
|
||||||
@@ -1368,6 +1368,18 @@ export function useWorkflowExecution() {
|
|||||||
},
|
},
|
||||||
|
|
||||||
onExecutionCompleted: (data) => {
|
onExecutionCompleted: (data) => {
|
||||||
|
if (
|
||||||
|
activeWorkflowId &&
|
||||||
|
executionIdRef.current &&
|
||||||
|
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
||||||
|
executionIdRef.current
|
||||||
|
)
|
||||||
|
return
|
||||||
|
|
||||||
|
if (activeWorkflowId) {
|
||||||
|
setCurrentExecutionId(activeWorkflowId, null)
|
||||||
|
}
|
||||||
|
|
||||||
executionResult = {
|
executionResult = {
|
||||||
success: data.success,
|
success: data.success,
|
||||||
output: data.output,
|
output: data.output,
|
||||||
@@ -1425,9 +1437,33 @@ export function useWorkflowExecution() {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const workflowExecState = activeWorkflowId
|
||||||
|
? useExecutionStore.getState().getWorkflowExecution(activeWorkflowId)
|
||||||
|
: null
|
||||||
|
if (activeWorkflowId && !workflowExecState?.isDebugging) {
|
||||||
|
setExecutionResult(executionResult)
|
||||||
|
setIsExecuting(activeWorkflowId, false)
|
||||||
|
setActiveBlocks(activeWorkflowId, new Set())
|
||||||
|
setTimeout(() => {
|
||||||
|
queryClient.invalidateQueries({ queryKey: subscriptionKeys.all })
|
||||||
|
}, 1000)
|
||||||
|
}
|
||||||
},
|
},
|
||||||
|
|
||||||
onExecutionError: (data) => {
|
onExecutionError: (data) => {
|
||||||
|
if (
|
||||||
|
activeWorkflowId &&
|
||||||
|
executionIdRef.current &&
|
||||||
|
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
||||||
|
executionIdRef.current
|
||||||
|
)
|
||||||
|
return
|
||||||
|
|
||||||
|
if (activeWorkflowId) {
|
||||||
|
setCurrentExecutionId(activeWorkflowId, null)
|
||||||
|
}
|
||||||
|
|
||||||
executionResult = {
|
executionResult = {
|
||||||
success: false,
|
success: false,
|
||||||
output: {},
|
output: {},
|
||||||
@@ -1441,43 +1477,53 @@ export function useWorkflowExecution() {
|
|||||||
const isPreExecutionError = accumulatedBlockLogs.length === 0
|
const isPreExecutionError = accumulatedBlockLogs.length === 0
|
||||||
handleExecutionErrorConsole({
|
handleExecutionErrorConsole({
|
||||||
workflowId: activeWorkflowId,
|
workflowId: activeWorkflowId,
|
||||||
executionId,
|
executionId: executionIdRef.current,
|
||||||
error: data.error,
|
error: data.error,
|
||||||
durationMs: data.duration,
|
durationMs: data.duration,
|
||||||
blockLogs: accumulatedBlockLogs,
|
blockLogs: accumulatedBlockLogs,
|
||||||
isPreExecutionError,
|
isPreExecutionError,
|
||||||
})
|
})
|
||||||
|
|
||||||
|
if (activeWorkflowId) {
|
||||||
|
setIsExecuting(activeWorkflowId, false)
|
||||||
|
setIsDebugging(activeWorkflowId, false)
|
||||||
|
setActiveBlocks(activeWorkflowId, new Set())
|
||||||
|
}
|
||||||
},
|
},
|
||||||
|
|
||||||
onExecutionCancelled: (data) => {
|
onExecutionCancelled: (data) => {
|
||||||
|
if (
|
||||||
|
activeWorkflowId &&
|
||||||
|
executionIdRef.current &&
|
||||||
|
useExecutionStore.getState().getCurrentExecutionId(activeWorkflowId) !==
|
||||||
|
executionIdRef.current
|
||||||
|
)
|
||||||
|
return
|
||||||
|
|
||||||
|
if (activeWorkflowId) {
|
||||||
|
setCurrentExecutionId(activeWorkflowId, null)
|
||||||
|
}
|
||||||
|
|
||||||
handleExecutionCancelledConsole({
|
handleExecutionCancelledConsole({
|
||||||
workflowId: activeWorkflowId,
|
workflowId: activeWorkflowId,
|
||||||
executionId,
|
executionId: executionIdRef.current,
|
||||||
durationMs: data?.duration,
|
durationMs: data?.duration,
|
||||||
})
|
})
|
||||||
|
|
||||||
|
if (activeWorkflowId) {
|
||||||
|
setIsExecuting(activeWorkflowId, false)
|
||||||
|
setIsDebugging(activeWorkflowId, false)
|
||||||
|
setActiveBlocks(activeWorkflowId, new Set())
|
||||||
|
}
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
return executionResult
|
return executionResult
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
// Don't log abort errors - they're intentional user actions
|
|
||||||
if (error.name === 'AbortError' || error.message?.includes('aborted')) {
|
if (error.name === 'AbortError' || error.message?.includes('aborted')) {
|
||||||
logger.info('Execution aborted by user')
|
logger.info('Execution aborted by user')
|
||||||
|
return executionResult
|
||||||
// Reset execution state
|
|
||||||
if (activeWorkflowId) {
|
|
||||||
setIsExecuting(activeWorkflowId, false)
|
|
||||||
setActiveBlocks(activeWorkflowId, new Set())
|
|
||||||
}
|
|
||||||
|
|
||||||
// Return gracefully without error
|
|
||||||
return {
|
|
||||||
success: false,
|
|
||||||
output: {},
|
|
||||||
metadata: { duration: 0 },
|
|
||||||
logs: [],
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.error('Server-side execution failed:', error)
|
logger.error('Server-side execution failed:', error)
|
||||||
@@ -1485,7 +1531,6 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Fallback: should never reach here
|
|
||||||
throw new Error('Server-side execution is required')
|
throw new Error('Server-side execution is required')
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1717,25 +1762,28 @@ export function useWorkflowExecution() {
|
|||||||
* Handles cancelling the current workflow execution
|
* Handles cancelling the current workflow execution
|
||||||
*/
|
*/
|
||||||
const handleCancelExecution = useCallback(() => {
|
const handleCancelExecution = useCallback(() => {
|
||||||
|
if (!activeWorkflowId) return
|
||||||
logger.info('Workflow execution cancellation requested')
|
logger.info('Workflow execution cancellation requested')
|
||||||
|
|
||||||
// Cancel the execution stream for this workflow (server-side)
|
const storedExecutionId = getCurrentExecutionId(activeWorkflowId)
|
||||||
executionStream.cancel(activeWorkflowId ?? undefined)
|
|
||||||
|
|
||||||
// Mark current chat execution as superseded so its cleanup won't affect new executions
|
if (storedExecutionId) {
|
||||||
currentChatExecutionIdRef.current = null
|
setCurrentExecutionId(activeWorkflowId, null)
|
||||||
|
fetch(`/api/workflows/${activeWorkflowId}/executions/${storedExecutionId}/cancel`, {
|
||||||
// Mark all running entries as canceled in the terminal
|
method: 'POST',
|
||||||
if (activeWorkflowId) {
|
}).catch(() => {})
|
||||||
cancelRunningEntries(activeWorkflowId)
|
handleExecutionCancelledConsole({
|
||||||
|
workflowId: activeWorkflowId,
|
||||||
// Reset execution state - this triggers chat stream cleanup via useEffect in chat.tsx
|
executionId: storedExecutionId,
|
||||||
setIsExecuting(activeWorkflowId, false)
|
})
|
||||||
setIsDebugging(activeWorkflowId, false)
|
|
||||||
setActiveBlocks(activeWorkflowId, new Set())
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// If in debug mode, also reset debug state
|
executionStream.cancel(activeWorkflowId)
|
||||||
|
currentChatExecutionIdRef.current = null
|
||||||
|
setIsExecuting(activeWorkflowId, false)
|
||||||
|
setIsDebugging(activeWorkflowId, false)
|
||||||
|
setActiveBlocks(activeWorkflowId, new Set())
|
||||||
|
|
||||||
if (isDebugging) {
|
if (isDebugging) {
|
||||||
resetDebugState()
|
resetDebugState()
|
||||||
}
|
}
|
||||||
@@ -1747,7 +1795,9 @@ export function useWorkflowExecution() {
|
|||||||
setIsDebugging,
|
setIsDebugging,
|
||||||
setActiveBlocks,
|
setActiveBlocks,
|
||||||
activeWorkflowId,
|
activeWorkflowId,
|
||||||
cancelRunningEntries,
|
getCurrentExecutionId,
|
||||||
|
setCurrentExecutionId,
|
||||||
|
handleExecutionCancelledConsole,
|
||||||
])
|
])
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -1847,7 +1897,7 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
setIsExecuting(workflowId, true)
|
setIsExecuting(workflowId, true)
|
||||||
const executionId = uuidv4()
|
const executionIdRef = { current: '' }
|
||||||
const accumulatedBlockLogs: BlockLog[] = []
|
const accumulatedBlockLogs: BlockLog[] = []
|
||||||
const accumulatedBlockStates = new Map<string, BlockState>()
|
const accumulatedBlockStates = new Map<string, BlockState>()
|
||||||
const executedBlockIds = new Set<string>()
|
const executedBlockIds = new Set<string>()
|
||||||
@@ -1856,7 +1906,7 @@ export function useWorkflowExecution() {
|
|||||||
try {
|
try {
|
||||||
const blockHandlers = buildBlockEventHandlers({
|
const blockHandlers = buildBlockEventHandlers({
|
||||||
workflowId,
|
workflowId,
|
||||||
executionId,
|
executionIdRef,
|
||||||
workflowEdges,
|
workflowEdges,
|
||||||
activeBlocksSet,
|
activeBlocksSet,
|
||||||
accumulatedBlockLogs,
|
accumulatedBlockLogs,
|
||||||
@@ -1871,6 +1921,10 @@ export function useWorkflowExecution() {
|
|||||||
startBlockId: blockId,
|
startBlockId: blockId,
|
||||||
sourceSnapshot: effectiveSnapshot,
|
sourceSnapshot: effectiveSnapshot,
|
||||||
input: workflowInput,
|
input: workflowInput,
|
||||||
|
onExecutionId: (id) => {
|
||||||
|
executionIdRef.current = id
|
||||||
|
setCurrentExecutionId(workflowId, id)
|
||||||
|
},
|
||||||
callbacks: {
|
callbacks: {
|
||||||
onBlockStarted: blockHandlers.onBlockStarted,
|
onBlockStarted: blockHandlers.onBlockStarted,
|
||||||
onBlockCompleted: blockHandlers.onBlockCompleted,
|
onBlockCompleted: blockHandlers.onBlockCompleted,
|
||||||
@@ -1878,7 +1932,6 @@ export function useWorkflowExecution() {
|
|||||||
|
|
||||||
onExecutionCompleted: (data) => {
|
onExecutionCompleted: (data) => {
|
||||||
if (data.success) {
|
if (data.success) {
|
||||||
// Add the start block (trigger) to executed blocks
|
|
||||||
executedBlockIds.add(blockId)
|
executedBlockIds.add(blockId)
|
||||||
|
|
||||||
const mergedBlockStates: Record<string, BlockState> = {
|
const mergedBlockStates: Record<string, BlockState> = {
|
||||||
@@ -1902,6 +1955,10 @@ export function useWorkflowExecution() {
|
|||||||
}
|
}
|
||||||
setLastExecutionSnapshot(workflowId, updatedSnapshot)
|
setLastExecutionSnapshot(workflowId, updatedSnapshot)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
setCurrentExecutionId(workflowId, null)
|
||||||
|
setIsExecuting(workflowId, false)
|
||||||
|
setActiveBlocks(workflowId, new Set())
|
||||||
},
|
},
|
||||||
|
|
||||||
onExecutionError: (data) => {
|
onExecutionError: (data) => {
|
||||||
@@ -1921,19 +1978,27 @@ export function useWorkflowExecution() {
|
|||||||
|
|
||||||
handleExecutionErrorConsole({
|
handleExecutionErrorConsole({
|
||||||
workflowId,
|
workflowId,
|
||||||
executionId,
|
executionId: executionIdRef.current,
|
||||||
error: data.error,
|
error: data.error,
|
||||||
durationMs: data.duration,
|
durationMs: data.duration,
|
||||||
blockLogs: accumulatedBlockLogs,
|
blockLogs: accumulatedBlockLogs,
|
||||||
})
|
})
|
||||||
|
|
||||||
|
setCurrentExecutionId(workflowId, null)
|
||||||
|
setIsExecuting(workflowId, false)
|
||||||
|
setActiveBlocks(workflowId, new Set())
|
||||||
},
|
},
|
||||||
|
|
||||||
onExecutionCancelled: (data) => {
|
onExecutionCancelled: (data) => {
|
||||||
handleExecutionCancelledConsole({
|
handleExecutionCancelledConsole({
|
||||||
workflowId,
|
workflowId,
|
||||||
executionId,
|
executionId: executionIdRef.current,
|
||||||
durationMs: data?.duration,
|
durationMs: data?.duration,
|
||||||
})
|
})
|
||||||
|
|
||||||
|
setCurrentExecutionId(workflowId, null)
|
||||||
|
setIsExecuting(workflowId, false)
|
||||||
|
setActiveBlocks(workflowId, new Set())
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
@@ -1942,14 +2007,20 @@ export function useWorkflowExecution() {
|
|||||||
logger.error('Run-from-block failed:', error)
|
logger.error('Run-from-block failed:', error)
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
setIsExecuting(workflowId, false)
|
const currentId = getCurrentExecutionId(workflowId)
|
||||||
setActiveBlocks(workflowId, new Set())
|
if (currentId === null || currentId === executionIdRef.current) {
|
||||||
|
setCurrentExecutionId(workflowId, null)
|
||||||
|
setIsExecuting(workflowId, false)
|
||||||
|
setActiveBlocks(workflowId, new Set())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
[
|
[
|
||||||
getLastExecutionSnapshot,
|
getLastExecutionSnapshot,
|
||||||
setLastExecutionSnapshot,
|
setLastExecutionSnapshot,
|
||||||
clearLastExecutionSnapshot,
|
clearLastExecutionSnapshot,
|
||||||
|
getCurrentExecutionId,
|
||||||
|
setCurrentExecutionId,
|
||||||
setIsExecuting,
|
setIsExecuting,
|
||||||
setActiveBlocks,
|
setActiveBlocks,
|
||||||
setBlockRunStatus,
|
setBlockRunStatus,
|
||||||
@@ -1979,29 +2050,213 @@ export function useWorkflowExecution() {
|
|||||||
|
|
||||||
const executionId = uuidv4()
|
const executionId = uuidv4()
|
||||||
try {
|
try {
|
||||||
const result = await executeWorkflow(
|
await executeWorkflow(undefined, undefined, executionId, undefined, 'manual', blockId)
|
||||||
undefined,
|
|
||||||
undefined,
|
|
||||||
executionId,
|
|
||||||
undefined,
|
|
||||||
'manual',
|
|
||||||
blockId
|
|
||||||
)
|
|
||||||
if (result && 'success' in result) {
|
|
||||||
setExecutionResult(result)
|
|
||||||
}
|
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
const errorResult = handleExecutionError(error, { executionId })
|
const errorResult = handleExecutionError(error, { executionId })
|
||||||
return errorResult
|
return errorResult
|
||||||
} finally {
|
} finally {
|
||||||
|
setCurrentExecutionId(workflowId, null)
|
||||||
setIsExecuting(workflowId, false)
|
setIsExecuting(workflowId, false)
|
||||||
setIsDebugging(workflowId, false)
|
setIsDebugging(workflowId, false)
|
||||||
setActiveBlocks(workflowId, new Set())
|
setActiveBlocks(workflowId, new Set())
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
[activeWorkflowId, setExecutionResult, setIsExecuting, setIsDebugging, setActiveBlocks]
|
[
|
||||||
|
activeWorkflowId,
|
||||||
|
setCurrentExecutionId,
|
||||||
|
setExecutionResult,
|
||||||
|
setIsExecuting,
|
||||||
|
setIsDebugging,
|
||||||
|
setActiveBlocks,
|
||||||
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
|
useEffect(() => {
|
||||||
|
if (!activeWorkflowId || !hasHydrated) return
|
||||||
|
|
||||||
|
const entries = useTerminalConsoleStore.getState().entries
|
||||||
|
const runningEntries = entries.filter(
|
||||||
|
(e) => e.isRunning && e.workflowId === activeWorkflowId && e.executionId
|
||||||
|
)
|
||||||
|
if (runningEntries.length === 0) return
|
||||||
|
|
||||||
|
if (activeReconnections.has(activeWorkflowId)) return
|
||||||
|
activeReconnections.add(activeWorkflowId)
|
||||||
|
|
||||||
|
executionStream.cancel(activeWorkflowId)
|
||||||
|
|
||||||
|
const sorted = [...runningEntries].sort((a, b) => {
|
||||||
|
const aTime = a.startedAt ? new Date(a.startedAt).getTime() : 0
|
||||||
|
const bTime = b.startedAt ? new Date(b.startedAt).getTime() : 0
|
||||||
|
return bTime - aTime
|
||||||
|
})
|
||||||
|
const executionId = sorted[0].executionId!
|
||||||
|
|
||||||
|
const otherExecutionIds = new Set(
|
||||||
|
sorted.filter((e) => e.executionId !== executionId).map((e) => e.executionId!)
|
||||||
|
)
|
||||||
|
if (otherExecutionIds.size > 0) {
|
||||||
|
cancelRunningEntries(activeWorkflowId)
|
||||||
|
}
|
||||||
|
|
||||||
|
setCurrentExecutionId(activeWorkflowId, executionId)
|
||||||
|
setIsExecuting(activeWorkflowId, true)
|
||||||
|
|
||||||
|
const workflowEdges = useWorkflowStore.getState().edges
|
||||||
|
const activeBlocksSet = new Set<string>()
|
||||||
|
const accumulatedBlockLogs: BlockLog[] = []
|
||||||
|
const accumulatedBlockStates = new Map<string, BlockState>()
|
||||||
|
const executedBlockIds = new Set<string>()
|
||||||
|
|
||||||
|
const executionIdRef = { current: executionId }
|
||||||
|
|
||||||
|
const handlers = buildBlockEventHandlers({
|
||||||
|
workflowId: activeWorkflowId,
|
||||||
|
executionIdRef,
|
||||||
|
workflowEdges,
|
||||||
|
activeBlocksSet,
|
||||||
|
accumulatedBlockLogs,
|
||||||
|
accumulatedBlockStates,
|
||||||
|
executedBlockIds,
|
||||||
|
consoleMode: 'update',
|
||||||
|
includeStartConsoleEntry: true,
|
||||||
|
})
|
||||||
|
|
||||||
|
const originalEntries = entries
|
||||||
|
.filter((e) => e.executionId === executionId)
|
||||||
|
.map((e) => ({ ...e }))
|
||||||
|
|
||||||
|
let cleared = false
|
||||||
|
let reconnectionComplete = false
|
||||||
|
let cleanupRan = false
|
||||||
|
const clearOnce = () => {
|
||||||
|
if (!cleared) {
|
||||||
|
cleared = true
|
||||||
|
clearExecutionEntries(executionId)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const reconnectWorkflowId = activeWorkflowId
|
||||||
|
|
||||||
|
executionStream
|
||||||
|
.reconnect({
|
||||||
|
workflowId: reconnectWorkflowId,
|
||||||
|
executionId,
|
||||||
|
callbacks: {
|
||||||
|
onBlockStarted: (data) => {
|
||||||
|
clearOnce()
|
||||||
|
handlers.onBlockStarted(data)
|
||||||
|
},
|
||||||
|
onBlockCompleted: (data) => {
|
||||||
|
clearOnce()
|
||||||
|
handlers.onBlockCompleted(data)
|
||||||
|
},
|
||||||
|
onBlockError: (data) => {
|
||||||
|
clearOnce()
|
||||||
|
handlers.onBlockError(data)
|
||||||
|
},
|
||||||
|
onExecutionCompleted: () => {
|
||||||
|
const currentId = useExecutionStore
|
||||||
|
.getState()
|
||||||
|
.getCurrentExecutionId(reconnectWorkflowId)
|
||||||
|
if (currentId !== executionId) {
|
||||||
|
reconnectionComplete = true
|
||||||
|
activeReconnections.delete(reconnectWorkflowId)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
clearOnce()
|
||||||
|
reconnectionComplete = true
|
||||||
|
activeReconnections.delete(reconnectWorkflowId)
|
||||||
|
setCurrentExecutionId(reconnectWorkflowId, null)
|
||||||
|
setIsExecuting(reconnectWorkflowId, false)
|
||||||
|
setActiveBlocks(reconnectWorkflowId, new Set())
|
||||||
|
},
|
||||||
|
onExecutionError: (data) => {
|
||||||
|
const currentId = useExecutionStore
|
||||||
|
.getState()
|
||||||
|
.getCurrentExecutionId(reconnectWorkflowId)
|
||||||
|
if (currentId !== executionId) {
|
||||||
|
reconnectionComplete = true
|
||||||
|
activeReconnections.delete(reconnectWorkflowId)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
clearOnce()
|
||||||
|
reconnectionComplete = true
|
||||||
|
activeReconnections.delete(reconnectWorkflowId)
|
||||||
|
setCurrentExecutionId(reconnectWorkflowId, null)
|
||||||
|
setIsExecuting(reconnectWorkflowId, false)
|
||||||
|
setActiveBlocks(reconnectWorkflowId, new Set())
|
||||||
|
handleExecutionErrorConsole({
|
||||||
|
workflowId: reconnectWorkflowId,
|
||||||
|
executionId,
|
||||||
|
error: data.error,
|
||||||
|
blockLogs: accumulatedBlockLogs,
|
||||||
|
})
|
||||||
|
},
|
||||||
|
onExecutionCancelled: () => {
|
||||||
|
const currentId = useExecutionStore
|
||||||
|
.getState()
|
||||||
|
.getCurrentExecutionId(reconnectWorkflowId)
|
||||||
|
if (currentId !== executionId) {
|
||||||
|
reconnectionComplete = true
|
||||||
|
activeReconnections.delete(reconnectWorkflowId)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
clearOnce()
|
||||||
|
reconnectionComplete = true
|
||||||
|
activeReconnections.delete(reconnectWorkflowId)
|
||||||
|
setCurrentExecutionId(reconnectWorkflowId, null)
|
||||||
|
setIsExecuting(reconnectWorkflowId, false)
|
||||||
|
setActiveBlocks(reconnectWorkflowId, new Set())
|
||||||
|
handleExecutionCancelledConsole({
|
||||||
|
workflowId: reconnectWorkflowId,
|
||||||
|
executionId,
|
||||||
|
})
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
.catch((error) => {
|
||||||
|
logger.warn('Execution reconnection failed', { executionId, error })
|
||||||
|
})
|
||||||
|
.finally(() => {
|
||||||
|
if (reconnectionComplete || cleanupRan) return
|
||||||
|
const currentId = useExecutionStore.getState().getCurrentExecutionId(reconnectWorkflowId)
|
||||||
|
if (currentId !== executionId) return
|
||||||
|
reconnectionComplete = true
|
||||||
|
activeReconnections.delete(reconnectWorkflowId)
|
||||||
|
clearExecutionEntries(executionId)
|
||||||
|
for (const entry of originalEntries) {
|
||||||
|
addConsole({
|
||||||
|
workflowId: entry.workflowId,
|
||||||
|
blockId: entry.blockId,
|
||||||
|
blockName: entry.blockName,
|
||||||
|
blockType: entry.blockType,
|
||||||
|
executionId: entry.executionId,
|
||||||
|
executionOrder: entry.executionOrder,
|
||||||
|
isRunning: false,
|
||||||
|
warning: 'Execution result unavailable — check the logs page',
|
||||||
|
})
|
||||||
|
}
|
||||||
|
setCurrentExecutionId(reconnectWorkflowId, null)
|
||||||
|
setIsExecuting(reconnectWorkflowId, false)
|
||||||
|
setActiveBlocks(reconnectWorkflowId, new Set())
|
||||||
|
})
|
||||||
|
|
||||||
|
return () => {
|
||||||
|
cleanupRan = true
|
||||||
|
executionStream.cancel(reconnectWorkflowId)
|
||||||
|
activeReconnections.delete(reconnectWorkflowId)
|
||||||
|
|
||||||
|
if (cleared && !reconnectionComplete) {
|
||||||
|
clearExecutionEntries(executionId)
|
||||||
|
for (const entry of originalEntries) {
|
||||||
|
addConsole(entry)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// eslint-disable-next-line react-hooks/exhaustive-deps
|
||||||
|
}, [activeWorkflowId, hasHydrated])
|
||||||
|
|
||||||
return {
|
return {
|
||||||
isExecuting,
|
isExecuting,
|
||||||
isDebugging,
|
isDebugging,
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
export { CancelSubscription } from './cancel-subscription'
|
export { CancelSubscription } from './cancel-subscription'
|
||||||
export { CreditBalance } from './credit-balance'
|
export { CreditBalance } from './credit-balance'
|
||||||
export { PlanCard, type PlanCardProps, type PlanFeature } from './plan-card'
|
export { PlanCard, type PlanCardProps, type PlanFeature } from './plan-card'
|
||||||
|
export { ReferralCode } from './referral-code'
|
||||||
|
|||||||
@@ -0,0 +1 @@
|
|||||||
|
export { ReferralCode } from './referral-code'
|
||||||
@@ -0,0 +1,103 @@
|
|||||||
|
'use client'
|
||||||
|
|
||||||
|
import { useState } from 'react'
|
||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { Button, Input, Label } from '@/components/emcn'
|
||||||
|
|
||||||
|
const logger = createLogger('ReferralCode')
|
||||||
|
|
||||||
|
interface ReferralCodeProps {
|
||||||
|
onRedeemComplete?: () => void
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Inline referral/promo code entry field with redeem button.
|
||||||
|
* One-time use per account — shows success or "already redeemed" state.
|
||||||
|
*/
|
||||||
|
export function ReferralCode({ onRedeemComplete }: ReferralCodeProps) {
|
||||||
|
const [code, setCode] = useState('')
|
||||||
|
const [isRedeeming, setIsRedeeming] = useState(false)
|
||||||
|
const [error, setError] = useState<string | null>(null)
|
||||||
|
const [success, setSuccess] = useState<{ bonusAmount: number } | null>(null)
|
||||||
|
|
||||||
|
const handleRedeem = async () => {
|
||||||
|
const trimmed = code.trim()
|
||||||
|
if (!trimmed || isRedeeming) return
|
||||||
|
|
||||||
|
setIsRedeeming(true)
|
||||||
|
setError(null)
|
||||||
|
|
||||||
|
try {
|
||||||
|
const response = await fetch('/api/referral-code/redeem', {
|
||||||
|
method: 'POST',
|
||||||
|
headers: { 'Content-Type': 'application/json' },
|
||||||
|
body: JSON.stringify({ code: trimmed }),
|
||||||
|
})
|
||||||
|
|
||||||
|
const data = await response.json()
|
||||||
|
|
||||||
|
if (!response.ok) {
|
||||||
|
throw new Error(data.error || 'Failed to redeem code')
|
||||||
|
}
|
||||||
|
|
||||||
|
if (data.redeemed) {
|
||||||
|
setSuccess({ bonusAmount: data.bonusAmount })
|
||||||
|
setCode('')
|
||||||
|
onRedeemComplete?.()
|
||||||
|
} else {
|
||||||
|
setError(data.error || 'Code could not be redeemed')
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
logger.error('Referral code redemption failed', { error: err })
|
||||||
|
setError(err instanceof Error ? err.message : 'Failed to redeem code')
|
||||||
|
} finally {
|
||||||
|
setIsRedeeming(false)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (success) {
|
||||||
|
return (
|
||||||
|
<div className='flex items-center justify-between'>
|
||||||
|
<Label>Referral Code</Label>
|
||||||
|
<span className='text-[12px] text-[var(--text-secondary)]'>
|
||||||
|
+${success.bonusAmount} credits applied
|
||||||
|
</span>
|
||||||
|
</div>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
return (
|
||||||
|
<div className='flex flex-col'>
|
||||||
|
<div className='flex items-center justify-between gap-[12px]'>
|
||||||
|
<Label className='shrink-0'>Referral Code</Label>
|
||||||
|
<div className='flex items-center gap-[8px]'>
|
||||||
|
<Input
|
||||||
|
type='text'
|
||||||
|
value={code}
|
||||||
|
onChange={(e) => {
|
||||||
|
setCode(e.target.value)
|
||||||
|
setError(null)
|
||||||
|
}}
|
||||||
|
onKeyDown={(e) => {
|
||||||
|
if (e.key === 'Enter') handleRedeem()
|
||||||
|
}}
|
||||||
|
placeholder='Enter code'
|
||||||
|
className='h-[32px] w-[140px] text-[12px]'
|
||||||
|
disabled={isRedeeming}
|
||||||
|
/>
|
||||||
|
<Button
|
||||||
|
variant='active'
|
||||||
|
className='h-[32px] shrink-0 rounded-[6px] text-[12px]'
|
||||||
|
onClick={handleRedeem}
|
||||||
|
disabled={isRedeeming || !code.trim()}
|
||||||
|
>
|
||||||
|
{isRedeeming ? 'Redeeming...' : 'Redeem'}
|
||||||
|
</Button>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
<div className='mt-[4px] min-h-[18px] text-right'>
|
||||||
|
{error && <span className='text-[11px] text-[var(--text-error)]'>{error}</span>}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
)
|
||||||
|
}
|
||||||
@@ -17,6 +17,7 @@ import {
|
|||||||
CancelSubscription,
|
CancelSubscription,
|
||||||
CreditBalance,
|
CreditBalance,
|
||||||
PlanCard,
|
PlanCard,
|
||||||
|
ReferralCode,
|
||||||
} from '@/app/workspace/[workspaceId]/w/components/sidebar/components/settings-modal/components/subscription/components'
|
} from '@/app/workspace/[workspaceId]/w/components/sidebar/components/settings-modal/components/subscription/components'
|
||||||
import {
|
import {
|
||||||
ENTERPRISE_PLAN_FEATURES,
|
ENTERPRISE_PLAN_FEATURES,
|
||||||
@@ -549,6 +550,10 @@ export function Subscription() {
|
|||||||
/>
|
/>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
|
{!subscription.isEnterprise && (
|
||||||
|
<ReferralCode onRedeemComplete={() => refetchSubscription()} />
|
||||||
|
)}
|
||||||
|
|
||||||
{/* Next Billing Date - hidden from team members */}
|
{/* Next Billing Date - hidden from team members */}
|
||||||
{subscription.isPaid &&
|
{subscription.isPaid &&
|
||||||
subscriptionData?.data?.periodEnd &&
|
subscriptionData?.data?.periodEnd &&
|
||||||
|
|||||||
@@ -4,12 +4,14 @@ import { useEffect } from 'react'
|
|||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { useRouter } from 'next/navigation'
|
import { useRouter } from 'next/navigation'
|
||||||
import { useSession } from '@/lib/auth/auth-client'
|
import { useSession } from '@/lib/auth/auth-client'
|
||||||
|
import { useReferralAttribution } from '@/hooks/use-referral-attribution'
|
||||||
|
|
||||||
const logger = createLogger('WorkspacePage')
|
const logger = createLogger('WorkspacePage')
|
||||||
|
|
||||||
export default function WorkspacePage() {
|
export default function WorkspacePage() {
|
||||||
const router = useRouter()
|
const router = useRouter()
|
||||||
const { data: session, isPending } = useSession()
|
const { data: session, isPending } = useSession()
|
||||||
|
useReferralAttribution()
|
||||||
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
const redirectToFirstWorkspace = async () => {
|
const redirectToFirstWorkspace = async () => {
|
||||||
|
|||||||
@@ -589,6 +589,7 @@ export async function executeScheduleJob(payload: ScheduleExecutionPayload) {
|
|||||||
|
|
||||||
export const scheduleExecution = task({
|
export const scheduleExecution = task({
|
||||||
id: 'schedule-execution',
|
id: 'schedule-execution',
|
||||||
|
machine: 'medium-1x',
|
||||||
retry: {
|
retry: {
|
||||||
maxAttempts: 1,
|
maxAttempts: 1,
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -669,6 +669,7 @@ async function executeWebhookJobInternal(
|
|||||||
|
|
||||||
export const webhookExecution = task({
|
export const webhookExecution = task({
|
||||||
id: 'webhook-execution',
|
id: 'webhook-execution',
|
||||||
|
machine: 'medium-1x',
|
||||||
retry: {
|
retry: {
|
||||||
maxAttempts: 1,
|
maxAttempts: 1,
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -197,5 +197,6 @@ export async function executeWorkflowJob(payload: WorkflowExecutionPayload) {
|
|||||||
|
|
||||||
export const workflowExecutionTask = task({
|
export const workflowExecutionTask = task({
|
||||||
id: 'workflow-execution',
|
id: 'workflow-execution',
|
||||||
|
machine: 'medium-1x',
|
||||||
run: executeWorkflowJob,
|
run: executeWorkflowJob,
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -10,9 +10,11 @@ import {
|
|||||||
getReasoningEffortValuesForModel,
|
getReasoningEffortValuesForModel,
|
||||||
getThinkingLevelsForModel,
|
getThinkingLevelsForModel,
|
||||||
getVerbosityValuesForModel,
|
getVerbosityValuesForModel,
|
||||||
|
MODELS_WITH_DEEP_RESEARCH,
|
||||||
MODELS_WITH_REASONING_EFFORT,
|
MODELS_WITH_REASONING_EFFORT,
|
||||||
MODELS_WITH_THINKING,
|
MODELS_WITH_THINKING,
|
||||||
MODELS_WITH_VERBOSITY,
|
MODELS_WITH_VERBOSITY,
|
||||||
|
MODELS_WITHOUT_MEMORY,
|
||||||
providers,
|
providers,
|
||||||
supportsTemperature,
|
supportsTemperature,
|
||||||
} from '@/providers/utils'
|
} from '@/providers/utils'
|
||||||
@@ -412,12 +414,22 @@ Return ONLY the JSON array.`,
|
|||||||
title: 'Tools',
|
title: 'Tools',
|
||||||
type: 'tool-input',
|
type: 'tool-input',
|
||||||
defaultValue: [],
|
defaultValue: [],
|
||||||
|
condition: {
|
||||||
|
field: 'model',
|
||||||
|
value: MODELS_WITH_DEEP_RESEARCH,
|
||||||
|
not: true,
|
||||||
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
id: 'skills',
|
id: 'skills',
|
||||||
title: 'Skills',
|
title: 'Skills',
|
||||||
type: 'skill-input',
|
type: 'skill-input',
|
||||||
defaultValue: [],
|
defaultValue: [],
|
||||||
|
condition: {
|
||||||
|
field: 'model',
|
||||||
|
value: MODELS_WITH_DEEP_RESEARCH,
|
||||||
|
not: true,
|
||||||
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
id: 'memoryType',
|
id: 'memoryType',
|
||||||
@@ -431,6 +443,11 @@ Return ONLY the JSON array.`,
|
|||||||
{ label: 'Sliding window (tokens)', id: 'sliding_window_tokens' },
|
{ label: 'Sliding window (tokens)', id: 'sliding_window_tokens' },
|
||||||
],
|
],
|
||||||
defaultValue: 'none',
|
defaultValue: 'none',
|
||||||
|
condition: {
|
||||||
|
field: 'model',
|
||||||
|
value: MODELS_WITHOUT_MEMORY,
|
||||||
|
not: true,
|
||||||
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
id: 'conversationId',
|
id: 'conversationId',
|
||||||
@@ -444,6 +461,7 @@ Return ONLY the JSON array.`,
|
|||||||
condition: {
|
condition: {
|
||||||
field: 'memoryType',
|
field: 'memoryType',
|
||||||
value: ['conversation', 'sliding_window', 'sliding_window_tokens'],
|
value: ['conversation', 'sliding_window', 'sliding_window_tokens'],
|
||||||
|
and: { field: 'model', value: MODELS_WITHOUT_MEMORY, not: true },
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -454,6 +472,7 @@ Return ONLY the JSON array.`,
|
|||||||
condition: {
|
condition: {
|
||||||
field: 'memoryType',
|
field: 'memoryType',
|
||||||
value: ['sliding_window'],
|
value: ['sliding_window'],
|
||||||
|
and: { field: 'model', value: MODELS_WITHOUT_MEMORY, not: true },
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -464,6 +483,7 @@ Return ONLY the JSON array.`,
|
|||||||
condition: {
|
condition: {
|
||||||
field: 'memoryType',
|
field: 'memoryType',
|
||||||
value: ['sliding_window_tokens'],
|
value: ['sliding_window_tokens'],
|
||||||
|
and: { field: 'model', value: MODELS_WITHOUT_MEMORY, not: true },
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -477,9 +497,13 @@ Return ONLY the JSON array.`,
|
|||||||
condition: () => ({
|
condition: () => ({
|
||||||
field: 'model',
|
field: 'model',
|
||||||
value: (() => {
|
value: (() => {
|
||||||
|
const deepResearch = new Set(MODELS_WITH_DEEP_RESEARCH.map((m) => m.toLowerCase()))
|
||||||
const allModels = Object.keys(getBaseModelProviders())
|
const allModels = Object.keys(getBaseModelProviders())
|
||||||
return allModels.filter(
|
return allModels.filter(
|
||||||
(model) => supportsTemperature(model) && getMaxTemperature(model) === 1
|
(model) =>
|
||||||
|
supportsTemperature(model) &&
|
||||||
|
getMaxTemperature(model) === 1 &&
|
||||||
|
!deepResearch.has(model.toLowerCase())
|
||||||
)
|
)
|
||||||
})(),
|
})(),
|
||||||
}),
|
}),
|
||||||
@@ -495,9 +519,13 @@ Return ONLY the JSON array.`,
|
|||||||
condition: () => ({
|
condition: () => ({
|
||||||
field: 'model',
|
field: 'model',
|
||||||
value: (() => {
|
value: (() => {
|
||||||
|
const deepResearch = new Set(MODELS_WITH_DEEP_RESEARCH.map((m) => m.toLowerCase()))
|
||||||
const allModels = Object.keys(getBaseModelProviders())
|
const allModels = Object.keys(getBaseModelProviders())
|
||||||
return allModels.filter(
|
return allModels.filter(
|
||||||
(model) => supportsTemperature(model) && getMaxTemperature(model) === 2
|
(model) =>
|
||||||
|
supportsTemperature(model) &&
|
||||||
|
getMaxTemperature(model) === 2 &&
|
||||||
|
!deepResearch.has(model.toLowerCase())
|
||||||
)
|
)
|
||||||
})(),
|
})(),
|
||||||
}),
|
}),
|
||||||
@@ -508,6 +536,11 @@ Return ONLY the JSON array.`,
|
|||||||
type: 'short-input',
|
type: 'short-input',
|
||||||
placeholder: 'Enter max tokens (e.g., 4096)...',
|
placeholder: 'Enter max tokens (e.g., 4096)...',
|
||||||
mode: 'advanced',
|
mode: 'advanced',
|
||||||
|
condition: {
|
||||||
|
field: 'model',
|
||||||
|
value: MODELS_WITH_DEEP_RESEARCH,
|
||||||
|
not: true,
|
||||||
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
id: 'responseFormat',
|
id: 'responseFormat',
|
||||||
@@ -515,6 +548,11 @@ Return ONLY the JSON array.`,
|
|||||||
type: 'code',
|
type: 'code',
|
||||||
placeholder: 'Enter JSON schema...',
|
placeholder: 'Enter JSON schema...',
|
||||||
language: 'json',
|
language: 'json',
|
||||||
|
condition: {
|
||||||
|
field: 'model',
|
||||||
|
value: MODELS_WITH_DEEP_RESEARCH,
|
||||||
|
not: true,
|
||||||
|
},
|
||||||
wandConfig: {
|
wandConfig: {
|
||||||
enabled: true,
|
enabled: true,
|
||||||
maintainHistory: true,
|
maintainHistory: true,
|
||||||
@@ -607,6 +645,16 @@ Example 3 (Array Input):
|
|||||||
generationType: 'json-schema',
|
generationType: 'json-schema',
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
id: 'previousInteractionId',
|
||||||
|
title: 'Previous Interaction ID',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'e.g., {{agent_1.interactionId}}',
|
||||||
|
condition: {
|
||||||
|
field: 'model',
|
||||||
|
value: MODELS_WITH_DEEP_RESEARCH,
|
||||||
|
},
|
||||||
|
},
|
||||||
],
|
],
|
||||||
tools: {
|
tools: {
|
||||||
access: [
|
access: [
|
||||||
@@ -770,5 +818,13 @@ Example 3 (Array Input):
|
|||||||
description: 'Provider timing information',
|
description: 'Provider timing information',
|
||||||
},
|
},
|
||||||
cost: { type: 'json', description: 'Cost of the API call' },
|
cost: { type: 'json', description: 'Cost of the API call' },
|
||||||
|
interactionId: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'Interaction ID for multi-turn deep research follow-ups',
|
||||||
|
condition: {
|
||||||
|
field: 'model',
|
||||||
|
value: MODELS_WITH_DEEP_RESEARCH,
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -394,6 +394,7 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
// Page Property Operations
|
// Page Property Operations
|
||||||
{ label: 'List Page Properties', id: 'list_page_properties' },
|
{ label: 'List Page Properties', id: 'list_page_properties' },
|
||||||
{ label: 'Create Page Property', id: 'create_page_property' },
|
{ label: 'Create Page Property', id: 'create_page_property' },
|
||||||
|
{ label: 'Delete Page Property', id: 'delete_page_property' },
|
||||||
// Search Operations
|
// Search Operations
|
||||||
{ label: 'Search Content', id: 'search' },
|
{ label: 'Search Content', id: 'search' },
|
||||||
{ label: 'Search in Space', id: 'search_in_space' },
|
{ label: 'Search in Space', id: 'search_in_space' },
|
||||||
@@ -414,6 +415,9 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
// Label Operations
|
// Label Operations
|
||||||
{ label: 'List Labels', id: 'list_labels' },
|
{ label: 'List Labels', id: 'list_labels' },
|
||||||
{ label: 'Add Label', id: 'add_label' },
|
{ label: 'Add Label', id: 'add_label' },
|
||||||
|
{ label: 'Delete Label', id: 'delete_label' },
|
||||||
|
{ label: 'Get Pages by Label', id: 'get_pages_by_label' },
|
||||||
|
{ label: 'List Space Labels', id: 'list_space_labels' },
|
||||||
// Space Operations
|
// Space Operations
|
||||||
{ label: 'Get Space', id: 'get_space' },
|
{ label: 'Get Space', id: 'get_space' },
|
||||||
{ label: 'List Spaces', id: 'list_spaces' },
|
{ label: 'List Spaces', id: 'list_spaces' },
|
||||||
@@ -485,6 +489,8 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
'search_in_space',
|
'search_in_space',
|
||||||
'get_space',
|
'get_space',
|
||||||
'list_spaces',
|
'list_spaces',
|
||||||
|
'get_pages_by_label',
|
||||||
|
'list_space_labels',
|
||||||
],
|
],
|
||||||
not: true,
|
not: true,
|
||||||
},
|
},
|
||||||
@@ -500,6 +506,8 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
'list_labels',
|
'list_labels',
|
||||||
'upload_attachment',
|
'upload_attachment',
|
||||||
'add_label',
|
'add_label',
|
||||||
|
'delete_label',
|
||||||
|
'delete_page_property',
|
||||||
'get_page_children',
|
'get_page_children',
|
||||||
'get_page_ancestors',
|
'get_page_ancestors',
|
||||||
'list_page_versions',
|
'list_page_versions',
|
||||||
@@ -527,6 +535,8 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
'search_in_space',
|
'search_in_space',
|
||||||
'get_space',
|
'get_space',
|
||||||
'list_spaces',
|
'list_spaces',
|
||||||
|
'get_pages_by_label',
|
||||||
|
'list_space_labels',
|
||||||
],
|
],
|
||||||
not: true,
|
not: true,
|
||||||
},
|
},
|
||||||
@@ -542,6 +552,8 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
'list_labels',
|
'list_labels',
|
||||||
'upload_attachment',
|
'upload_attachment',
|
||||||
'add_label',
|
'add_label',
|
||||||
|
'delete_label',
|
||||||
|
'delete_page_property',
|
||||||
'get_page_children',
|
'get_page_children',
|
||||||
'get_page_ancestors',
|
'get_page_ancestors',
|
||||||
'list_page_versions',
|
'list_page_versions',
|
||||||
@@ -566,6 +578,7 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
'search_in_space',
|
'search_in_space',
|
||||||
'create_blogpost',
|
'create_blogpost',
|
||||||
'list_blogposts_in_space',
|
'list_blogposts_in_space',
|
||||||
|
'list_space_labels',
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@@ -601,6 +614,14 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
required: true,
|
required: true,
|
||||||
condition: { field: 'operation', value: 'create_page_property' },
|
condition: { field: 'operation', value: 'create_page_property' },
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
id: 'propertyId',
|
||||||
|
title: 'Property ID',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Enter property ID to delete',
|
||||||
|
required: true,
|
||||||
|
condition: { field: 'operation', value: 'delete_page_property' },
|
||||||
|
},
|
||||||
{
|
{
|
||||||
id: 'title',
|
id: 'title',
|
||||||
title: 'Title',
|
title: 'Title',
|
||||||
@@ -694,7 +715,7 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
type: 'short-input',
|
type: 'short-input',
|
||||||
placeholder: 'Enter label name',
|
placeholder: 'Enter label name',
|
||||||
required: true,
|
required: true,
|
||||||
condition: { field: 'operation', value: 'add_label' },
|
condition: { field: 'operation', value: ['add_label', 'delete_label'] },
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
id: 'labelPrefix',
|
id: 'labelPrefix',
|
||||||
@@ -709,6 +730,14 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
value: () => 'global',
|
value: () => 'global',
|
||||||
condition: { field: 'operation', value: 'add_label' },
|
condition: { field: 'operation', value: 'add_label' },
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
id: 'labelId',
|
||||||
|
title: 'Label ID',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Enter label ID',
|
||||||
|
required: true,
|
||||||
|
condition: { field: 'operation', value: 'get_pages_by_label' },
|
||||||
|
},
|
||||||
{
|
{
|
||||||
id: 'blogPostStatus',
|
id: 'blogPostStatus',
|
||||||
title: 'Status',
|
title: 'Status',
|
||||||
@@ -759,6 +788,8 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
'list_page_versions',
|
'list_page_versions',
|
||||||
'list_page_properties',
|
'list_page_properties',
|
||||||
'list_labels',
|
'list_labels',
|
||||||
|
'get_pages_by_label',
|
||||||
|
'list_space_labels',
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@@ -780,6 +811,8 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
'list_page_versions',
|
'list_page_versions',
|
||||||
'list_page_properties',
|
'list_page_properties',
|
||||||
'list_labels',
|
'list_labels',
|
||||||
|
'get_pages_by_label',
|
||||||
|
'list_space_labels',
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@@ -800,6 +833,7 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
// Property Tools
|
// Property Tools
|
||||||
'confluence_list_page_properties',
|
'confluence_list_page_properties',
|
||||||
'confluence_create_page_property',
|
'confluence_create_page_property',
|
||||||
|
'confluence_delete_page_property',
|
||||||
// Search Tools
|
// Search Tools
|
||||||
'confluence_search',
|
'confluence_search',
|
||||||
'confluence_search_in_space',
|
'confluence_search_in_space',
|
||||||
@@ -820,6 +854,9 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
// Label Tools
|
// Label Tools
|
||||||
'confluence_list_labels',
|
'confluence_list_labels',
|
||||||
'confluence_add_label',
|
'confluence_add_label',
|
||||||
|
'confluence_delete_label',
|
||||||
|
'confluence_get_pages_by_label',
|
||||||
|
'confluence_list_space_labels',
|
||||||
// Space Tools
|
// Space Tools
|
||||||
'confluence_get_space',
|
'confluence_get_space',
|
||||||
'confluence_list_spaces',
|
'confluence_list_spaces',
|
||||||
@@ -852,6 +889,8 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
return 'confluence_list_page_properties'
|
return 'confluence_list_page_properties'
|
||||||
case 'create_page_property':
|
case 'create_page_property':
|
||||||
return 'confluence_create_page_property'
|
return 'confluence_create_page_property'
|
||||||
|
case 'delete_page_property':
|
||||||
|
return 'confluence_delete_page_property'
|
||||||
// Search Operations
|
// Search Operations
|
||||||
case 'search':
|
case 'search':
|
||||||
return 'confluence_search'
|
return 'confluence_search'
|
||||||
@@ -887,6 +926,12 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
return 'confluence_list_labels'
|
return 'confluence_list_labels'
|
||||||
case 'add_label':
|
case 'add_label':
|
||||||
return 'confluence_add_label'
|
return 'confluence_add_label'
|
||||||
|
case 'delete_label':
|
||||||
|
return 'confluence_delete_label'
|
||||||
|
case 'get_pages_by_label':
|
||||||
|
return 'confluence_get_pages_by_label'
|
||||||
|
case 'list_space_labels':
|
||||||
|
return 'confluence_list_space_labels'
|
||||||
// Space Operations
|
// Space Operations
|
||||||
case 'get_space':
|
case 'get_space':
|
||||||
return 'confluence_get_space'
|
return 'confluence_get_space'
|
||||||
@@ -908,7 +953,9 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
versionNumber,
|
versionNumber,
|
||||||
propertyKey,
|
propertyKey,
|
||||||
propertyValue,
|
propertyValue,
|
||||||
|
propertyId,
|
||||||
labelPrefix,
|
labelPrefix,
|
||||||
|
labelId,
|
||||||
blogPostStatus,
|
blogPostStatus,
|
||||||
purge,
|
purge,
|
||||||
bodyFormat,
|
bodyFormat,
|
||||||
@@ -959,7 +1006,9 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Operations that support cursor pagination
|
// Operations that support generic cursor pagination.
|
||||||
|
// get_pages_by_label and list_space_labels have dedicated handlers
|
||||||
|
// below that pass cursor along with their required params (labelId, spaceId).
|
||||||
const supportsCursor = [
|
const supportsCursor = [
|
||||||
'list_attachments',
|
'list_attachments',
|
||||||
'list_spaces',
|
'list_spaces',
|
||||||
@@ -996,6 +1045,35 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (operation === 'delete_page_property') {
|
||||||
|
return {
|
||||||
|
credential,
|
||||||
|
pageId: effectivePageId,
|
||||||
|
operation,
|
||||||
|
propertyId,
|
||||||
|
...rest,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (operation === 'get_pages_by_label') {
|
||||||
|
return {
|
||||||
|
credential,
|
||||||
|
operation,
|
||||||
|
labelId,
|
||||||
|
cursor: cursor || undefined,
|
||||||
|
...rest,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (operation === 'list_space_labels') {
|
||||||
|
return {
|
||||||
|
credential,
|
||||||
|
operation,
|
||||||
|
cursor: cursor || undefined,
|
||||||
|
...rest,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if (operation === 'upload_attachment') {
|
if (operation === 'upload_attachment') {
|
||||||
const normalizedFile = normalizeFileInput(attachmentFile, { single: true })
|
const normalizedFile = normalizeFileInput(attachmentFile, { single: true })
|
||||||
if (!normalizedFile) {
|
if (!normalizedFile) {
|
||||||
@@ -1044,7 +1122,9 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
attachmentFileName: { type: 'string', description: 'Custom file name for attachment' },
|
attachmentFileName: { type: 'string', description: 'Custom file name for attachment' },
|
||||||
attachmentComment: { type: 'string', description: 'Comment for the attachment' },
|
attachmentComment: { type: 'string', description: 'Comment for the attachment' },
|
||||||
labelName: { type: 'string', description: 'Label name' },
|
labelName: { type: 'string', description: 'Label name' },
|
||||||
|
labelId: { type: 'string', description: 'Label identifier' },
|
||||||
labelPrefix: { type: 'string', description: 'Label prefix (global, my, team, system)' },
|
labelPrefix: { type: 'string', description: 'Label prefix (global, my, team, system)' },
|
||||||
|
propertyId: { type: 'string', description: 'Property identifier' },
|
||||||
blogPostStatus: { type: 'string', description: 'Blog post status (current or draft)' },
|
blogPostStatus: { type: 'string', description: 'Blog post status (current or draft)' },
|
||||||
purge: { type: 'boolean', description: 'Permanently delete instead of moving to trash' },
|
purge: { type: 'boolean', description: 'Permanently delete instead of moving to trash' },
|
||||||
bodyFormat: { type: 'string', description: 'Body format for comments' },
|
bodyFormat: { type: 'string', description: 'Body format for comments' },
|
||||||
@@ -1080,6 +1160,7 @@ export const ConfluenceV2Block: BlockConfig<ConfluenceResponse> = {
|
|||||||
// Label Results
|
// Label Results
|
||||||
labels: { type: 'array', description: 'List of labels' },
|
labels: { type: 'array', description: 'List of labels' },
|
||||||
labelName: { type: 'string', description: 'Label name' },
|
labelName: { type: 'string', description: 'Label name' },
|
||||||
|
labelId: { type: 'string', description: 'Label identifier' },
|
||||||
// Space Results
|
// Space Results
|
||||||
spaces: { type: 'array', description: 'List of spaces' },
|
spaces: { type: 'array', description: 'List of spaces' },
|
||||||
spaceId: { type: 'string', description: 'Space identifier' },
|
spaceId: { type: 'string', description: 'Space identifier' },
|
||||||
|
|||||||
201
apps/sim/blocks/blocks/google_books.ts
Normal file
201
apps/sim/blocks/blocks/google_books.ts
Normal file
@@ -0,0 +1,201 @@
|
|||||||
|
import { GoogleBooksIcon } from '@/components/icons'
|
||||||
|
import type { BlockConfig } from '@/blocks/types'
|
||||||
|
import { AuthMode } from '@/blocks/types'
|
||||||
|
|
||||||
|
export const GoogleBooksBlock: BlockConfig = {
|
||||||
|
type: 'google_books',
|
||||||
|
name: 'Google Books',
|
||||||
|
description: 'Search and retrieve book information',
|
||||||
|
authMode: AuthMode.ApiKey,
|
||||||
|
longDescription:
|
||||||
|
'Search for books using the Google Books API. Find volumes by title, author, ISBN, or keywords, and retrieve detailed information about specific books including descriptions, ratings, and publication details.',
|
||||||
|
docsLink: 'https://docs.sim.ai/tools/google_books',
|
||||||
|
category: 'tools',
|
||||||
|
bgColor: '#E0E0E0',
|
||||||
|
icon: GoogleBooksIcon,
|
||||||
|
|
||||||
|
subBlocks: [
|
||||||
|
{
|
||||||
|
id: 'operation',
|
||||||
|
title: 'Operation',
|
||||||
|
type: 'dropdown',
|
||||||
|
options: [
|
||||||
|
{ label: 'Search Volumes', id: 'volume_search' },
|
||||||
|
{ label: 'Get Volume Details', id: 'volume_details' },
|
||||||
|
],
|
||||||
|
value: () => 'volume_search',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'apiKey',
|
||||||
|
title: 'API Key',
|
||||||
|
type: 'short-input',
|
||||||
|
password: true,
|
||||||
|
placeholder: 'Enter your Google Books API key',
|
||||||
|
required: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'query',
|
||||||
|
title: 'Search Query',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'e.g., intitle:harry potter inauthor:rowling',
|
||||||
|
condition: { field: 'operation', value: 'volume_search' },
|
||||||
|
required: { field: 'operation', value: 'volume_search' },
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'filter',
|
||||||
|
title: 'Filter',
|
||||||
|
type: 'dropdown',
|
||||||
|
options: [
|
||||||
|
{ label: 'None', id: '' },
|
||||||
|
{ label: 'Partial Preview', id: 'partial' },
|
||||||
|
{ label: 'Full Preview', id: 'full' },
|
||||||
|
{ label: 'Free eBooks', id: 'free-ebooks' },
|
||||||
|
{ label: 'Paid eBooks', id: 'paid-ebooks' },
|
||||||
|
{ label: 'All eBooks', id: 'ebooks' },
|
||||||
|
],
|
||||||
|
condition: { field: 'operation', value: 'volume_search' },
|
||||||
|
mode: 'advanced',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'printType',
|
||||||
|
title: 'Print Type',
|
||||||
|
type: 'dropdown',
|
||||||
|
options: [
|
||||||
|
{ label: 'All', id: 'all' },
|
||||||
|
{ label: 'Books', id: 'books' },
|
||||||
|
{ label: 'Magazines', id: 'magazines' },
|
||||||
|
],
|
||||||
|
value: () => 'all',
|
||||||
|
condition: { field: 'operation', value: 'volume_search' },
|
||||||
|
mode: 'advanced',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'orderBy',
|
||||||
|
title: 'Order By',
|
||||||
|
type: 'dropdown',
|
||||||
|
options: [
|
||||||
|
{ label: 'Relevance', id: 'relevance' },
|
||||||
|
{ label: 'Newest', id: 'newest' },
|
||||||
|
],
|
||||||
|
value: () => 'relevance',
|
||||||
|
condition: { field: 'operation', value: 'volume_search' },
|
||||||
|
mode: 'advanced',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'maxResults',
|
||||||
|
title: 'Max Results',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Number of results (1-40)',
|
||||||
|
condition: { field: 'operation', value: 'volume_search' },
|
||||||
|
mode: 'advanced',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'startIndex',
|
||||||
|
title: 'Start Index',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Starting index for pagination',
|
||||||
|
condition: { field: 'operation', value: 'volume_search' },
|
||||||
|
mode: 'advanced',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'langRestrict',
|
||||||
|
title: 'Language',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'ISO 639-1 code (e.g., en, es, fr)',
|
||||||
|
condition: { field: 'operation', value: 'volume_search' },
|
||||||
|
mode: 'advanced',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'volumeId',
|
||||||
|
title: 'Volume ID',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Google Books volume ID',
|
||||||
|
condition: { field: 'operation', value: 'volume_details' },
|
||||||
|
required: { field: 'operation', value: 'volume_details' },
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'projection',
|
||||||
|
title: 'Projection',
|
||||||
|
type: 'dropdown',
|
||||||
|
options: [
|
||||||
|
{ label: 'Full', id: 'full' },
|
||||||
|
{ label: 'Lite', id: 'lite' },
|
||||||
|
],
|
||||||
|
value: () => 'full',
|
||||||
|
condition: { field: 'operation', value: 'volume_details' },
|
||||||
|
mode: 'advanced',
|
||||||
|
},
|
||||||
|
],
|
||||||
|
|
||||||
|
tools: {
|
||||||
|
access: ['google_books_volume_search', 'google_books_volume_details'],
|
||||||
|
config: {
|
||||||
|
tool: (params) => `google_books_${params.operation}`,
|
||||||
|
params: (params) => {
|
||||||
|
const { operation, ...rest } = params
|
||||||
|
|
||||||
|
let maxResults: number | undefined
|
||||||
|
if (params.maxResults) {
|
||||||
|
maxResults = Number.parseInt(params.maxResults, 10)
|
||||||
|
if (Number.isNaN(maxResults)) {
|
||||||
|
maxResults = undefined
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let startIndex: number | undefined
|
||||||
|
if (params.startIndex) {
|
||||||
|
startIndex = Number.parseInt(params.startIndex, 10)
|
||||||
|
if (Number.isNaN(startIndex)) {
|
||||||
|
startIndex = undefined
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
...rest,
|
||||||
|
maxResults,
|
||||||
|
startIndex,
|
||||||
|
filter: params.filter || undefined,
|
||||||
|
printType: params.printType || undefined,
|
||||||
|
orderBy: params.orderBy || undefined,
|
||||||
|
projection: params.projection || undefined,
|
||||||
|
}
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
inputs: {
|
||||||
|
operation: { type: 'string', description: 'Operation to perform' },
|
||||||
|
apiKey: { type: 'string', description: 'Google Books API key' },
|
||||||
|
query: { type: 'string', description: 'Search query' },
|
||||||
|
filter: { type: 'string', description: 'Filter by availability' },
|
||||||
|
printType: { type: 'string', description: 'Print type filter' },
|
||||||
|
orderBy: { type: 'string', description: 'Sort order' },
|
||||||
|
maxResults: { type: 'string', description: 'Maximum number of results' },
|
||||||
|
startIndex: { type: 'string', description: 'Starting index for pagination' },
|
||||||
|
langRestrict: { type: 'string', description: 'Language restriction' },
|
||||||
|
volumeId: { type: 'string', description: 'Volume ID for details' },
|
||||||
|
projection: { type: 'string', description: 'Projection level' },
|
||||||
|
},
|
||||||
|
|
||||||
|
outputs: {
|
||||||
|
totalItems: { type: 'number', description: 'Total number of matching results' },
|
||||||
|
volumes: { type: 'json', description: 'List of matching volumes' },
|
||||||
|
id: { type: 'string', description: 'Volume ID' },
|
||||||
|
title: { type: 'string', description: 'Book title' },
|
||||||
|
subtitle: { type: 'string', description: 'Book subtitle' },
|
||||||
|
authors: { type: 'json', description: 'List of authors' },
|
||||||
|
publisher: { type: 'string', description: 'Publisher name' },
|
||||||
|
publishedDate: { type: 'string', description: 'Publication date' },
|
||||||
|
description: { type: 'string', description: 'Book description' },
|
||||||
|
pageCount: { type: 'number', description: 'Number of pages' },
|
||||||
|
categories: { type: 'json', description: 'Book categories' },
|
||||||
|
averageRating: { type: 'number', description: 'Average rating (1-5)' },
|
||||||
|
ratingsCount: { type: 'number', description: 'Number of ratings' },
|
||||||
|
language: { type: 'string', description: 'Language code' },
|
||||||
|
previewLink: { type: 'string', description: 'Link to preview on Google Books' },
|
||||||
|
infoLink: { type: 'string', description: 'Link to info page' },
|
||||||
|
thumbnailUrl: { type: 'string', description: 'Book cover thumbnail URL' },
|
||||||
|
isbn10: { type: 'string', description: 'ISBN-10 identifier' },
|
||||||
|
isbn13: { type: 'string', description: 'ISBN-13 identifier' },
|
||||||
|
},
|
||||||
|
}
|
||||||
@@ -58,6 +58,16 @@ export const S3Block: BlockConfig<S3Response> = {
|
|||||||
},
|
},
|
||||||
required: true,
|
required: true,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
id: 'getObjectRegion',
|
||||||
|
title: 'AWS Region',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Used when S3 URL does not include region',
|
||||||
|
condition: {
|
||||||
|
field: 'operation',
|
||||||
|
value: ['get_object'],
|
||||||
|
},
|
||||||
|
},
|
||||||
{
|
{
|
||||||
id: 'bucketName',
|
id: 'bucketName',
|
||||||
title: 'Bucket Name',
|
title: 'Bucket Name',
|
||||||
@@ -291,34 +301,11 @@ export const S3Block: BlockConfig<S3Response> = {
|
|||||||
if (!params.s3Uri) {
|
if (!params.s3Uri) {
|
||||||
throw new Error('S3 Object URL is required')
|
throw new Error('S3 Object URL is required')
|
||||||
}
|
}
|
||||||
|
return {
|
||||||
// Parse S3 URI for get_object
|
accessKeyId: params.accessKeyId,
|
||||||
try {
|
secretAccessKey: params.secretAccessKey,
|
||||||
const url = new URL(params.s3Uri)
|
region: params.getObjectRegion || params.region,
|
||||||
const hostname = url.hostname
|
s3Uri: params.s3Uri,
|
||||||
const bucketName = hostname.split('.')[0]
|
|
||||||
const regionMatch = hostname.match(/s3[.-]([^.]+)\.amazonaws\.com/)
|
|
||||||
const region = regionMatch ? regionMatch[1] : params.region
|
|
||||||
const objectKey = url.pathname.startsWith('/')
|
|
||||||
? url.pathname.substring(1)
|
|
||||||
: url.pathname
|
|
||||||
|
|
||||||
if (!bucketName || !objectKey) {
|
|
||||||
throw new Error('Could not parse S3 URL')
|
|
||||||
}
|
|
||||||
|
|
||||||
return {
|
|
||||||
accessKeyId: params.accessKeyId,
|
|
||||||
secretAccessKey: params.secretAccessKey,
|
|
||||||
region,
|
|
||||||
bucketName,
|
|
||||||
objectKey,
|
|
||||||
s3Uri: params.s3Uri,
|
|
||||||
}
|
|
||||||
} catch (_error) {
|
|
||||||
throw new Error(
|
|
||||||
'Invalid S3 Object URL format. Expected: https://bucket-name.s3.region.amazonaws.com/path/to/file'
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -401,6 +388,7 @@ export const S3Block: BlockConfig<S3Response> = {
|
|||||||
acl: { type: 'string', description: 'Access control list' },
|
acl: { type: 'string', description: 'Access control list' },
|
||||||
// Download inputs
|
// Download inputs
|
||||||
s3Uri: { type: 'string', description: 'S3 object URL' },
|
s3Uri: { type: 'string', description: 'S3 object URL' },
|
||||||
|
getObjectRegion: { type: 'string', description: 'Optional AWS region override for downloads' },
|
||||||
// List inputs
|
// List inputs
|
||||||
prefix: { type: 'string', description: 'Prefix filter' },
|
prefix: { type: 'string', description: 'Prefix filter' },
|
||||||
maxKeys: { type: 'number', description: 'Maximum results' },
|
maxKeys: { type: 'number', description: 'Maximum results' },
|
||||||
|
|||||||
@@ -39,6 +39,7 @@ import { GitHubBlock, GitHubV2Block } from '@/blocks/blocks/github'
|
|||||||
import { GitLabBlock } from '@/blocks/blocks/gitlab'
|
import { GitLabBlock } from '@/blocks/blocks/gitlab'
|
||||||
import { GmailBlock, GmailV2Block } from '@/blocks/blocks/gmail'
|
import { GmailBlock, GmailV2Block } from '@/blocks/blocks/gmail'
|
||||||
import { GoogleSearchBlock } from '@/blocks/blocks/google'
|
import { GoogleSearchBlock } from '@/blocks/blocks/google'
|
||||||
|
import { GoogleBooksBlock } from '@/blocks/blocks/google_books'
|
||||||
import { GoogleCalendarBlock, GoogleCalendarV2Block } from '@/blocks/blocks/google_calendar'
|
import { GoogleCalendarBlock, GoogleCalendarV2Block } from '@/blocks/blocks/google_calendar'
|
||||||
import { GoogleDocsBlock } from '@/blocks/blocks/google_docs'
|
import { GoogleDocsBlock } from '@/blocks/blocks/google_docs'
|
||||||
import { GoogleDriveBlock } from '@/blocks/blocks/google_drive'
|
import { GoogleDriveBlock } from '@/blocks/blocks/google_drive'
|
||||||
@@ -214,6 +215,7 @@ export const registry: Record<string, BlockConfig> = {
|
|||||||
gmail_v2: GmailV2Block,
|
gmail_v2: GmailV2Block,
|
||||||
google_calendar: GoogleCalendarBlock,
|
google_calendar: GoogleCalendarBlock,
|
||||||
google_calendar_v2: GoogleCalendarV2Block,
|
google_calendar_v2: GoogleCalendarV2Block,
|
||||||
|
google_books: GoogleBooksBlock,
|
||||||
google_docs: GoogleDocsBlock,
|
google_docs: GoogleDocsBlock,
|
||||||
google_drive: GoogleDriveBlock,
|
google_drive: GoogleDriveBlock,
|
||||||
google_forms: GoogleFormsBlock,
|
google_forms: GoogleFormsBlock,
|
||||||
|
|||||||
@@ -196,6 +196,8 @@ export interface SubBlockConfig {
|
|||||||
type: SubBlockType
|
type: SubBlockType
|
||||||
mode?: 'basic' | 'advanced' | 'both' | 'trigger' // Default is 'both' if not specified. 'trigger' means only shown in trigger mode
|
mode?: 'basic' | 'advanced' | 'both' | 'trigger' // Default is 'both' if not specified. 'trigger' means only shown in trigger mode
|
||||||
canonicalParamId?: string
|
canonicalParamId?: string
|
||||||
|
/** Controls parameter visibility in agent/tool-input context */
|
||||||
|
paramVisibility?: 'user-or-llm' | 'user-only' | 'llm-only' | 'hidden'
|
||||||
required?:
|
required?:
|
||||||
| boolean
|
| boolean
|
||||||
| {
|
| {
|
||||||
|
|||||||
@@ -1157,6 +1157,21 @@ export function AirweaveIcon(props: SVGProps<SVGSVGElement>) {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function GoogleBooksIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
|
return (
|
||||||
|
<svg {...props} xmlns='http://www.w3.org/2000/svg' viewBox='0 0 478.633 540.068'>
|
||||||
|
<path
|
||||||
|
fill='#1C51A4'
|
||||||
|
d='M449.059,218.231L245.519,99.538l-0.061,193.23c0.031,1.504-0.368,2.977-1.166,4.204c-0.798,1.258-1.565,1.995-2.915,2.547c-1.35,0.552-2.792,0.706-4.204,0.399c-1.412-0.307-2.7-1.043-3.713-2.117l-69.166-70.609l-69.381,70.179c-1.013,0.982-2.301,1.657-3.652,1.903c-1.381,0.246-2.792,0.092-4.081-0.491c-1.289-0.583-1.626-0.522-2.394-1.749c-0.767-1.197-1.197-2.608-1.197-4.081L85.031,6.007l-2.915-1.289C43.973-11.638,0,16.409,0,59.891v420.306c0,46.029,49.312,74.782,88.775,51.767l360.285-210.138C488.491,298.782,488.491,241.246,449.059,218.231z'
|
||||||
|
/>
|
||||||
|
<path
|
||||||
|
fill='#80D7FB'
|
||||||
|
d='M88.805,8.124c-2.179-1.289-4.419-2.363-6.659-3.345l0.123,288.663c0,1.442,0.43,2.854,1.197,4.081c0.767,1.197,1.872,2.148,3.161,2.731c1.289,0.583,2.7,0.736,4.081,0.491c1.381-0.246,2.639-0.921,3.652-1.903l69.749-69.688l69.811,69.749c1.013,1.074,2.301,1.81,3.713,2.117c1.412,0.307,2.884,0.153,4.204-0.399c1.319-0.552,2.455-1.565,3.253-2.792c0.798-1.258,1.197-2.731,1.166-4.204V99.998L88.805,8.124z'
|
||||||
|
/>
|
||||||
|
</svg>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
export function GoogleDocsIcon(props: SVGProps<SVGSVGElement>) {
|
export function GoogleDocsIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
return (
|
return (
|
||||||
<svg
|
<svg
|
||||||
|
|||||||
@@ -2,8 +2,8 @@
|
|||||||
slug: enterprise
|
slug: enterprise
|
||||||
title: 'Build with Sim for Enterprise'
|
title: 'Build with Sim for Enterprise'
|
||||||
description: 'Access control, BYOK, self-hosted deployments, on-prem Copilot, SSO & SAML, whitelabeling, Admin API, and flexible data retention—enterprise features for teams with strict security and compliance requirements.'
|
description: 'Access control, BYOK, self-hosted deployments, on-prem Copilot, SSO & SAML, whitelabeling, Admin API, and flexible data retention—enterprise features for teams with strict security and compliance requirements.'
|
||||||
date: 2026-01-23
|
date: 2026-02-11
|
||||||
updated: 2026-01-23
|
updated: 2026-02-11
|
||||||
authors:
|
authors:
|
||||||
- vik
|
- vik
|
||||||
readingTime: 10
|
readingTime: 10
|
||||||
@@ -13,8 +13,8 @@ ogAlt: 'Sim Enterprise features overview'
|
|||||||
about: ['Enterprise Software', 'Security', 'Compliance', 'Self-Hosting']
|
about: ['Enterprise Software', 'Security', 'Compliance', 'Self-Hosting']
|
||||||
timeRequired: PT10M
|
timeRequired: PT10M
|
||||||
canonical: https://sim.ai/studio/enterprise
|
canonical: https://sim.ai/studio/enterprise
|
||||||
featured: false
|
featured: true
|
||||||
draft: true
|
draft: false
|
||||||
---
|
---
|
||||||
|
|
||||||
We've been working with security teams at larger organizations to bring Sim into environments with strict compliance and data handling requirements. This post covers the enterprise capabilities we've built: granular access control, bring-your-own-keys, self-hosted deployments, on-prem Copilot, SSO & SAML, whitelabeling, compliance, and programmatic management via the Admin API.
|
We've been working with security teams at larger organizations to bring Sim into environments with strict compliance and data handling requirements. This post covers the enterprise capabilities we've built: granular access control, bring-your-own-keys, self-hosted deployments, on-prem Copilot, SSO & SAML, whitelabeling, compliance, and programmatic management via the Admin API.
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
import { setupGlobalFetchMock } from '@sim/testing'
|
||||||
import { afterEach, beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
import { afterEach, beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
||||||
import { getAllBlocks } from '@/blocks'
|
import { getAllBlocks } from '@/blocks'
|
||||||
import { BlockType, isMcpTool } from '@/executor/constants'
|
import { BlockType, isMcpTool } from '@/executor/constants'
|
||||||
@@ -61,6 +62,30 @@ vi.mock('@/providers', () => ({
|
|||||||
}),
|
}),
|
||||||
}))
|
}))
|
||||||
|
|
||||||
|
vi.mock('@/executor/utils/http', () => ({
|
||||||
|
buildAuthHeaders: vi.fn().mockResolvedValue({ 'Content-Type': 'application/json' }),
|
||||||
|
buildAPIUrl: vi.fn((path: string, params?: Record<string, string>) => {
|
||||||
|
const url = new URL(path, 'http://localhost:3000')
|
||||||
|
if (params) {
|
||||||
|
for (const [key, value] of Object.entries(params)) {
|
||||||
|
if (value !== undefined && value !== null) {
|
||||||
|
url.searchParams.set(key, value)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return url
|
||||||
|
}),
|
||||||
|
extractAPIErrorMessage: vi.fn(async (response: Response) => {
|
||||||
|
const defaultMessage = `API request failed with status ${response.status}`
|
||||||
|
try {
|
||||||
|
const errorData = await response.json()
|
||||||
|
return errorData.error || defaultMessage
|
||||||
|
} catch {
|
||||||
|
return defaultMessage
|
||||||
|
}
|
||||||
|
}),
|
||||||
|
}))
|
||||||
|
|
||||||
vi.mock('@sim/db', () => ({
|
vi.mock('@sim/db', () => ({
|
||||||
db: {
|
db: {
|
||||||
select: vi.fn().mockReturnValue({
|
select: vi.fn().mockReturnValue({
|
||||||
@@ -84,7 +109,7 @@ vi.mock('@sim/db/schema', () => ({
|
|||||||
},
|
},
|
||||||
}))
|
}))
|
||||||
|
|
||||||
global.fetch = Object.assign(vi.fn(), { preconnect: vi.fn() }) as typeof fetch
|
setupGlobalFetchMock()
|
||||||
|
|
||||||
const mockGetAllBlocks = getAllBlocks as Mock
|
const mockGetAllBlocks = getAllBlocks as Mock
|
||||||
const mockExecuteTool = executeTool as Mock
|
const mockExecuteTool = executeTool as Mock
|
||||||
@@ -1901,5 +1926,301 @@ describe('AgentBlockHandler', () => {
|
|||||||
|
|
||||||
expect(discoveryCalls[0].url).toContain('serverId=mcp-legacy-server')
|
expect(discoveryCalls[0].url).toContain('serverId=mcp-legacy-server')
|
||||||
})
|
})
|
||||||
|
|
||||||
|
describe('customToolId resolution - DB as source of truth', () => {
|
||||||
|
const staleInlineSchema = {
|
||||||
|
function: {
|
||||||
|
name: 'formatReport',
|
||||||
|
description: 'Formats a report',
|
||||||
|
parameters: {
|
||||||
|
type: 'object',
|
||||||
|
properties: {
|
||||||
|
title: { type: 'string', description: 'Report title' },
|
||||||
|
content: { type: 'string', description: 'Report content' },
|
||||||
|
},
|
||||||
|
required: ['title', 'content'],
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
const dbSchema = {
|
||||||
|
function: {
|
||||||
|
name: 'formatReport',
|
||||||
|
description: 'Formats a report',
|
||||||
|
parameters: {
|
||||||
|
type: 'object',
|
||||||
|
properties: {
|
||||||
|
title: { type: 'string', description: 'Report title' },
|
||||||
|
content: { type: 'string', description: 'Report content' },
|
||||||
|
format: { type: 'string', description: 'Output format' },
|
||||||
|
},
|
||||||
|
required: ['title', 'content', 'format'],
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
const staleInlineCode = 'return { title, content };'
|
||||||
|
const dbCode = 'return { title, content, format };'
|
||||||
|
|
||||||
|
function mockFetchForCustomTool(toolId: string) {
|
||||||
|
mockFetch.mockImplementation((url: string) => {
|
||||||
|
if (typeof url === 'string' && url.includes('/api/tools/custom')) {
|
||||||
|
return Promise.resolve({
|
||||||
|
ok: true,
|
||||||
|
headers: { get: () => null },
|
||||||
|
json: () =>
|
||||||
|
Promise.resolve({
|
||||||
|
data: [
|
||||||
|
{
|
||||||
|
id: toolId,
|
||||||
|
title: 'formatReport',
|
||||||
|
schema: dbSchema,
|
||||||
|
code: dbCode,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
return Promise.resolve({
|
||||||
|
ok: true,
|
||||||
|
headers: { get: () => null },
|
||||||
|
json: () => Promise.resolve({}),
|
||||||
|
})
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
function mockFetchFailure() {
|
||||||
|
mockFetch.mockImplementation((url: string) => {
|
||||||
|
if (typeof url === 'string' && url.includes('/api/tools/custom')) {
|
||||||
|
return Promise.resolve({
|
||||||
|
ok: false,
|
||||||
|
status: 500,
|
||||||
|
headers: { get: () => null },
|
||||||
|
json: () => Promise.resolve({}),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
return Promise.resolve({
|
||||||
|
ok: true,
|
||||||
|
headers: { get: () => null },
|
||||||
|
json: () => Promise.resolve({}),
|
||||||
|
})
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
Object.defineProperty(global, 'window', {
|
||||||
|
value: undefined,
|
||||||
|
writable: true,
|
||||||
|
configurable: true,
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should always fetch latest schema from DB when customToolId is present', async () => {
|
||||||
|
const toolId = 'custom-tool-123'
|
||||||
|
mockFetchForCustomTool(toolId)
|
||||||
|
|
||||||
|
const inputs = {
|
||||||
|
model: 'gpt-4o',
|
||||||
|
userPrompt: 'Format a report',
|
||||||
|
apiKey: 'test-api-key',
|
||||||
|
tools: [
|
||||||
|
{
|
||||||
|
type: 'custom-tool',
|
||||||
|
customToolId: toolId,
|
||||||
|
title: 'formatReport',
|
||||||
|
schema: staleInlineSchema,
|
||||||
|
code: staleInlineCode,
|
||||||
|
usageControl: 'auto' as const,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}
|
||||||
|
|
||||||
|
mockGetProviderFromModel.mockReturnValue('openai')
|
||||||
|
|
||||||
|
await handler.execute(mockContext, mockBlock, inputs)
|
||||||
|
|
||||||
|
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||||
|
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||||
|
const tools = providerCall[1].tools
|
||||||
|
|
||||||
|
expect(tools.length).toBe(1)
|
||||||
|
// DB schema wins over stale inline — includes format param
|
||||||
|
expect(tools[0].parameters.required).toContain('format')
|
||||||
|
expect(tools[0].parameters.properties).toHaveProperty('format')
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should fetch from DB when customToolId has no inline schema', async () => {
|
||||||
|
const toolId = 'custom-tool-123'
|
||||||
|
mockFetchForCustomTool(toolId)
|
||||||
|
|
||||||
|
const inputs = {
|
||||||
|
model: 'gpt-4o',
|
||||||
|
userPrompt: 'Format a report',
|
||||||
|
apiKey: 'test-api-key',
|
||||||
|
tools: [
|
||||||
|
{
|
||||||
|
type: 'custom-tool',
|
||||||
|
customToolId: toolId,
|
||||||
|
usageControl: 'auto' as const,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}
|
||||||
|
|
||||||
|
mockGetProviderFromModel.mockReturnValue('openai')
|
||||||
|
|
||||||
|
await handler.execute(mockContext, mockBlock, inputs)
|
||||||
|
|
||||||
|
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||||
|
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||||
|
const tools = providerCall[1].tools
|
||||||
|
|
||||||
|
expect(tools.length).toBe(1)
|
||||||
|
expect(tools[0].name).toBe('formatReport')
|
||||||
|
expect(tools[0].parameters.required).toContain('format')
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should fall back to inline schema when DB fetch fails and inline exists', async () => {
|
||||||
|
mockFetchFailure()
|
||||||
|
|
||||||
|
const inputs = {
|
||||||
|
model: 'gpt-4o',
|
||||||
|
userPrompt: 'Format a report',
|
||||||
|
apiKey: 'test-api-key',
|
||||||
|
tools: [
|
||||||
|
{
|
||||||
|
type: 'custom-tool',
|
||||||
|
customToolId: 'custom-tool-123',
|
||||||
|
title: 'formatReport',
|
||||||
|
schema: staleInlineSchema,
|
||||||
|
code: staleInlineCode,
|
||||||
|
usageControl: 'auto' as const,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}
|
||||||
|
|
||||||
|
mockGetProviderFromModel.mockReturnValue('openai')
|
||||||
|
|
||||||
|
await handler.execute(mockContext, mockBlock, inputs)
|
||||||
|
|
||||||
|
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||||
|
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||||
|
const tools = providerCall[1].tools
|
||||||
|
|
||||||
|
expect(tools.length).toBe(1)
|
||||||
|
expect(tools[0].name).toBe('formatReport')
|
||||||
|
expect(tools[0].parameters.required).not.toContain('format')
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should return null when DB fetch fails and no inline schema exists', async () => {
|
||||||
|
mockFetchFailure()
|
||||||
|
|
||||||
|
const inputs = {
|
||||||
|
model: 'gpt-4o',
|
||||||
|
userPrompt: 'Format a report',
|
||||||
|
apiKey: 'test-api-key',
|
||||||
|
tools: [
|
||||||
|
{
|
||||||
|
type: 'custom-tool',
|
||||||
|
customToolId: 'custom-tool-123',
|
||||||
|
usageControl: 'auto' as const,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}
|
||||||
|
|
||||||
|
mockGetProviderFromModel.mockReturnValue('openai')
|
||||||
|
|
||||||
|
await handler.execute(mockContext, mockBlock, inputs)
|
||||||
|
|
||||||
|
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||||
|
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||||
|
const tools = providerCall[1].tools
|
||||||
|
|
||||||
|
expect(tools.length).toBe(0)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should use DB code for executeFunction when customToolId resolves', async () => {
|
||||||
|
const toolId = 'custom-tool-123'
|
||||||
|
mockFetchForCustomTool(toolId)
|
||||||
|
|
||||||
|
let capturedTools: any[] = []
|
||||||
|
Promise.all = vi.fn().mockImplementation((promises: Promise<any>[]) => {
|
||||||
|
const result = originalPromiseAll.call(Promise, promises)
|
||||||
|
result.then((tools: any[]) => {
|
||||||
|
if (tools?.length) {
|
||||||
|
capturedTools = tools.filter((t) => t !== null)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
return result
|
||||||
|
})
|
||||||
|
|
||||||
|
const inputs = {
|
||||||
|
model: 'gpt-4o',
|
||||||
|
userPrompt: 'Format a report',
|
||||||
|
apiKey: 'test-api-key',
|
||||||
|
tools: [
|
||||||
|
{
|
||||||
|
type: 'custom-tool',
|
||||||
|
customToolId: toolId,
|
||||||
|
title: 'formatReport',
|
||||||
|
schema: staleInlineSchema,
|
||||||
|
code: staleInlineCode,
|
||||||
|
usageControl: 'auto' as const,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}
|
||||||
|
|
||||||
|
mockGetProviderFromModel.mockReturnValue('openai')
|
||||||
|
|
||||||
|
await handler.execute(mockContext, mockBlock, inputs)
|
||||||
|
|
||||||
|
expect(capturedTools.length).toBe(1)
|
||||||
|
expect(typeof capturedTools[0].executeFunction).toBe('function')
|
||||||
|
|
||||||
|
await capturedTools[0].executeFunction({ title: 'Q1', format: 'pdf' })
|
||||||
|
|
||||||
|
expect(mockExecuteTool).toHaveBeenCalledWith(
|
||||||
|
'function_execute',
|
||||||
|
expect.objectContaining({
|
||||||
|
code: dbCode,
|
||||||
|
}),
|
||||||
|
false,
|
||||||
|
expect.any(Object)
|
||||||
|
)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should not fetch from DB when no customToolId is present', async () => {
|
||||||
|
const inputs = {
|
||||||
|
model: 'gpt-4o',
|
||||||
|
userPrompt: 'Use the tool',
|
||||||
|
apiKey: 'test-api-key',
|
||||||
|
tools: [
|
||||||
|
{
|
||||||
|
type: 'custom-tool',
|
||||||
|
title: 'formatReport',
|
||||||
|
schema: staleInlineSchema,
|
||||||
|
code: staleInlineCode,
|
||||||
|
usageControl: 'auto' as const,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}
|
||||||
|
|
||||||
|
mockGetProviderFromModel.mockReturnValue('openai')
|
||||||
|
|
||||||
|
await handler.execute(mockContext, mockBlock, inputs)
|
||||||
|
|
||||||
|
const customToolFetches = mockFetch.mock.calls.filter(
|
||||||
|
(call: any[]) => typeof call[0] === 'string' && call[0].includes('/api/tools/custom')
|
||||||
|
)
|
||||||
|
expect(customToolFetches.length).toBe(0)
|
||||||
|
|
||||||
|
expect(mockExecuteProviderRequest).toHaveBeenCalled()
|
||||||
|
const providerCall = mockExecuteProviderRequest.mock.calls[0]
|
||||||
|
const tools = providerCall[1].tools
|
||||||
|
|
||||||
|
expect(tools.length).toBe(1)
|
||||||
|
expect(tools[0].name).toBe('formatReport')
|
||||||
|
expect(tools[0].parameters.required).not.toContain('format')
|
||||||
|
})
|
||||||
|
})
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -62,9 +62,12 @@ export class AgentBlockHandler implements BlockHandler {
|
|||||||
await validateModelProvider(ctx.userId, model, ctx)
|
await validateModelProvider(ctx.userId, model, ctx)
|
||||||
|
|
||||||
const providerId = getProviderFromModel(model)
|
const providerId = getProviderFromModel(model)
|
||||||
const formattedTools = await this.formatTools(ctx, filteredInputs.tools || [])
|
const formattedTools = await this.formatTools(
|
||||||
|
ctx,
|
||||||
|
filteredInputs.tools || [],
|
||||||
|
block.canonicalModes
|
||||||
|
)
|
||||||
|
|
||||||
// Resolve skill metadata for progressive disclosure
|
|
||||||
const skillInputs = filteredInputs.skills ?? []
|
const skillInputs = filteredInputs.skills ?? []
|
||||||
let skillMetadata: Array<{ name: string; description: string }> = []
|
let skillMetadata: Array<{ name: string; description: string }> = []
|
||||||
if (skillInputs.length > 0 && ctx.workspaceId) {
|
if (skillInputs.length > 0 && ctx.workspaceId) {
|
||||||
@@ -221,7 +224,11 @@ export class AgentBlockHandler implements BlockHandler {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
private async formatTools(ctx: ExecutionContext, inputTools: ToolInput[]): Promise<any[]> {
|
private async formatTools(
|
||||||
|
ctx: ExecutionContext,
|
||||||
|
inputTools: ToolInput[],
|
||||||
|
canonicalModes?: Record<string, 'basic' | 'advanced'>
|
||||||
|
): Promise<any[]> {
|
||||||
if (!Array.isArray(inputTools)) return []
|
if (!Array.isArray(inputTools)) return []
|
||||||
|
|
||||||
const filtered = inputTools.filter((tool) => {
|
const filtered = inputTools.filter((tool) => {
|
||||||
@@ -249,7 +256,7 @@ export class AgentBlockHandler implements BlockHandler {
|
|||||||
if (tool.type === 'custom-tool' && (tool.schema || tool.customToolId)) {
|
if (tool.type === 'custom-tool' && (tool.schema || tool.customToolId)) {
|
||||||
return await this.createCustomTool(ctx, tool)
|
return await this.createCustomTool(ctx, tool)
|
||||||
}
|
}
|
||||||
return this.transformBlockTool(ctx, tool)
|
return this.transformBlockTool(ctx, tool, canonicalModes)
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(`[AgentHandler] Error creating tool:`, { tool, error })
|
logger.error(`[AgentHandler] Error creating tool:`, { tool, error })
|
||||||
return null
|
return null
|
||||||
@@ -272,15 +279,16 @@ export class AgentBlockHandler implements BlockHandler {
|
|||||||
let code = tool.code
|
let code = tool.code
|
||||||
let title = tool.title
|
let title = tool.title
|
||||||
|
|
||||||
if (tool.customToolId && !schema) {
|
if (tool.customToolId) {
|
||||||
const resolved = await this.fetchCustomToolById(ctx, tool.customToolId)
|
const resolved = await this.fetchCustomToolById(ctx, tool.customToolId)
|
||||||
if (!resolved) {
|
if (resolved) {
|
||||||
|
schema = resolved.schema
|
||||||
|
code = resolved.code
|
||||||
|
title = resolved.title
|
||||||
|
} else if (!schema) {
|
||||||
logger.error(`Custom tool not found: ${tool.customToolId}`)
|
logger.error(`Custom tool not found: ${tool.customToolId}`)
|
||||||
return null
|
return null
|
||||||
}
|
}
|
||||||
schema = resolved.schema
|
|
||||||
code = resolved.code
|
|
||||||
title = resolved.title
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!schema?.function) {
|
if (!schema?.function) {
|
||||||
@@ -719,12 +727,17 @@ export class AgentBlockHandler implements BlockHandler {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private async transformBlockTool(ctx: ExecutionContext, tool: ToolInput) {
|
private async transformBlockTool(
|
||||||
|
ctx: ExecutionContext,
|
||||||
|
tool: ToolInput,
|
||||||
|
canonicalModes?: Record<string, 'basic' | 'advanced'>
|
||||||
|
) {
|
||||||
const transformedTool = await transformBlockTool(tool, {
|
const transformedTool = await transformBlockTool(tool, {
|
||||||
selectedOperation: tool.operation,
|
selectedOperation: tool.operation,
|
||||||
getAllBlocks,
|
getAllBlocks,
|
||||||
getToolAsync: (toolId: string) => getToolAsync(toolId, ctx.workflowId),
|
getToolAsync: (toolId: string) => getToolAsync(toolId, ctx.workflowId),
|
||||||
getTool,
|
getTool,
|
||||||
|
canonicalModes,
|
||||||
})
|
})
|
||||||
|
|
||||||
if (transformedTool) {
|
if (transformedTool) {
|
||||||
@@ -999,6 +1012,7 @@ export class AgentBlockHandler implements BlockHandler {
|
|||||||
reasoningEffort: inputs.reasoningEffort,
|
reasoningEffort: inputs.reasoningEffort,
|
||||||
verbosity: inputs.verbosity,
|
verbosity: inputs.verbosity,
|
||||||
thinkingLevel: inputs.thinkingLevel,
|
thinkingLevel: inputs.thinkingLevel,
|
||||||
|
previousInteractionId: inputs.previousInteractionId,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1069,6 +1083,7 @@ export class AgentBlockHandler implements BlockHandler {
|
|||||||
reasoningEffort: providerRequest.reasoningEffort,
|
reasoningEffort: providerRequest.reasoningEffort,
|
||||||
verbosity: providerRequest.verbosity,
|
verbosity: providerRequest.verbosity,
|
||||||
thinkingLevel: providerRequest.thinkingLevel,
|
thinkingLevel: providerRequest.thinkingLevel,
|
||||||
|
previousInteractionId: providerRequest.previousInteractionId,
|
||||||
})
|
})
|
||||||
|
|
||||||
return this.processProviderResponse(response, block, responseFormat)
|
return this.processProviderResponse(response, block, responseFormat)
|
||||||
@@ -1269,6 +1284,7 @@ export class AgentBlockHandler implements BlockHandler {
|
|||||||
content: result.content,
|
content: result.content,
|
||||||
model: result.model,
|
model: result.model,
|
||||||
...this.createResponseMetadata(result),
|
...this.createResponseMetadata(result),
|
||||||
|
...(result.interactionId && { interactionId: result.interactionId }),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -20,6 +20,8 @@ export interface AgentInputs {
|
|||||||
conversationId?: string // Required for all non-none memory types
|
conversationId?: string // Required for all non-none memory types
|
||||||
slidingWindowSize?: string // For message-based sliding window
|
slidingWindowSize?: string // For message-based sliding window
|
||||||
slidingWindowTokens?: string // For token-based sliding window
|
slidingWindowTokens?: string // For token-based sliding window
|
||||||
|
// Deep research multi-turn
|
||||||
|
previousInteractionId?: string // Interactions API previous interaction reference
|
||||||
// LLM parameters
|
// LLM parameters
|
||||||
temperature?: string
|
temperature?: string
|
||||||
maxTokens?: string
|
maxTokens?: string
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ import { db } from '@sim/db'
|
|||||||
import { account } from '@sim/db/schema'
|
import { account } from '@sim/db/schema'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { eq } from 'drizzle-orm'
|
import { eq } from 'drizzle-orm'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { refreshTokenIfNeeded } from '@/app/api/auth/oauth/utils'
|
import { refreshTokenIfNeeded } from '@/app/api/auth/oauth/utils'
|
||||||
import { generateRouterPrompt, generateRouterV2Prompt } from '@/blocks/blocks/router'
|
import { generateRouterPrompt, generateRouterV2Prompt } from '@/blocks/blocks/router'
|
||||||
import type { BlockOutput } from '@/blocks/types'
|
import type { BlockOutput } from '@/blocks/types'
|
||||||
@@ -79,7 +79,7 @@ export class RouterBlockHandler implements BlockHandler {
|
|||||||
const providerId = getProviderFromModel(routerConfig.model)
|
const providerId = getProviderFromModel(routerConfig.model)
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const url = new URL('/api/providers', getBaseUrl())
|
const url = new URL('/api/providers', getInternalApiBaseUrl())
|
||||||
if (ctx.userId) url.searchParams.set('userId', ctx.userId)
|
if (ctx.userId) url.searchParams.set('userId', ctx.userId)
|
||||||
|
|
||||||
const messages = [{ role: 'user', content: routerConfig.prompt }]
|
const messages = [{ role: 'user', content: routerConfig.prompt }]
|
||||||
@@ -209,7 +209,7 @@ export class RouterBlockHandler implements BlockHandler {
|
|||||||
const providerId = getProviderFromModel(routerConfig.model)
|
const providerId = getProviderFromModel(routerConfig.model)
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const url = new URL('/api/providers', getBaseUrl())
|
const url = new URL('/api/providers', getInternalApiBaseUrl())
|
||||||
if (ctx.userId) url.searchParams.set('userId', ctx.userId)
|
if (ctx.userId) url.searchParams.set('userId', ctx.userId)
|
||||||
|
|
||||||
const messages = [{ role: 'user', content: routerConfig.context }]
|
const messages = [{ role: 'user', content: routerConfig.context }]
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
import { setupGlobalFetchMock } from '@sim/testing'
|
||||||
import { beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
import { beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
||||||
import { BlockType } from '@/executor/constants'
|
import { BlockType } from '@/executor/constants'
|
||||||
import { WorkflowBlockHandler } from '@/executor/handlers/workflow/workflow-handler'
|
import { WorkflowBlockHandler } from '@/executor/handlers/workflow/workflow-handler'
|
||||||
@@ -9,7 +10,7 @@ vi.mock('@/lib/auth/internal', () => ({
|
|||||||
}))
|
}))
|
||||||
|
|
||||||
// Mock fetch globally
|
// Mock fetch globally
|
||||||
global.fetch = vi.fn()
|
setupGlobalFetchMock()
|
||||||
|
|
||||||
describe('WorkflowBlockHandler', () => {
|
describe('WorkflowBlockHandler', () => {
|
||||||
let handler: WorkflowBlockHandler
|
let handler: WorkflowBlockHandler
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
import { generateInternalToken } from '@/lib/auth/internal'
|
import { generateInternalToken } from '@/lib/auth/internal'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getBaseUrl, getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { HTTP } from '@/executor/constants'
|
import { HTTP } from '@/executor/constants'
|
||||||
|
|
||||||
export async function buildAuthHeaders(): Promise<Record<string, string>> {
|
export async function buildAuthHeaders(): Promise<Record<string, string>> {
|
||||||
@@ -16,7 +16,8 @@ export async function buildAuthHeaders(): Promise<Record<string, string>> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export function buildAPIUrl(path: string, params?: Record<string, string>): URL {
|
export function buildAPIUrl(path: string, params?: Record<string, string>): URL {
|
||||||
const url = new URL(path, getBaseUrl())
|
const baseUrl = path.startsWith('/api/') ? getInternalApiBaseUrl() : getBaseUrl()
|
||||||
|
const url = new URL(path, baseUrl)
|
||||||
|
|
||||||
if (params) {
|
if (params) {
|
||||||
for (const [key, value] of Object.entries(params)) {
|
for (const [key, value] of Object.entries(params)) {
|
||||||
|
|||||||
@@ -423,7 +423,7 @@ interface GenerateVersionDescriptionVariables {
|
|||||||
|
|
||||||
const VERSION_DESCRIPTION_SYSTEM_PROMPT = `You are writing deployment version descriptions for a workflow automation platform.
|
const VERSION_DESCRIPTION_SYSTEM_PROMPT = `You are writing deployment version descriptions for a workflow automation platform.
|
||||||
|
|
||||||
Write a brief, factual description (1-3 sentences, under 400 characters) that states what changed between versions.
|
Write a brief, factual description (1-3 sentences, under 2000 characters) that states what changed between versions.
|
||||||
|
|
||||||
Guidelines:
|
Guidelines:
|
||||||
- Use the specific values provided (credential names, channel names, model names)
|
- Use the specific values provided (credential names, channel names, model names)
|
||||||
|
|||||||
@@ -642,6 +642,10 @@ export function useDeployChildWorkflow() {
|
|||||||
queryClient.invalidateQueries({
|
queryClient.invalidateQueries({
|
||||||
queryKey: workflowKeys.deploymentStatus(variables.workflowId),
|
queryKey: workflowKeys.deploymentStatus(variables.workflowId),
|
||||||
})
|
})
|
||||||
|
// Invalidate workflow state so tool input mappings refresh
|
||||||
|
queryClient.invalidateQueries({
|
||||||
|
queryKey: workflowKeys.state(variables.workflowId),
|
||||||
|
})
|
||||||
// Also invalidate deployment queries
|
// Also invalidate deployment queries
|
||||||
queryClient.invalidateQueries({
|
queryClient.invalidateQueries({
|
||||||
queryKey: deploymentKeys.info(variables.workflowId),
|
queryKey: deploymentKeys.info(variables.workflowId),
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { useCallback, useRef } from 'react'
|
import { useCallback } from 'react'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import type {
|
import type {
|
||||||
BlockCompletedData,
|
BlockCompletedData,
|
||||||
@@ -16,6 +16,18 @@ import type { SerializableExecutionState } from '@/executor/execution/types'
|
|||||||
|
|
||||||
const logger = createLogger('useExecutionStream')
|
const logger = createLogger('useExecutionStream')
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Detects errors caused by the browser killing a fetch (page refresh, navigation, tab close).
|
||||||
|
* These should be treated as clean disconnects, not execution errors.
|
||||||
|
*/
|
||||||
|
function isClientDisconnectError(error: any): boolean {
|
||||||
|
if (error.name === 'AbortError') return true
|
||||||
|
const msg = (error.message ?? '').toLowerCase()
|
||||||
|
return (
|
||||||
|
msg.includes('network error') || msg.includes('failed to fetch') || msg.includes('load failed')
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Processes SSE events from a response body and invokes appropriate callbacks.
|
* Processes SSE events from a response body and invokes appropriate callbacks.
|
||||||
*/
|
*/
|
||||||
@@ -121,6 +133,7 @@ export interface ExecuteStreamOptions {
|
|||||||
parallels?: Record<string, any>
|
parallels?: Record<string, any>
|
||||||
}
|
}
|
||||||
stopAfterBlockId?: string
|
stopAfterBlockId?: string
|
||||||
|
onExecutionId?: (executionId: string) => void
|
||||||
callbacks?: ExecutionStreamCallbacks
|
callbacks?: ExecutionStreamCallbacks
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -129,30 +142,40 @@ export interface ExecuteFromBlockOptions {
|
|||||||
startBlockId: string
|
startBlockId: string
|
||||||
sourceSnapshot: SerializableExecutionState
|
sourceSnapshot: SerializableExecutionState
|
||||||
input?: any
|
input?: any
|
||||||
|
onExecutionId?: (executionId: string) => void
|
||||||
callbacks?: ExecutionStreamCallbacks
|
callbacks?: ExecutionStreamCallbacks
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export interface ReconnectStreamOptions {
|
||||||
|
workflowId: string
|
||||||
|
executionId: string
|
||||||
|
fromEventId?: number
|
||||||
|
callbacks?: ExecutionStreamCallbacks
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Module-level map shared across all hook instances.
|
||||||
|
* Ensures ANY instance can cancel streams started by ANY other instance,
|
||||||
|
* which is critical for SPA navigation where the original hook instance unmounts
|
||||||
|
* but the SSE stream must be cancellable from the new instance.
|
||||||
|
*/
|
||||||
|
const sharedAbortControllers = new Map<string, AbortController>()
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Hook for executing workflows via server-side SSE streaming.
|
* Hook for executing workflows via server-side SSE streaming.
|
||||||
* Supports concurrent executions via per-workflow AbortController maps.
|
* Supports concurrent executions via per-workflow AbortController maps.
|
||||||
*/
|
*/
|
||||||
export function useExecutionStream() {
|
export function useExecutionStream() {
|
||||||
const abortControllersRef = useRef<Map<string, AbortController>>(new Map())
|
|
||||||
const currentExecutionsRef = useRef<Map<string, { workflowId: string; executionId: string }>>(
|
|
||||||
new Map()
|
|
||||||
)
|
|
||||||
|
|
||||||
const execute = useCallback(async (options: ExecuteStreamOptions) => {
|
const execute = useCallback(async (options: ExecuteStreamOptions) => {
|
||||||
const { workflowId, callbacks = {}, ...payload } = options
|
const { workflowId, callbacks = {}, onExecutionId, ...payload } = options
|
||||||
|
|
||||||
const existing = abortControllersRef.current.get(workflowId)
|
const existing = sharedAbortControllers.get(workflowId)
|
||||||
if (existing) {
|
if (existing) {
|
||||||
existing.abort()
|
existing.abort()
|
||||||
}
|
}
|
||||||
|
|
||||||
const abortController = new AbortController()
|
const abortController = new AbortController()
|
||||||
abortControllersRef.current.set(workflowId, abortController)
|
sharedAbortControllers.set(workflowId, abortController)
|
||||||
currentExecutionsRef.current.delete(workflowId)
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
||||||
@@ -177,42 +200,48 @@ export function useExecutionStream() {
|
|||||||
throw new Error('No response body')
|
throw new Error('No response body')
|
||||||
}
|
}
|
||||||
|
|
||||||
const executionId = response.headers.get('X-Execution-Id')
|
const serverExecutionId = response.headers.get('X-Execution-Id')
|
||||||
if (executionId) {
|
if (serverExecutionId) {
|
||||||
currentExecutionsRef.current.set(workflowId, { workflowId, executionId })
|
onExecutionId?.(serverExecutionId)
|
||||||
}
|
}
|
||||||
|
|
||||||
const reader = response.body.getReader()
|
const reader = response.body.getReader()
|
||||||
await processSSEStream(reader, callbacks, 'Execution')
|
await processSSEStream(reader, callbacks, 'Execution')
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
if (error.name === 'AbortError') {
|
if (isClientDisconnectError(error)) {
|
||||||
logger.info('Execution stream cancelled')
|
logger.info('Execution stream disconnected (page unload or abort)')
|
||||||
callbacks.onExecutionCancelled?.({ duration: 0 })
|
return
|
||||||
} else {
|
|
||||||
logger.error('Execution stream error:', error)
|
|
||||||
callbacks.onExecutionError?.({
|
|
||||||
error: error.message || 'Unknown error',
|
|
||||||
duration: 0,
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
|
logger.error('Execution stream error:', error)
|
||||||
|
callbacks.onExecutionError?.({
|
||||||
|
error: error.message || 'Unknown error',
|
||||||
|
duration: 0,
|
||||||
|
})
|
||||||
throw error
|
throw error
|
||||||
} finally {
|
} finally {
|
||||||
abortControllersRef.current.delete(workflowId)
|
if (sharedAbortControllers.get(workflowId) === abortController) {
|
||||||
currentExecutionsRef.current.delete(workflowId)
|
sharedAbortControllers.delete(workflowId)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}, [])
|
}, [])
|
||||||
|
|
||||||
const executeFromBlock = useCallback(async (options: ExecuteFromBlockOptions) => {
|
const executeFromBlock = useCallback(async (options: ExecuteFromBlockOptions) => {
|
||||||
const { workflowId, startBlockId, sourceSnapshot, input, callbacks = {} } = options
|
const {
|
||||||
|
workflowId,
|
||||||
|
startBlockId,
|
||||||
|
sourceSnapshot,
|
||||||
|
input,
|
||||||
|
onExecutionId,
|
||||||
|
callbacks = {},
|
||||||
|
} = options
|
||||||
|
|
||||||
const existing = abortControllersRef.current.get(workflowId)
|
const existing = sharedAbortControllers.get(workflowId)
|
||||||
if (existing) {
|
if (existing) {
|
||||||
existing.abort()
|
existing.abort()
|
||||||
}
|
}
|
||||||
|
|
||||||
const abortController = new AbortController()
|
const abortController = new AbortController()
|
||||||
abortControllersRef.current.set(workflowId, abortController)
|
sharedAbortControllers.set(workflowId, abortController)
|
||||||
currentExecutionsRef.current.delete(workflowId)
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
const response = await fetch(`/api/workflows/${workflowId}/execute`, {
|
||||||
@@ -246,64 +275,80 @@ export function useExecutionStream() {
|
|||||||
throw new Error('No response body')
|
throw new Error('No response body')
|
||||||
}
|
}
|
||||||
|
|
||||||
const executionId = response.headers.get('X-Execution-Id')
|
const serverExecutionId = response.headers.get('X-Execution-Id')
|
||||||
if (executionId) {
|
if (serverExecutionId) {
|
||||||
currentExecutionsRef.current.set(workflowId, { workflowId, executionId })
|
onExecutionId?.(serverExecutionId)
|
||||||
}
|
}
|
||||||
|
|
||||||
const reader = response.body.getReader()
|
const reader = response.body.getReader()
|
||||||
await processSSEStream(reader, callbacks, 'Run-from-block')
|
await processSSEStream(reader, callbacks, 'Run-from-block')
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
if (error.name === 'AbortError') {
|
if (isClientDisconnectError(error)) {
|
||||||
logger.info('Run-from-block execution cancelled')
|
logger.info('Run-from-block stream disconnected (page unload or abort)')
|
||||||
callbacks.onExecutionCancelled?.({ duration: 0 })
|
return
|
||||||
} else {
|
|
||||||
logger.error('Run-from-block execution error:', error)
|
|
||||||
callbacks.onExecutionError?.({
|
|
||||||
error: error.message || 'Unknown error',
|
|
||||||
duration: 0,
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
|
logger.error('Run-from-block execution error:', error)
|
||||||
|
callbacks.onExecutionError?.({
|
||||||
|
error: error.message || 'Unknown error',
|
||||||
|
duration: 0,
|
||||||
|
})
|
||||||
throw error
|
throw error
|
||||||
} finally {
|
} finally {
|
||||||
abortControllersRef.current.delete(workflowId)
|
if (sharedAbortControllers.get(workflowId) === abortController) {
|
||||||
currentExecutionsRef.current.delete(workflowId)
|
sharedAbortControllers.delete(workflowId)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}, [])
|
||||||
|
|
||||||
|
const reconnect = useCallback(async (options: ReconnectStreamOptions) => {
|
||||||
|
const { workflowId, executionId, fromEventId = 0, callbacks = {} } = options
|
||||||
|
|
||||||
|
const existing = sharedAbortControllers.get(workflowId)
|
||||||
|
if (existing) {
|
||||||
|
existing.abort()
|
||||||
|
}
|
||||||
|
|
||||||
|
const abortController = new AbortController()
|
||||||
|
sharedAbortControllers.set(workflowId, abortController)
|
||||||
|
try {
|
||||||
|
const response = await fetch(
|
||||||
|
`/api/workflows/${workflowId}/executions/${executionId}/stream?from=${fromEventId}`,
|
||||||
|
{ signal: abortController.signal }
|
||||||
|
)
|
||||||
|
if (!response.ok) throw new Error(`Reconnect failed (${response.status})`)
|
||||||
|
if (!response.body) throw new Error('No response body')
|
||||||
|
|
||||||
|
await processSSEStream(response.body.getReader(), callbacks, 'Reconnect')
|
||||||
|
} catch (error: any) {
|
||||||
|
if (isClientDisconnectError(error)) return
|
||||||
|
logger.error('Reconnection stream error:', error)
|
||||||
|
throw error
|
||||||
|
} finally {
|
||||||
|
if (sharedAbortControllers.get(workflowId) === abortController) {
|
||||||
|
sharedAbortControllers.delete(workflowId)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}, [])
|
}, [])
|
||||||
|
|
||||||
const cancel = useCallback((workflowId?: string) => {
|
const cancel = useCallback((workflowId?: string) => {
|
||||||
if (workflowId) {
|
if (workflowId) {
|
||||||
const execution = currentExecutionsRef.current.get(workflowId)
|
const controller = sharedAbortControllers.get(workflowId)
|
||||||
if (execution) {
|
|
||||||
fetch(`/api/workflows/${execution.workflowId}/executions/${execution.executionId}/cancel`, {
|
|
||||||
method: 'POST',
|
|
||||||
}).catch(() => {})
|
|
||||||
}
|
|
||||||
|
|
||||||
const controller = abortControllersRef.current.get(workflowId)
|
|
||||||
if (controller) {
|
if (controller) {
|
||||||
controller.abort()
|
controller.abort()
|
||||||
abortControllersRef.current.delete(workflowId)
|
sharedAbortControllers.delete(workflowId)
|
||||||
}
|
}
|
||||||
currentExecutionsRef.current.delete(workflowId)
|
|
||||||
} else {
|
} else {
|
||||||
for (const [, execution] of currentExecutionsRef.current) {
|
for (const [, controller] of sharedAbortControllers) {
|
||||||
fetch(`/api/workflows/${execution.workflowId}/executions/${execution.executionId}/cancel`, {
|
|
||||||
method: 'POST',
|
|
||||||
}).catch(() => {})
|
|
||||||
}
|
|
||||||
|
|
||||||
for (const [, controller] of abortControllersRef.current) {
|
|
||||||
controller.abort()
|
controller.abort()
|
||||||
}
|
}
|
||||||
abortControllersRef.current.clear()
|
sharedAbortControllers.clear()
|
||||||
currentExecutionsRef.current.clear()
|
|
||||||
}
|
}
|
||||||
}, [])
|
}, [])
|
||||||
|
|
||||||
return {
|
return {
|
||||||
execute,
|
execute,
|
||||||
executeFromBlock,
|
executeFromBlock,
|
||||||
|
reconnect,
|
||||||
cancel,
|
cancel,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
46
apps/sim/hooks/use-referral-attribution.ts
Normal file
46
apps/sim/hooks/use-referral-attribution.ts
Normal file
@@ -0,0 +1,46 @@
|
|||||||
|
'use client'
|
||||||
|
|
||||||
|
import { useEffect, useRef } from 'react'
|
||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
|
||||||
|
const logger = createLogger('ReferralAttribution')
|
||||||
|
|
||||||
|
const COOKIE_NAME = 'sim_utm'
|
||||||
|
|
||||||
|
const TERMINAL_REASONS = new Set([
|
||||||
|
'invalid_cookie',
|
||||||
|
'no_utm_cookie',
|
||||||
|
'no_matching_campaign',
|
||||||
|
'already_attributed',
|
||||||
|
])
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Fires a one-shot `POST /api/attribution` when a `sim_utm` cookie is present.
|
||||||
|
* Retries on transient failures; stops on terminal outcomes.
|
||||||
|
*/
|
||||||
|
export function useReferralAttribution() {
|
||||||
|
const calledRef = useRef(false)
|
||||||
|
|
||||||
|
useEffect(() => {
|
||||||
|
if (calledRef.current) return
|
||||||
|
if (!document.cookie.includes(COOKIE_NAME)) return
|
||||||
|
|
||||||
|
calledRef.current = true
|
||||||
|
|
||||||
|
fetch('/api/attribution', { method: 'POST' })
|
||||||
|
.then((res) => res.json())
|
||||||
|
.then((data) => {
|
||||||
|
if (data.attributed) {
|
||||||
|
logger.info('Referral attribution successful', { bonusAmount: data.bonusAmount })
|
||||||
|
} else if (data.error || TERMINAL_REASONS.has(data.reason)) {
|
||||||
|
logger.info('Referral attribution skipped', { reason: data.reason || data.error })
|
||||||
|
} else {
|
||||||
|
calledRef.current = false
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.catch((err) => {
|
||||||
|
logger.warn('Referral attribution failed, will retry', { error: err })
|
||||||
|
calledRef.current = false
|
||||||
|
})
|
||||||
|
}, [])
|
||||||
|
}
|
||||||
64
apps/sim/lib/billing/credits/bonus.ts
Normal file
64
apps/sim/lib/billing/credits/bonus.ts
Normal file
@@ -0,0 +1,64 @@
|
|||||||
|
import { db } from '@sim/db'
|
||||||
|
import { organization, userStats } from '@sim/db/schema'
|
||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { eq, sql } from 'drizzle-orm'
|
||||||
|
import { getHighestPrioritySubscription } from '@/lib/billing/core/subscription'
|
||||||
|
import type { DbOrTx } from '@/lib/db/types'
|
||||||
|
|
||||||
|
const logger = createLogger('BonusCredits')
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Apply bonus credits to a user (e.g. referral bonuses, promotional codes).
|
||||||
|
*
|
||||||
|
* Detects the user's current plan and routes credits accordingly:
|
||||||
|
* - Free/Pro: adds to `userStats.creditBalance` and increments `currentUsageLimit`
|
||||||
|
* - Team/Enterprise: adds to `organization.creditBalance` and increments `orgUsageLimit`
|
||||||
|
*
|
||||||
|
* Uses direct increment (not recalculation) so it works correctly for free-tier
|
||||||
|
* users where `setUsageLimitForCredits` would compute planBase=0 and skip the update.
|
||||||
|
*
|
||||||
|
* @param tx - Optional Drizzle transaction context. When provided, all DB writes
|
||||||
|
* participate in the caller's transaction for atomicity.
|
||||||
|
*/
|
||||||
|
export async function applyBonusCredits(
|
||||||
|
userId: string,
|
||||||
|
amount: number,
|
||||||
|
tx?: DbOrTx
|
||||||
|
): Promise<void> {
|
||||||
|
const dbCtx = tx ?? db
|
||||||
|
const subscription = await getHighestPrioritySubscription(userId)
|
||||||
|
const isTeamOrEnterprise = subscription?.plan === 'team' || subscription?.plan === 'enterprise'
|
||||||
|
|
||||||
|
if (isTeamOrEnterprise && subscription?.referenceId) {
|
||||||
|
const orgId = subscription.referenceId
|
||||||
|
|
||||||
|
await dbCtx
|
||||||
|
.update(organization)
|
||||||
|
.set({
|
||||||
|
creditBalance: sql`${organization.creditBalance} + ${amount}`,
|
||||||
|
orgUsageLimit: sql`COALESCE(${organization.orgUsageLimit}, '0')::decimal + ${amount}`,
|
||||||
|
})
|
||||||
|
.where(eq(organization.id, orgId))
|
||||||
|
|
||||||
|
logger.info('Applied bonus credits to organization', {
|
||||||
|
userId,
|
||||||
|
organizationId: orgId,
|
||||||
|
plan: subscription.plan,
|
||||||
|
amount,
|
||||||
|
})
|
||||||
|
} else {
|
||||||
|
await dbCtx
|
||||||
|
.update(userStats)
|
||||||
|
.set({
|
||||||
|
creditBalance: sql`${userStats.creditBalance} + ${amount}`,
|
||||||
|
currentUsageLimit: sql`COALESCE(${userStats.currentUsageLimit}, '0')::decimal + ${amount}`,
|
||||||
|
})
|
||||||
|
.where(eq(userStats.userId, userId))
|
||||||
|
|
||||||
|
logger.info('Applied bonus credits to user', {
|
||||||
|
userId,
|
||||||
|
plan: subscription?.plan || 'free',
|
||||||
|
amount,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -20,6 +20,8 @@ export interface BuildPayloadParams {
|
|||||||
fileAttachments?: Array<{ id: string; key: string; size: number; [key: string]: unknown }>
|
fileAttachments?: Array<{ id: string; key: string; size: number; [key: string]: unknown }>
|
||||||
commands?: string[]
|
commands?: string[]
|
||||||
chatId?: string
|
chatId?: string
|
||||||
|
conversationId?: string
|
||||||
|
prefetch?: boolean
|
||||||
implicitFeedback?: string
|
implicitFeedback?: string
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -64,6 +66,10 @@ export async function buildCopilotRequestPayload(
|
|||||||
fileAttachments,
|
fileAttachments,
|
||||||
commands,
|
commands,
|
||||||
chatId,
|
chatId,
|
||||||
|
conversationId,
|
||||||
|
prefetch,
|
||||||
|
conversationHistory,
|
||||||
|
implicitFeedback,
|
||||||
} = params
|
} = params
|
||||||
|
|
||||||
const selectedModel = options.selectedModel
|
const selectedModel = options.selectedModel
|
||||||
@@ -154,6 +160,12 @@ export async function buildCopilotRequestPayload(
|
|||||||
version: SIM_AGENT_VERSION,
|
version: SIM_AGENT_VERSION,
|
||||||
...(contexts && contexts.length > 0 ? { context: contexts } : {}),
|
...(contexts && contexts.length > 0 ? { context: contexts } : {}),
|
||||||
...(chatId ? { chatId } : {}),
|
...(chatId ? { chatId } : {}),
|
||||||
|
...(conversationId ? { conversationId } : {}),
|
||||||
|
...(Array.isArray(conversationHistory) && conversationHistory.length > 0
|
||||||
|
? { conversationHistory }
|
||||||
|
: {}),
|
||||||
|
...(typeof prefetch === 'boolean' ? { prefetch } : {}),
|
||||||
|
...(implicitFeedback ? { implicitFeedback } : {}),
|
||||||
...(processedFileContents.length > 0 ? { fileAttachments: processedFileContents } : {}),
|
...(processedFileContents.length > 0 ? { fileAttachments: processedFileContents } : {}),
|
||||||
...(integrationTools.length > 0 ? { integrationTools } : {}),
|
...(integrationTools.length > 0 ? { integrationTools } : {}),
|
||||||
...(credentials ? { credentials } : {}),
|
...(credentials ? { credentials } : {}),
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import { serializeMessagesForDB } from './serialization'
|
|||||||
|
|
||||||
const logger = createLogger('CopilotMessagePersistence')
|
const logger = createLogger('CopilotMessagePersistence')
|
||||||
|
|
||||||
interface PersistParams {
|
export async function persistMessages(params: {
|
||||||
chatId: string
|
chatId: string
|
||||||
messages: CopilotMessage[]
|
messages: CopilotMessage[]
|
||||||
sensitiveCredentialIds?: Set<string>
|
sensitiveCredentialIds?: Set<string>
|
||||||
@@ -13,29 +13,24 @@ interface PersistParams {
|
|||||||
mode?: string
|
mode?: string
|
||||||
model?: string
|
model?: string
|
||||||
conversationId?: string
|
conversationId?: string
|
||||||
}
|
}): Promise<boolean> {
|
||||||
|
|
||||||
/** Builds the JSON body used by both fetch and sendBeacon persistence paths. */
|
|
||||||
function buildPersistBody(params: PersistParams): string {
|
|
||||||
const dbMessages = serializeMessagesForDB(
|
|
||||||
params.messages,
|
|
||||||
params.sensitiveCredentialIds ?? new Set<string>()
|
|
||||||
)
|
|
||||||
return JSON.stringify({
|
|
||||||
chatId: params.chatId,
|
|
||||||
messages: dbMessages,
|
|
||||||
...(params.planArtifact !== undefined ? { planArtifact: params.planArtifact } : {}),
|
|
||||||
...(params.mode || params.model ? { config: { mode: params.mode, model: params.model } } : {}),
|
|
||||||
...(params.conversationId ? { conversationId: params.conversationId } : {}),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
export async function persistMessages(params: PersistParams): Promise<boolean> {
|
|
||||||
try {
|
try {
|
||||||
|
const dbMessages = serializeMessagesForDB(
|
||||||
|
params.messages,
|
||||||
|
params.sensitiveCredentialIds ?? new Set<string>()
|
||||||
|
)
|
||||||
const response = await fetch(COPILOT_UPDATE_MESSAGES_API_PATH, {
|
const response = await fetch(COPILOT_UPDATE_MESSAGES_API_PATH, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
headers: { 'Content-Type': 'application/json' },
|
headers: { 'Content-Type': 'application/json' },
|
||||||
body: buildPersistBody(params),
|
body: JSON.stringify({
|
||||||
|
chatId: params.chatId,
|
||||||
|
messages: dbMessages,
|
||||||
|
...(params.planArtifact !== undefined ? { planArtifact: params.planArtifact } : {}),
|
||||||
|
...(params.mode || params.model
|
||||||
|
? { config: { mode: params.mode, model: params.model } }
|
||||||
|
: {}),
|
||||||
|
...(params.conversationId ? { conversationId: params.conversationId } : {}),
|
||||||
|
}),
|
||||||
})
|
})
|
||||||
return response.ok
|
return response.ok
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
@@ -46,27 +41,3 @@ export async function persistMessages(params: PersistParams): Promise<boolean> {
|
|||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Persists messages using navigator.sendBeacon, which is reliable during page unload.
|
|
||||||
* Unlike fetch, sendBeacon is guaranteed to be queued even when the page is being torn down.
|
|
||||||
*/
|
|
||||||
export function persistMessagesBeacon(params: PersistParams): boolean {
|
|
||||||
try {
|
|
||||||
const body = buildPersistBody(params)
|
|
||||||
const blob = new Blob([body], { type: 'application/json' })
|
|
||||||
const sent = navigator.sendBeacon(COPILOT_UPDATE_MESSAGES_API_PATH, blob)
|
|
||||||
if (!sent) {
|
|
||||||
logger.warn('sendBeacon returned false — browser may have rejected the request', {
|
|
||||||
chatId: params.chatId,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
return sent
|
|
||||||
} catch (error) {
|
|
||||||
logger.warn('Failed to persist messages via sendBeacon', {
|
|
||||||
chatId: params.chatId,
|
|
||||||
error: error instanceof Error ? error.message : String(error),
|
|
||||||
})
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
import { db } from '@sim/db'
|
import { db } from '@sim/db'
|
||||||
import { workflow } from '@sim/db/schema'
|
import { customTools, workflow } from '@sim/db/schema'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { eq } from 'drizzle-orm'
|
import { and, desc, eq, isNull, or } from 'drizzle-orm'
|
||||||
import { SIM_AGENT_API_URL } from '@/lib/copilot/constants'
|
import { SIM_AGENT_API_URL } from '@/lib/copilot/constants'
|
||||||
import type {
|
import type {
|
||||||
ExecutionContext,
|
ExecutionContext,
|
||||||
@@ -12,6 +12,7 @@ import { routeExecution } from '@/lib/copilot/tools/server/router'
|
|||||||
import { env } from '@/lib/core/config/env'
|
import { env } from '@/lib/core/config/env'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { getEffectiveDecryptedEnv } from '@/lib/environment/utils'
|
import { getEffectiveDecryptedEnv } from '@/lib/environment/utils'
|
||||||
|
import { upsertCustomTools } from '@/lib/workflows/custom-tools/operations'
|
||||||
import { getTool, resolveToolId } from '@/tools/utils'
|
import { getTool, resolveToolId } from '@/tools/utils'
|
||||||
import {
|
import {
|
||||||
executeCheckDeploymentStatus,
|
executeCheckDeploymentStatus,
|
||||||
@@ -76,6 +77,247 @@ import {
|
|||||||
|
|
||||||
const logger = createLogger('CopilotToolExecutor')
|
const logger = createLogger('CopilotToolExecutor')
|
||||||
|
|
||||||
|
type ManageCustomToolOperation = 'add' | 'edit' | 'delete' | 'list'
|
||||||
|
|
||||||
|
interface ManageCustomToolSchema {
|
||||||
|
type: 'function'
|
||||||
|
function: {
|
||||||
|
name: string
|
||||||
|
description?: string
|
||||||
|
parameters: Record<string, unknown>
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
interface ManageCustomToolParams {
|
||||||
|
operation?: string
|
||||||
|
toolId?: string
|
||||||
|
schema?: ManageCustomToolSchema
|
||||||
|
code?: string
|
||||||
|
title?: string
|
||||||
|
workspaceId?: string
|
||||||
|
}
|
||||||
|
|
||||||
|
async function executeManageCustomTool(
|
||||||
|
rawParams: Record<string, unknown>,
|
||||||
|
context: ExecutionContext
|
||||||
|
): Promise<ToolCallResult> {
|
||||||
|
const params = rawParams as ManageCustomToolParams
|
||||||
|
const operation = String(params.operation || '').toLowerCase() as ManageCustomToolOperation
|
||||||
|
const workspaceId = params.workspaceId || context.workspaceId
|
||||||
|
|
||||||
|
if (!operation) {
|
||||||
|
return { success: false, error: "Missing required 'operation' argument" }
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
if (operation === 'list') {
|
||||||
|
const toolsForUser = workspaceId
|
||||||
|
? await db
|
||||||
|
.select()
|
||||||
|
.from(customTools)
|
||||||
|
.where(
|
||||||
|
or(
|
||||||
|
eq(customTools.workspaceId, workspaceId),
|
||||||
|
and(isNull(customTools.workspaceId), eq(customTools.userId, context.userId))
|
||||||
|
)
|
||||||
|
)
|
||||||
|
.orderBy(desc(customTools.createdAt))
|
||||||
|
: await db
|
||||||
|
.select()
|
||||||
|
.from(customTools)
|
||||||
|
.where(and(isNull(customTools.workspaceId), eq(customTools.userId, context.userId)))
|
||||||
|
.orderBy(desc(customTools.createdAt))
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
success: true,
|
||||||
|
operation,
|
||||||
|
tools: toolsForUser,
|
||||||
|
count: toolsForUser.length,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (operation === 'add') {
|
||||||
|
if (!workspaceId) {
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
error: "workspaceId is required for operation 'add'",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (!params.schema || !params.code) {
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
error: "Both 'schema' and 'code' are required for operation 'add'",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const title = params.title || params.schema.function?.name
|
||||||
|
if (!title) {
|
||||||
|
return { success: false, error: "Missing tool title or schema.function.name for 'add'" }
|
||||||
|
}
|
||||||
|
|
||||||
|
const resultTools = await upsertCustomTools({
|
||||||
|
tools: [
|
||||||
|
{
|
||||||
|
title,
|
||||||
|
schema: params.schema,
|
||||||
|
code: params.code,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
workspaceId,
|
||||||
|
userId: context.userId,
|
||||||
|
})
|
||||||
|
const created = resultTools.find((tool) => tool.title === title)
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
success: true,
|
||||||
|
operation,
|
||||||
|
toolId: created?.id,
|
||||||
|
title,
|
||||||
|
message: `Created custom tool "${title}"`,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (operation === 'edit') {
|
||||||
|
if (!workspaceId) {
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
error: "workspaceId is required for operation 'edit'",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (!params.toolId) {
|
||||||
|
return { success: false, error: "'toolId' is required for operation 'edit'" }
|
||||||
|
}
|
||||||
|
if (!params.schema && !params.code) {
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
error: "At least one of 'schema' or 'code' is required for operation 'edit'",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const workspaceTool = await db
|
||||||
|
.select()
|
||||||
|
.from(customTools)
|
||||||
|
.where(and(eq(customTools.id, params.toolId), eq(customTools.workspaceId, workspaceId)))
|
||||||
|
.limit(1)
|
||||||
|
|
||||||
|
const legacyTool =
|
||||||
|
workspaceTool.length === 0
|
||||||
|
? await db
|
||||||
|
.select()
|
||||||
|
.from(customTools)
|
||||||
|
.where(
|
||||||
|
and(
|
||||||
|
eq(customTools.id, params.toolId),
|
||||||
|
isNull(customTools.workspaceId),
|
||||||
|
eq(customTools.userId, context.userId)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
.limit(1)
|
||||||
|
: []
|
||||||
|
|
||||||
|
const existing = workspaceTool[0] || legacyTool[0]
|
||||||
|
if (!existing) {
|
||||||
|
return { success: false, error: `Custom tool not found: ${params.toolId}` }
|
||||||
|
}
|
||||||
|
|
||||||
|
const mergedSchema = params.schema || (existing.schema as ManageCustomToolSchema)
|
||||||
|
const mergedCode = params.code || existing.code
|
||||||
|
const title = params.title || mergedSchema.function?.name || existing.title
|
||||||
|
|
||||||
|
await upsertCustomTools({
|
||||||
|
tools: [
|
||||||
|
{
|
||||||
|
id: params.toolId,
|
||||||
|
title,
|
||||||
|
schema: mergedSchema,
|
||||||
|
code: mergedCode,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
workspaceId,
|
||||||
|
userId: context.userId,
|
||||||
|
})
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
success: true,
|
||||||
|
operation,
|
||||||
|
toolId: params.toolId,
|
||||||
|
title,
|
||||||
|
message: `Updated custom tool "${title}"`,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (operation === 'delete') {
|
||||||
|
if (!params.toolId) {
|
||||||
|
return { success: false, error: "'toolId' is required for operation 'delete'" }
|
||||||
|
}
|
||||||
|
|
||||||
|
const workspaceDelete =
|
||||||
|
workspaceId != null
|
||||||
|
? await db
|
||||||
|
.delete(customTools)
|
||||||
|
.where(
|
||||||
|
and(eq(customTools.id, params.toolId), eq(customTools.workspaceId, workspaceId))
|
||||||
|
)
|
||||||
|
.returning({ id: customTools.id })
|
||||||
|
: []
|
||||||
|
|
||||||
|
const legacyDelete =
|
||||||
|
workspaceDelete.length === 0
|
||||||
|
? await db
|
||||||
|
.delete(customTools)
|
||||||
|
.where(
|
||||||
|
and(
|
||||||
|
eq(customTools.id, params.toolId),
|
||||||
|
isNull(customTools.workspaceId),
|
||||||
|
eq(customTools.userId, context.userId)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
.returning({ id: customTools.id })
|
||||||
|
: []
|
||||||
|
|
||||||
|
const deleted = workspaceDelete[0] || legacyDelete[0]
|
||||||
|
if (!deleted) {
|
||||||
|
return { success: false, error: `Custom tool not found: ${params.toolId}` }
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
success: true,
|
||||||
|
operation,
|
||||||
|
toolId: params.toolId,
|
||||||
|
message: 'Deleted custom tool',
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
error: `Unsupported operation for manage_custom_tool: ${operation}`,
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('manage_custom_tool execution failed', {
|
||||||
|
operation,
|
||||||
|
workspaceId,
|
||||||
|
userId: context.userId,
|
||||||
|
error: error instanceof Error ? error.message : String(error),
|
||||||
|
})
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
error: error instanceof Error ? error.message : 'Failed to manage custom tool',
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
const SERVER_TOOLS = new Set<string>([
|
const SERVER_TOOLS = new Set<string>([
|
||||||
'get_blocks_and_tools',
|
'get_blocks_and_tools',
|
||||||
'get_blocks_metadata',
|
'get_blocks_metadata',
|
||||||
@@ -161,6 +403,19 @@ const SIM_WORKFLOW_TOOL_HANDLERS: Record<
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
oauth_request_access: async (p, _c) => {
|
||||||
|
const providerName = (p.providerName || p.provider_name || 'the provider') as string
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
success: true,
|
||||||
|
status: 'requested',
|
||||||
|
providerName,
|
||||||
|
message: `Requested ${providerName} OAuth connection. The user should complete the OAuth modal in the UI, then retry credential-dependent actions.`,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
manage_custom_tool: (p, c) => executeManageCustomTool(p, c),
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|||||||
@@ -220,6 +220,7 @@ export const env = createEnv({
|
|||||||
SOCKET_SERVER_URL: z.string().url().optional(), // WebSocket server URL for real-time features
|
SOCKET_SERVER_URL: z.string().url().optional(), // WebSocket server URL for real-time features
|
||||||
SOCKET_PORT: z.number().optional(), // Port for WebSocket server
|
SOCKET_PORT: z.number().optional(), // Port for WebSocket server
|
||||||
PORT: z.number().optional(), // Main application port
|
PORT: z.number().optional(), // Main application port
|
||||||
|
INTERNAL_API_BASE_URL: z.string().optional(), // Optional internal base URL for server-side self-calls; must include protocol if set (e.g., http://sim-app.namespace.svc.cluster.local:3000)
|
||||||
ALLOWED_ORIGINS: z.string().optional(), // CORS allowed origins
|
ALLOWED_ORIGINS: z.string().optional(), // CORS allowed origins
|
||||||
|
|
||||||
// OAuth Integration Credentials - All optional, enables third-party integrations
|
// OAuth Integration Credentials - All optional, enables third-party integrations
|
||||||
|
|||||||
@@ -1,6 +1,19 @@
|
|||||||
import { getEnv } from '@/lib/core/config/env'
|
import { getEnv } from '@/lib/core/config/env'
|
||||||
import { isProd } from '@/lib/core/config/feature-flags'
|
import { isProd } from '@/lib/core/config/feature-flags'
|
||||||
|
|
||||||
|
function hasHttpProtocol(url: string): boolean {
|
||||||
|
return /^https?:\/\//i.test(url)
|
||||||
|
}
|
||||||
|
|
||||||
|
function normalizeBaseUrl(url: string): string {
|
||||||
|
if (hasHttpProtocol(url)) {
|
||||||
|
return url
|
||||||
|
}
|
||||||
|
|
||||||
|
const protocol = isProd ? 'https://' : 'http://'
|
||||||
|
return `${protocol}${url}`
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Returns the base URL of the application from NEXT_PUBLIC_APP_URL
|
* Returns the base URL of the application from NEXT_PUBLIC_APP_URL
|
||||||
* This ensures webhooks, callbacks, and other integrations always use the correct public URL
|
* This ensures webhooks, callbacks, and other integrations always use the correct public URL
|
||||||
@@ -8,7 +21,7 @@ import { isProd } from '@/lib/core/config/feature-flags'
|
|||||||
* @throws Error if NEXT_PUBLIC_APP_URL is not configured
|
* @throws Error if NEXT_PUBLIC_APP_URL is not configured
|
||||||
*/
|
*/
|
||||||
export function getBaseUrl(): string {
|
export function getBaseUrl(): string {
|
||||||
const baseUrl = getEnv('NEXT_PUBLIC_APP_URL')
|
const baseUrl = getEnv('NEXT_PUBLIC_APP_URL')?.trim()
|
||||||
|
|
||||||
if (!baseUrl) {
|
if (!baseUrl) {
|
||||||
throw new Error(
|
throw new Error(
|
||||||
@@ -16,12 +29,26 @@ export function getBaseUrl(): string {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (baseUrl.startsWith('http://') || baseUrl.startsWith('https://')) {
|
return normalizeBaseUrl(baseUrl)
|
||||||
return baseUrl
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns the base URL used by server-side internal API calls.
|
||||||
|
* Falls back to NEXT_PUBLIC_APP_URL when INTERNAL_API_BASE_URL is not set.
|
||||||
|
*/
|
||||||
|
export function getInternalApiBaseUrl(): string {
|
||||||
|
const internalBaseUrl = getEnv('INTERNAL_API_BASE_URL')?.trim()
|
||||||
|
if (!internalBaseUrl) {
|
||||||
|
return getBaseUrl()
|
||||||
}
|
}
|
||||||
|
|
||||||
const protocol = isProd ? 'https://' : 'http://'
|
if (!hasHttpProtocol(internalBaseUrl)) {
|
||||||
return `${protocol}${baseUrl}`
|
throw new Error(
|
||||||
|
'INTERNAL_API_BASE_URL must include protocol (http:// or https://), e.g. http://sim-app.default.svc.cluster.local:3000'
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
return internalBaseUrl
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|||||||
246
apps/sim/lib/execution/event-buffer.ts
Normal file
246
apps/sim/lib/execution/event-buffer.ts
Normal file
@@ -0,0 +1,246 @@
|
|||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { getRedisClient } from '@/lib/core/config/redis'
|
||||||
|
import type { ExecutionEvent } from '@/lib/workflows/executor/execution-events'
|
||||||
|
|
||||||
|
const logger = createLogger('ExecutionEventBuffer')
|
||||||
|
|
||||||
|
const REDIS_PREFIX = 'execution:stream:'
|
||||||
|
const TTL_SECONDS = 60 * 60 // 1 hour
|
||||||
|
const EVENT_LIMIT = 1000
|
||||||
|
const RESERVE_BATCH = 100
|
||||||
|
const FLUSH_INTERVAL_MS = 15
|
||||||
|
const FLUSH_MAX_BATCH = 200
|
||||||
|
|
||||||
|
function getEventsKey(executionId: string) {
|
||||||
|
return `${REDIS_PREFIX}${executionId}:events`
|
||||||
|
}
|
||||||
|
|
||||||
|
function getSeqKey(executionId: string) {
|
||||||
|
return `${REDIS_PREFIX}${executionId}:seq`
|
||||||
|
}
|
||||||
|
|
||||||
|
function getMetaKey(executionId: string) {
|
||||||
|
return `${REDIS_PREFIX}${executionId}:meta`
|
||||||
|
}
|
||||||
|
|
||||||
|
export type ExecutionStreamStatus = 'active' | 'complete' | 'error' | 'cancelled'
|
||||||
|
|
||||||
|
export interface ExecutionStreamMeta {
|
||||||
|
status: ExecutionStreamStatus
|
||||||
|
userId?: string
|
||||||
|
workflowId?: string
|
||||||
|
updatedAt?: string
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ExecutionEventEntry {
|
||||||
|
eventId: number
|
||||||
|
executionId: string
|
||||||
|
event: ExecutionEvent
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ExecutionEventWriter {
|
||||||
|
write: (event: ExecutionEvent) => Promise<ExecutionEventEntry>
|
||||||
|
flush: () => Promise<void>
|
||||||
|
close: () => Promise<void>
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function setExecutionMeta(
|
||||||
|
executionId: string,
|
||||||
|
meta: Partial<ExecutionStreamMeta>
|
||||||
|
): Promise<void> {
|
||||||
|
const redis = getRedisClient()
|
||||||
|
if (!redis) {
|
||||||
|
logger.warn('setExecutionMeta: Redis client unavailable', { executionId })
|
||||||
|
return
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
const key = getMetaKey(executionId)
|
||||||
|
const payload: Record<string, string> = {
|
||||||
|
updatedAt: new Date().toISOString(),
|
||||||
|
}
|
||||||
|
if (meta.status) payload.status = meta.status
|
||||||
|
if (meta.userId) payload.userId = meta.userId
|
||||||
|
if (meta.workflowId) payload.workflowId = meta.workflowId
|
||||||
|
await redis.hset(key, payload)
|
||||||
|
await redis.expire(key, TTL_SECONDS)
|
||||||
|
} catch (error) {
|
||||||
|
logger.warn('Failed to update execution meta', {
|
||||||
|
executionId,
|
||||||
|
error: error instanceof Error ? error.message : String(error),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function getExecutionMeta(executionId: string): Promise<ExecutionStreamMeta | null> {
|
||||||
|
const redis = getRedisClient()
|
||||||
|
if (!redis) {
|
||||||
|
logger.warn('getExecutionMeta: Redis client unavailable', { executionId })
|
||||||
|
return null
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
const key = getMetaKey(executionId)
|
||||||
|
const meta = await redis.hgetall(key)
|
||||||
|
if (!meta || Object.keys(meta).length === 0) return null
|
||||||
|
return meta as unknown as ExecutionStreamMeta
|
||||||
|
} catch (error) {
|
||||||
|
logger.warn('Failed to read execution meta', {
|
||||||
|
executionId,
|
||||||
|
error: error instanceof Error ? error.message : String(error),
|
||||||
|
})
|
||||||
|
return null
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function readExecutionEvents(
|
||||||
|
executionId: string,
|
||||||
|
afterEventId: number
|
||||||
|
): Promise<ExecutionEventEntry[]> {
|
||||||
|
const redis = getRedisClient()
|
||||||
|
if (!redis) return []
|
||||||
|
try {
|
||||||
|
const raw = await redis.zrangebyscore(getEventsKey(executionId), afterEventId + 1, '+inf')
|
||||||
|
return raw
|
||||||
|
.map((entry) => {
|
||||||
|
try {
|
||||||
|
return JSON.parse(entry) as ExecutionEventEntry
|
||||||
|
} catch {
|
||||||
|
return null
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.filter((entry): entry is ExecutionEventEntry => Boolean(entry))
|
||||||
|
} catch (error) {
|
||||||
|
logger.warn('Failed to read execution events', {
|
||||||
|
executionId,
|
||||||
|
error: error instanceof Error ? error.message : String(error),
|
||||||
|
})
|
||||||
|
return []
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export function createExecutionEventWriter(executionId: string): ExecutionEventWriter {
|
||||||
|
const redis = getRedisClient()
|
||||||
|
if (!redis) {
|
||||||
|
logger.warn(
|
||||||
|
'createExecutionEventWriter: Redis client unavailable, events will not be buffered',
|
||||||
|
{
|
||||||
|
executionId,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
return {
|
||||||
|
write: async (event) => ({ eventId: 0, executionId, event }),
|
||||||
|
flush: async () => {},
|
||||||
|
close: async () => {},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let pending: ExecutionEventEntry[] = []
|
||||||
|
let nextEventId = 0
|
||||||
|
let maxReservedId = 0
|
||||||
|
let flushTimer: ReturnType<typeof setTimeout> | null = null
|
||||||
|
|
||||||
|
const scheduleFlush = () => {
|
||||||
|
if (flushTimer) return
|
||||||
|
flushTimer = setTimeout(() => {
|
||||||
|
flushTimer = null
|
||||||
|
void flush()
|
||||||
|
}, FLUSH_INTERVAL_MS)
|
||||||
|
}
|
||||||
|
|
||||||
|
const reserveIds = async (minCount: number) => {
|
||||||
|
const reserveCount = Math.max(RESERVE_BATCH, minCount)
|
||||||
|
const newMax = await redis.incrby(getSeqKey(executionId), reserveCount)
|
||||||
|
const startId = newMax - reserveCount + 1
|
||||||
|
if (nextEventId === 0 || nextEventId > maxReservedId) {
|
||||||
|
nextEventId = startId
|
||||||
|
maxReservedId = newMax
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let flushPromise: Promise<void> | null = null
|
||||||
|
let closed = false
|
||||||
|
const inflightWrites = new Set<Promise<ExecutionEventEntry>>()
|
||||||
|
|
||||||
|
const doFlush = async () => {
|
||||||
|
if (pending.length === 0) return
|
||||||
|
const batch = pending
|
||||||
|
pending = []
|
||||||
|
try {
|
||||||
|
const key = getEventsKey(executionId)
|
||||||
|
const zaddArgs: (string | number)[] = []
|
||||||
|
for (const entry of batch) {
|
||||||
|
zaddArgs.push(entry.eventId, JSON.stringify(entry))
|
||||||
|
}
|
||||||
|
const pipeline = redis.pipeline()
|
||||||
|
pipeline.zadd(key, ...zaddArgs)
|
||||||
|
pipeline.expire(key, TTL_SECONDS)
|
||||||
|
pipeline.expire(getSeqKey(executionId), TTL_SECONDS)
|
||||||
|
pipeline.zremrangebyrank(key, 0, -EVENT_LIMIT - 1)
|
||||||
|
await pipeline.exec()
|
||||||
|
} catch (error) {
|
||||||
|
logger.warn('Failed to flush execution events', {
|
||||||
|
executionId,
|
||||||
|
batchSize: batch.length,
|
||||||
|
error: error instanceof Error ? error.message : String(error),
|
||||||
|
stack: error instanceof Error ? error.stack : undefined,
|
||||||
|
})
|
||||||
|
pending = batch.concat(pending)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const flush = async () => {
|
||||||
|
if (flushPromise) {
|
||||||
|
await flushPromise
|
||||||
|
return
|
||||||
|
}
|
||||||
|
flushPromise = doFlush()
|
||||||
|
try {
|
||||||
|
await flushPromise
|
||||||
|
} finally {
|
||||||
|
flushPromise = null
|
||||||
|
if (pending.length > 0) scheduleFlush()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const writeCore = async (event: ExecutionEvent): Promise<ExecutionEventEntry> => {
|
||||||
|
if (closed) return { eventId: 0, executionId, event }
|
||||||
|
if (nextEventId === 0 || nextEventId > maxReservedId) {
|
||||||
|
await reserveIds(1)
|
||||||
|
}
|
||||||
|
const eventId = nextEventId++
|
||||||
|
const entry: ExecutionEventEntry = { eventId, executionId, event }
|
||||||
|
pending.push(entry)
|
||||||
|
if (pending.length >= FLUSH_MAX_BATCH) {
|
||||||
|
await flush()
|
||||||
|
} else {
|
||||||
|
scheduleFlush()
|
||||||
|
}
|
||||||
|
return entry
|
||||||
|
}
|
||||||
|
|
||||||
|
const write = (event: ExecutionEvent): Promise<ExecutionEventEntry> => {
|
||||||
|
const p = writeCore(event)
|
||||||
|
inflightWrites.add(p)
|
||||||
|
const remove = () => inflightWrites.delete(p)
|
||||||
|
p.then(remove, remove)
|
||||||
|
return p
|
||||||
|
}
|
||||||
|
|
||||||
|
const close = async () => {
|
||||||
|
closed = true
|
||||||
|
if (flushTimer) {
|
||||||
|
clearTimeout(flushTimer)
|
||||||
|
flushTimer = null
|
||||||
|
}
|
||||||
|
if (inflightWrites.size > 0) {
|
||||||
|
await Promise.allSettled(inflightWrites)
|
||||||
|
}
|
||||||
|
if (flushPromise) {
|
||||||
|
await flushPromise
|
||||||
|
}
|
||||||
|
if (pending.length > 0) {
|
||||||
|
await doFlush()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return { write, flush, close }
|
||||||
|
}
|
||||||
@@ -2,7 +2,7 @@ import { db } from '@sim/db'
|
|||||||
import { account } from '@sim/db/schema'
|
import { account } from '@sim/db/schema'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { eq } from 'drizzle-orm'
|
import { eq } from 'drizzle-orm'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { refreshTokenIfNeeded } from '@/app/api/auth/oauth/utils'
|
import { refreshTokenIfNeeded } from '@/app/api/auth/oauth/utils'
|
||||||
import { executeProviderRequest } from '@/providers'
|
import { executeProviderRequest } from '@/providers'
|
||||||
import { getProviderFromModel } from '@/providers/utils'
|
import { getProviderFromModel } from '@/providers/utils'
|
||||||
@@ -61,7 +61,7 @@ async function queryKnowledgeBase(
|
|||||||
})
|
})
|
||||||
|
|
||||||
// Call the knowledge base search API directly
|
// Call the knowledge base search API directly
|
||||||
const searchUrl = `${getBaseUrl()}/api/knowledge/search`
|
const searchUrl = `${getInternalApiBaseUrl()}/api/knowledge/search`
|
||||||
|
|
||||||
const response = await fetch(searchUrl, {
|
const response = await fetch(searchUrl, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
|
|||||||
@@ -539,8 +539,8 @@ async function executeMistralOCRRequest(
|
|||||||
const isInternalRoute = url.startsWith('/')
|
const isInternalRoute = url.startsWith('/')
|
||||||
|
|
||||||
if (isInternalRoute) {
|
if (isInternalRoute) {
|
||||||
const { getBaseUrl } = await import('@/lib/core/utils/urls')
|
const { getInternalApiBaseUrl } = await import('@/lib/core/utils/urls')
|
||||||
url = `${getBaseUrl()}${url}`
|
url = `${getInternalApiBaseUrl()}${url}`
|
||||||
}
|
}
|
||||||
|
|
||||||
let headers =
|
let headers =
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { createEnvMock, createMockLogger } from '@sim/testing'
|
import { createEnvMock, loggerMock } from '@sim/testing'
|
||||||
import { beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
import { beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -10,10 +10,6 @@ import { beforeEach, describe, expect, it, type Mock, vi } from 'vitest'
|
|||||||
* mock functions can intercept.
|
* mock functions can intercept.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
const loggerMock = vi.hoisted(() => ({
|
|
||||||
createLogger: () => createMockLogger(),
|
|
||||||
}))
|
|
||||||
|
|
||||||
const mockSend = vi.fn()
|
const mockSend = vi.fn()
|
||||||
const mockBatchSend = vi.fn()
|
const mockBatchSend = vi.fn()
|
||||||
const mockAzureBeginSend = vi.fn()
|
const mockAzureBeginSend = vi.fn()
|
||||||
|
|||||||
@@ -1,20 +1,8 @@
|
|||||||
import { createEnvMock, createMockLogger } from '@sim/testing'
|
import { createEnvMock, databaseMock, loggerMock } from '@sim/testing'
|
||||||
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
||||||
import type { EmailType } from '@/lib/messaging/email/mailer'
|
import type { EmailType } from '@/lib/messaging/email/mailer'
|
||||||
|
|
||||||
const loggerMock = vi.hoisted(() => ({
|
vi.mock('@sim/db', () => databaseMock)
|
||||||
createLogger: () => createMockLogger(),
|
|
||||||
}))
|
|
||||||
|
|
||||||
const mockDb = vi.hoisted(() => ({
|
|
||||||
select: vi.fn(),
|
|
||||||
insert: vi.fn(),
|
|
||||||
update: vi.fn(),
|
|
||||||
}))
|
|
||||||
|
|
||||||
vi.mock('@sim/db', () => ({
|
|
||||||
db: mockDb,
|
|
||||||
}))
|
|
||||||
|
|
||||||
vi.mock('@sim/db/schema', () => ({
|
vi.mock('@sim/db/schema', () => ({
|
||||||
user: { id: 'id', email: 'email' },
|
user: { id: 'id', email: 'email' },
|
||||||
@@ -30,6 +18,8 @@ vi.mock('drizzle-orm', () => ({
|
|||||||
eq: vi.fn((a, b) => ({ type: 'eq', left: a, right: b })),
|
eq: vi.fn((a, b) => ({ type: 'eq', left: a, right: b })),
|
||||||
}))
|
}))
|
||||||
|
|
||||||
|
const mockDb = databaseMock.db as Record<string, ReturnType<typeof vi.fn>>
|
||||||
|
|
||||||
vi.mock('@/lib/core/config/env', () => createEnvMock({ BETTER_AUTH_SECRET: 'test-secret-key' }))
|
vi.mock('@/lib/core/config/env', () => createEnvMock({ BETTER_AUTH_SECRET: 'test-secret-key' }))
|
||||||
|
|
||||||
vi.mock('@sim/logger', () => loggerMock)
|
vi.mock('@sim/logger', () => loggerMock)
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ import { and, eq, isNull, or, sql } from 'drizzle-orm'
|
|||||||
import { nanoid } from 'nanoid'
|
import { nanoid } from 'nanoid'
|
||||||
import { isOrganizationOnTeamOrEnterprisePlan } from '@/lib/billing'
|
import { isOrganizationOnTeamOrEnterprisePlan } from '@/lib/billing'
|
||||||
import { pollingIdempotency } from '@/lib/core/idempotency/service'
|
import { pollingIdempotency } from '@/lib/core/idempotency/service'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { getOAuthToken, refreshAccessTokenIfNeeded } from '@/app/api/auth/oauth/utils'
|
import { getOAuthToken, refreshAccessTokenIfNeeded } from '@/app/api/auth/oauth/utils'
|
||||||
import type { GmailAttachment } from '@/tools/gmail/types'
|
import type { GmailAttachment } from '@/tools/gmail/types'
|
||||||
import { downloadAttachments, extractAttachmentInfo } from '@/tools/gmail/utils'
|
import { downloadAttachments, extractAttachmentInfo } from '@/tools/gmail/utils'
|
||||||
@@ -691,7 +691,7 @@ async function processEmails(
|
|||||||
`[${requestId}] Sending ${config.includeRawEmail ? 'simplified + raw' : 'simplified'} email payload for ${email.id}`
|
`[${requestId}] Sending ${config.includeRawEmail ? 'simplified + raw' : 'simplified'} email payload for ${email.id}`
|
||||||
)
|
)
|
||||||
|
|
||||||
const webhookUrl = `${getBaseUrl()}/api/webhooks/trigger/${webhookData.path}`
|
const webhookUrl = `${getInternalApiBaseUrl()}/api/webhooks/trigger/${webhookData.path}`
|
||||||
|
|
||||||
const response = await fetch(webhookUrl, {
|
const response = await fetch(webhookUrl, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ import type { FetchMessageObject, MailboxLockObject } from 'imapflow'
|
|||||||
import { ImapFlow } from 'imapflow'
|
import { ImapFlow } from 'imapflow'
|
||||||
import { nanoid } from 'nanoid'
|
import { nanoid } from 'nanoid'
|
||||||
import { pollingIdempotency } from '@/lib/core/idempotency/service'
|
import { pollingIdempotency } from '@/lib/core/idempotency/service'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { MAX_CONSECUTIVE_FAILURES } from '@/triggers/constants'
|
import { MAX_CONSECUTIVE_FAILURES } from '@/triggers/constants'
|
||||||
|
|
||||||
const logger = createLogger('ImapPollingService')
|
const logger = createLogger('ImapPollingService')
|
||||||
@@ -639,7 +639,7 @@ async function processEmails(
|
|||||||
timestamp: new Date().toISOString(),
|
timestamp: new Date().toISOString(),
|
||||||
}
|
}
|
||||||
|
|
||||||
const webhookUrl = `${getBaseUrl()}/api/webhooks/trigger/${webhookData.path}`
|
const webhookUrl = `${getInternalApiBaseUrl()}/api/webhooks/trigger/${webhookData.path}`
|
||||||
|
|
||||||
const response = await fetch(webhookUrl, {
|
const response = await fetch(webhookUrl, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ import { htmlToText } from 'html-to-text'
|
|||||||
import { nanoid } from 'nanoid'
|
import { nanoid } from 'nanoid'
|
||||||
import { isOrganizationOnTeamOrEnterprisePlan } from '@/lib/billing'
|
import { isOrganizationOnTeamOrEnterprisePlan } from '@/lib/billing'
|
||||||
import { pollingIdempotency } from '@/lib/core/idempotency'
|
import { pollingIdempotency } from '@/lib/core/idempotency'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { getOAuthToken, refreshAccessTokenIfNeeded } from '@/app/api/auth/oauth/utils'
|
import { getOAuthToken, refreshAccessTokenIfNeeded } from '@/app/api/auth/oauth/utils'
|
||||||
import { MAX_CONSECUTIVE_FAILURES } from '@/triggers/constants'
|
import { MAX_CONSECUTIVE_FAILURES } from '@/triggers/constants'
|
||||||
|
|
||||||
@@ -601,7 +601,7 @@ async function processOutlookEmails(
|
|||||||
`[${requestId}] Processing email: ${email.subject} from ${email.from?.emailAddress?.address}`
|
`[${requestId}] Processing email: ${email.subject} from ${email.from?.emailAddress?.address}`
|
||||||
)
|
)
|
||||||
|
|
||||||
const webhookUrl = `${getBaseUrl()}/api/webhooks/trigger/${webhookData.path}`
|
const webhookUrl = `${getInternalApiBaseUrl()}/api/webhooks/trigger/${webhookData.path}`
|
||||||
|
|
||||||
const response = await fetch(webhookUrl, {
|
const response = await fetch(webhookUrl, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ import {
|
|||||||
secureFetchWithPinnedIP,
|
secureFetchWithPinnedIP,
|
||||||
validateUrlWithDNS,
|
validateUrlWithDNS,
|
||||||
} from '@/lib/core/security/input-validation.server'
|
} from '@/lib/core/security/input-validation.server'
|
||||||
import { getBaseUrl } from '@/lib/core/utils/urls'
|
import { getInternalApiBaseUrl } from '@/lib/core/utils/urls'
|
||||||
import { MAX_CONSECUTIVE_FAILURES } from '@/triggers/constants'
|
import { MAX_CONSECUTIVE_FAILURES } from '@/triggers/constants'
|
||||||
|
|
||||||
const logger = createLogger('RssPollingService')
|
const logger = createLogger('RssPollingService')
|
||||||
@@ -376,7 +376,7 @@ async function processRssItems(
|
|||||||
timestamp: new Date().toISOString(),
|
timestamp: new Date().toISOString(),
|
||||||
}
|
}
|
||||||
|
|
||||||
const webhookUrl = `${getBaseUrl()}/api/webhooks/trigger/${webhookData.path}`
|
const webhookUrl = `${getInternalApiBaseUrl()}/api/webhooks/trigger/${webhookData.path}`
|
||||||
|
|
||||||
const response = await fetch(webhookUrl, {
|
const response = await fetch(webhookUrl, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
|
|||||||
@@ -2364,6 +2364,261 @@ describe('hasWorkflowChanged', () => {
|
|||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
describe('Trigger Config Normalization (False Positive Prevention)', () => {
|
||||||
|
it.concurrent(
|
||||||
|
'should not detect change when deployed has null fields but current has values from triggerConfig',
|
||||||
|
() => {
|
||||||
|
// Core scenario: deployed state has null individual fields, current state has
|
||||||
|
// values populated from triggerConfig at runtime by populateTriggerFieldsFromConfig
|
||||||
|
const deployedState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
botToken: { id: 'botToken', type: 'short-input', value: null },
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const currentState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
||||||
|
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
it.concurrent(
|
||||||
|
'should detect change when user edits a trigger field to a different value',
|
||||||
|
() => {
|
||||||
|
const deployedState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'old-secret' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const currentState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'new-secret' },
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'old-secret' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(hasWorkflowChanged(currentState, deployedState)).toBe(true)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
it.concurrent('should not detect change when both sides have no triggerConfig', () => {
|
||||||
|
const deployedState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const currentState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent(
|
||||||
|
'should not detect change when deployed has empty fields and triggerConfig populates them',
|
||||||
|
() => {
|
||||||
|
// Empty string is also treated as "empty" by normalizeTriggerConfigValues
|
||||||
|
const deployedState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: '' },
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const currentState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
it.concurrent('should not detect change when triggerId differs', () => {
|
||||||
|
const deployedState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
model: { value: 'gpt-4' },
|
||||||
|
triggerId: { value: null },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const currentState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
model: { value: 'gpt-4' },
|
||||||
|
triggerId: { value: 'slack_webhook' },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent(
|
||||||
|
'should not detect change for namespaced system subBlock IDs like samplePayload_slack_webhook',
|
||||||
|
() => {
|
||||||
|
const deployedState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
model: { value: 'gpt-4' },
|
||||||
|
samplePayload_slack_webhook: { value: 'old payload' },
|
||||||
|
triggerInstructions_slack_webhook: { value: 'old instructions' },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const currentState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
model: { value: 'gpt-4' },
|
||||||
|
samplePayload_slack_webhook: { value: 'new payload' },
|
||||||
|
triggerInstructions_slack_webhook: { value: 'new instructions' },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
it.concurrent(
|
||||||
|
'should handle mixed scenario: some fields from triggerConfig, some user-edited',
|
||||||
|
() => {
|
||||||
|
const deployedState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
botToken: { id: 'botToken', type: 'short-input', value: null },
|
||||||
|
includeFiles: { id: 'includeFiles', type: 'switch', value: false },
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const currentState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
||||||
|
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
||||||
|
includeFiles: { id: 'includeFiles', type: 'switch', value: true },
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
// includeFiles changed from false to true — this IS a real change
|
||||||
|
expect(hasWorkflowChanged(currentState, deployedState)).toBe(true)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
})
|
||||||
|
|
||||||
describe('Trigger Runtime Metadata (Should Not Trigger Change)', () => {
|
describe('Trigger Runtime Metadata (Should Not Trigger Change)', () => {
|
||||||
it.concurrent('should not detect change when webhookId differs', () => {
|
it.concurrent('should not detect change when webhookId differs', () => {
|
||||||
const deployedState = createWorkflowState({
|
const deployedState = createWorkflowState({
|
||||||
|
|||||||
@@ -9,6 +9,7 @@ import {
|
|||||||
normalizeLoop,
|
normalizeLoop,
|
||||||
normalizeParallel,
|
normalizeParallel,
|
||||||
normalizeSubBlockValue,
|
normalizeSubBlockValue,
|
||||||
|
normalizeTriggerConfigValues,
|
||||||
normalizeValue,
|
normalizeValue,
|
||||||
normalizeVariables,
|
normalizeVariables,
|
||||||
sanitizeVariable,
|
sanitizeVariable,
|
||||||
@@ -172,14 +173,18 @@ export function generateWorkflowDiffSummary(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Normalize trigger config values for both states before comparison
|
||||||
|
const normalizedCurrentSubs = normalizeTriggerConfigValues(currentSubBlocks)
|
||||||
|
const normalizedPreviousSubs = normalizeTriggerConfigValues(previousSubBlocks)
|
||||||
|
|
||||||
// Compare subBlocks using shared helper for filtering (single source of truth)
|
// Compare subBlocks using shared helper for filtering (single source of truth)
|
||||||
const allSubBlockIds = filterSubBlockIds([
|
const allSubBlockIds = filterSubBlockIds([
|
||||||
...new Set([...Object.keys(currentSubBlocks), ...Object.keys(previousSubBlocks)]),
|
...new Set([...Object.keys(normalizedCurrentSubs), ...Object.keys(normalizedPreviousSubs)]),
|
||||||
])
|
])
|
||||||
|
|
||||||
for (const subId of allSubBlockIds) {
|
for (const subId of allSubBlockIds) {
|
||||||
const currentSub = currentSubBlocks[subId] as Record<string, unknown> | undefined
|
const currentSub = normalizedCurrentSubs[subId] as Record<string, unknown> | undefined
|
||||||
const previousSub = previousSubBlocks[subId] as Record<string, unknown> | undefined
|
const previousSub = normalizedPreviousSubs[subId] as Record<string, unknown> | undefined
|
||||||
|
|
||||||
if (!currentSub || !previousSub) {
|
if (!currentSub || !previousSub) {
|
||||||
changes.push({
|
changes.push({
|
||||||
|
|||||||
@@ -4,10 +4,12 @@
|
|||||||
import { describe, expect, it } from 'vitest'
|
import { describe, expect, it } from 'vitest'
|
||||||
import type { Loop, Parallel } from '@/stores/workflows/workflow/types'
|
import type { Loop, Parallel } from '@/stores/workflows/workflow/types'
|
||||||
import {
|
import {
|
||||||
|
filterSubBlockIds,
|
||||||
normalizedStringify,
|
normalizedStringify,
|
||||||
normalizeEdge,
|
normalizeEdge,
|
||||||
normalizeLoop,
|
normalizeLoop,
|
||||||
normalizeParallel,
|
normalizeParallel,
|
||||||
|
normalizeTriggerConfigValues,
|
||||||
normalizeValue,
|
normalizeValue,
|
||||||
sanitizeInputFormat,
|
sanitizeInputFormat,
|
||||||
sanitizeTools,
|
sanitizeTools,
|
||||||
@@ -584,4 +586,226 @@ describe('Workflow Normalization Utilities', () => {
|
|||||||
expect(result2).toBe(result3)
|
expect(result2).toBe(result3)
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
describe('filterSubBlockIds', () => {
|
||||||
|
it.concurrent('should exclude exact SYSTEM_SUBBLOCK_IDS', () => {
|
||||||
|
const ids = ['signingSecret', 'samplePayload', 'triggerInstructions', 'botToken']
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual(['botToken', 'signingSecret'])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should exclude namespaced SYSTEM_SUBBLOCK_IDS (prefix matching)', () => {
|
||||||
|
const ids = [
|
||||||
|
'signingSecret',
|
||||||
|
'samplePayload_slack_webhook',
|
||||||
|
'triggerInstructions_slack_webhook',
|
||||||
|
'webhookUrlDisplay_slack_webhook',
|
||||||
|
'botToken',
|
||||||
|
]
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual(['botToken', 'signingSecret'])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should exclude exact TRIGGER_RUNTIME_SUBBLOCK_IDS', () => {
|
||||||
|
const ids = ['webhookId', 'triggerPath', 'triggerConfig', 'triggerId', 'signingSecret']
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual(['signingSecret'])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should not exclude IDs that merely contain a system ID substring', () => {
|
||||||
|
const ids = ['mySamplePayload', 'notSamplePayload']
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual(['mySamplePayload', 'notSamplePayload'])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should return sorted results', () => {
|
||||||
|
const ids = ['zebra', 'alpha', 'middle']
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual(['alpha', 'middle', 'zebra'])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should handle empty array', () => {
|
||||||
|
expect(filterSubBlockIds([])).toEqual([])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should handle all IDs being excluded', () => {
|
||||||
|
const ids = ['webhookId', 'triggerPath', 'samplePayload', 'triggerConfig']
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual([])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should exclude setupScript and scheduleInfo namespaced variants', () => {
|
||||||
|
const ids = ['setupScript_google_sheets_row', 'scheduleInfo_cron_trigger', 'realField']
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual(['realField'])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should exclude triggerCredentials namespaced variants', () => {
|
||||||
|
const ids = ['triggerCredentials_slack_webhook', 'signingSecret']
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual(['signingSecret'])
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should exclude synthetic tool-input subBlock IDs', () => {
|
||||||
|
const ids = [
|
||||||
|
'toolConfig',
|
||||||
|
'toolConfig-tool-0-query',
|
||||||
|
'toolConfig-tool-0-url',
|
||||||
|
'toolConfig-tool-1-status',
|
||||||
|
'systemPrompt',
|
||||||
|
]
|
||||||
|
const result = filterSubBlockIds(ids)
|
||||||
|
expect(result).toEqual(['systemPrompt', 'toolConfig'])
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('normalizeTriggerConfigValues', () => {
|
||||||
|
it.concurrent('should return subBlocks unchanged when no triggerConfig exists', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'secret123' },
|
||||||
|
botToken: { id: 'botToken', type: 'short-input', value: 'token456' },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect(result).toEqual(subBlocks)
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should return subBlocks unchanged when triggerConfig value is null', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: { id: 'triggerConfig', type: 'short-input', value: null },
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect(result).toEqual(subBlocks)
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent(
|
||||||
|
'should return subBlocks unchanged when triggerConfig value is not an object',
|
||||||
|
() => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: { id: 'triggerConfig', type: 'short-input', value: 'string-value' },
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect(result).toEqual(subBlocks)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
it.concurrent('should populate null individual fields from triggerConfig', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123', botToken: 'token456' },
|
||||||
|
},
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
botToken: { id: 'botToken', type: 'short-input', value: null },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
||||||
|
expect((result.botToken as Record<string, unknown>).value).toBe('token456')
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should populate undefined individual fields from triggerConfig', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123' },
|
||||||
|
},
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: undefined },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should populate empty string individual fields from triggerConfig', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123' },
|
||||||
|
},
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: '' },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect((result.signingSecret as Record<string, unknown>).value).toBe('secret123')
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should NOT overwrite existing non-empty individual field values', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'old-secret' },
|
||||||
|
},
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: 'user-edited-secret' },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect((result.signingSecret as Record<string, unknown>).value).toBe('user-edited-secret')
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should skip triggerConfig fields that are null/undefined', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: null, botToken: undefined },
|
||||||
|
},
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
botToken: { id: 'botToken', type: 'short-input', value: null },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect((result.signingSecret as Record<string, unknown>).value).toBe(null)
|
||||||
|
expect((result.botToken as Record<string, unknown>).value).toBe(null)
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should skip fields from triggerConfig that have no matching subBlock', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { nonExistentField: 'value123' },
|
||||||
|
},
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
expect(result.nonExistentField).toBeUndefined()
|
||||||
|
expect((result.signingSecret as Record<string, unknown>).value).toBe(null)
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should not mutate the original subBlocks object', () => {
|
||||||
|
const original = {
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123' },
|
||||||
|
},
|
||||||
|
signingSecret: { id: 'signingSecret', type: 'short-input', value: null },
|
||||||
|
}
|
||||||
|
normalizeTriggerConfigValues(original)
|
||||||
|
expect((original.signingSecret as Record<string, unknown>).value).toBe(null)
|
||||||
|
})
|
||||||
|
|
||||||
|
it.concurrent('should preserve other subBlock properties when populating value', () => {
|
||||||
|
const subBlocks = {
|
||||||
|
triggerConfig: {
|
||||||
|
id: 'triggerConfig',
|
||||||
|
type: 'short-input',
|
||||||
|
value: { signingSecret: 'secret123' },
|
||||||
|
},
|
||||||
|
signingSecret: {
|
||||||
|
id: 'signingSecret',
|
||||||
|
type: 'short-input',
|
||||||
|
value: null,
|
||||||
|
placeholder: 'Enter signing secret',
|
||||||
|
},
|
||||||
|
}
|
||||||
|
const result = normalizeTriggerConfigValues(subBlocks)
|
||||||
|
const normalized = result.signingSecret as Record<string, unknown>
|
||||||
|
expect(normalized.value).toBe('secret123')
|
||||||
|
expect(normalized.id).toBe('signingSecret')
|
||||||
|
expect(normalized.type).toBe('short-input')
|
||||||
|
expect(normalized.placeholder).toBe('Enter signing secret')
|
||||||
|
})
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -411,17 +411,63 @@ export function extractBlockFieldsForComparison(block: BlockState): ExtractedBlo
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Filters subBlock IDs to exclude system and trigger runtime subBlocks.
|
* Pattern matching synthetic subBlock IDs created by ToolSubBlockRenderer.
|
||||||
|
* These IDs follow the format `{subBlockId}-tool-{index}-{paramId}` and are
|
||||||
|
* mirrors of values already stored in toolConfig.value.tools[N].params.
|
||||||
|
*/
|
||||||
|
const SYNTHETIC_TOOL_SUBBLOCK_RE = /-tool-\d+-/
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Filters subBlock IDs to exclude system, trigger runtime, and synthetic tool subBlocks.
|
||||||
*
|
*
|
||||||
* @param subBlockIds - Array of subBlock IDs to filter
|
* @param subBlockIds - Array of subBlock IDs to filter
|
||||||
* @returns Filtered and sorted array of subBlock IDs
|
* @returns Filtered and sorted array of subBlock IDs
|
||||||
*/
|
*/
|
||||||
export function filterSubBlockIds(subBlockIds: string[]): string[] {
|
export function filterSubBlockIds(subBlockIds: string[]): string[] {
|
||||||
return subBlockIds
|
return subBlockIds
|
||||||
.filter((id) => !SYSTEM_SUBBLOCK_IDS.includes(id) && !TRIGGER_RUNTIME_SUBBLOCK_IDS.includes(id))
|
.filter((id) => {
|
||||||
|
if (TRIGGER_RUNTIME_SUBBLOCK_IDS.includes(id)) return false
|
||||||
|
if (SYSTEM_SUBBLOCK_IDS.some((sysId) => id === sysId || id.startsWith(`${sysId}_`)))
|
||||||
|
return false
|
||||||
|
if (SYNTHETIC_TOOL_SUBBLOCK_RE.test(id)) return false
|
||||||
|
return true
|
||||||
|
})
|
||||||
.sort()
|
.sort()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Normalizes trigger block subBlocks by populating null/empty individual fields
|
||||||
|
* from the triggerConfig aggregate subBlock. This compensates for the runtime
|
||||||
|
* population done by populateTriggerFieldsFromConfig, ensuring consistent
|
||||||
|
* comparison between client state (with populated values) and deployed state
|
||||||
|
* (with null values from DB).
|
||||||
|
*/
|
||||||
|
export function normalizeTriggerConfigValues(
|
||||||
|
subBlocks: Record<string, unknown>
|
||||||
|
): Record<string, unknown> {
|
||||||
|
const triggerConfigSub = subBlocks.triggerConfig as Record<string, unknown> | undefined
|
||||||
|
const triggerConfigValue = triggerConfigSub?.value
|
||||||
|
if (!triggerConfigValue || typeof triggerConfigValue !== 'object') {
|
||||||
|
return subBlocks
|
||||||
|
}
|
||||||
|
|
||||||
|
const result = { ...subBlocks }
|
||||||
|
for (const [fieldId, configValue] of Object.entries(
|
||||||
|
triggerConfigValue as Record<string, unknown>
|
||||||
|
)) {
|
||||||
|
if (configValue === null || configValue === undefined) continue
|
||||||
|
const existingSub = result[fieldId] as Record<string, unknown> | undefined
|
||||||
|
if (
|
||||||
|
existingSub &&
|
||||||
|
(existingSub.value === null || existingSub.value === undefined || existingSub.value === '')
|
||||||
|
) {
|
||||||
|
result[fieldId] = { ...existingSub, value: configValue }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return result
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Normalizes a subBlock value with sanitization for specific subBlock types.
|
* Normalizes a subBlock value with sanitization for specific subBlock types.
|
||||||
* Sanitizes: tools (removes isExpanded), inputFormat (removes collapsed)
|
* Sanitizes: tools (removes isExpanded), inputFormat (removes collapsed)
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user