mirror of
https://github.com/simstudioai/sim.git
synced 2026-01-15 01:47:59 -05:00
Compare commits
15 Commits
feat/group
...
fix/copilo
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
3ab9b91445 | ||
|
|
a36bdd8729 | ||
|
|
d5bd97de32 | ||
|
|
bd7009e316 | ||
|
|
4f04b1efea | ||
|
|
258e96d6b5 | ||
|
|
4b026ad54d | ||
|
|
f6b7c15dc4 | ||
|
|
70ed19fcdb | ||
|
|
d6e4c91e81 | ||
|
|
e3fa40af11 | ||
|
|
6e0055f847 | ||
|
|
ebbe67aae3 | ||
|
|
2b49d15ec8 | ||
|
|
3d037c9b74 |
@@ -552,6 +552,53 @@ All fields automatically have:
|
|||||||
- `mode: 'trigger'` - Only shown in trigger mode
|
- `mode: 'trigger'` - Only shown in trigger mode
|
||||||
- `condition: { field: 'selectedTriggerId', value: triggerId }` - Only shown when this trigger is selected
|
- `condition: { field: 'selectedTriggerId', value: triggerId }` - Only shown when this trigger is selected
|
||||||
|
|
||||||
|
## Trigger Outputs & Webhook Input Formatting
|
||||||
|
|
||||||
|
### Important: Two Sources of Truth
|
||||||
|
|
||||||
|
There are two related but separate concerns:
|
||||||
|
|
||||||
|
1. **Trigger `outputs`** - Schema/contract defining what fields SHOULD be available. Used by UI for tag dropdown.
|
||||||
|
2. **`formatWebhookInput`** - Implementation that transforms raw webhook payload into actual data. Located in `apps/sim/lib/webhooks/utils.server.ts`.
|
||||||
|
|
||||||
|
**These MUST be aligned.** The fields returned by `formatWebhookInput` should match what's defined in trigger `outputs`. If they differ:
|
||||||
|
- Tag dropdown shows fields that don't exist (broken variable resolution)
|
||||||
|
- Or actual data has fields not shown in dropdown (users can't discover them)
|
||||||
|
|
||||||
|
### When to Add a formatWebhookInput Handler
|
||||||
|
|
||||||
|
- **Simple providers**: If the raw webhook payload structure already matches your outputs, you don't need a handler. The generic fallback returns `body` directly.
|
||||||
|
- **Complex providers**: If you need to transform, flatten, extract nested data, compute fields, or handle conditional logic, add a handler.
|
||||||
|
|
||||||
|
### Adding a Handler
|
||||||
|
|
||||||
|
In `apps/sim/lib/webhooks/utils.server.ts`, add a handler block:
|
||||||
|
|
||||||
|
```typescript
|
||||||
|
if (foundWebhook.provider === '{service}') {
|
||||||
|
// Transform raw webhook body to match trigger outputs
|
||||||
|
return {
|
||||||
|
eventType: body.type,
|
||||||
|
resourceId: body.data?.id || '',
|
||||||
|
timestamp: body.created_at,
|
||||||
|
resource: body.data,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**Key rules:**
|
||||||
|
- Return fields that match your trigger `outputs` definition exactly
|
||||||
|
- No wrapper objects like `webhook: { data: ... }` or `{service}: { ... }`
|
||||||
|
- No duplication (don't spread body AND add individual fields)
|
||||||
|
- Use `null` for missing optional data, not empty objects with empty strings
|
||||||
|
|
||||||
|
### Verify Alignment
|
||||||
|
|
||||||
|
Run the alignment checker:
|
||||||
|
```bash
|
||||||
|
bunx scripts/check-trigger-alignment.ts {service}
|
||||||
|
```
|
||||||
|
|
||||||
## Trigger Outputs
|
## Trigger Outputs
|
||||||
|
|
||||||
Trigger outputs use the same schema as block outputs (NOT tool outputs).
|
Trigger outputs use the same schema as block outputs (NOT tool outputs).
|
||||||
@@ -649,6 +696,11 @@ export const {service}WebhookTrigger: TriggerConfig = {
|
|||||||
- [ ] Added `delete{Service}Webhook` function to `provider-subscriptions.ts`
|
- [ ] Added `delete{Service}Webhook` function to `provider-subscriptions.ts`
|
||||||
- [ ] Added provider to `cleanupExternalWebhook` function
|
- [ ] Added provider to `cleanupExternalWebhook` function
|
||||||
|
|
||||||
|
### Webhook Input Formatting
|
||||||
|
- [ ] Added handler in `apps/sim/lib/webhooks/utils.server.ts` (if custom formatting needed)
|
||||||
|
- [ ] Handler returns fields matching trigger `outputs` exactly
|
||||||
|
- [ ] Run `bunx scripts/check-trigger-alignment.ts {service}` to verify alignment
|
||||||
|
|
||||||
### Testing
|
### Testing
|
||||||
- [ ] Run `bun run type-check` to verify no TypeScript errors
|
- [ ] Run `bun run type-check` to verify no TypeScript errors
|
||||||
- [ ] Restart dev server to pick up new triggers
|
- [ ] Restart dev server to pick up new triggers
|
||||||
|
|||||||
@@ -1855,17 +1855,25 @@ export function LinearIcon(props: React.SVGProps<SVGSVGElement>) {
|
|||||||
|
|
||||||
export function LemlistIcon(props: SVGProps<SVGSVGElement>) {
|
export function LemlistIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
return (
|
return (
|
||||||
<svg
|
<svg {...props} xmlns='http://www.w3.org/2000/svg' viewBox='0 0 180 181' fill='none'>
|
||||||
{...props}
|
<path
|
||||||
xmlns='http://www.w3.org/2000/svg'
|
fillRule='evenodd'
|
||||||
viewBox='0 0 24 24'
|
clipRule='evenodd'
|
||||||
width='24'
|
d='M32.0524 0.919922H147.948C165.65 0.919922 180 15.2703 180 32.9723V148.867C180 166.57 165.65 180.92 147.948 180.92H32.0524C14.3504 180.92 0 166.57 0 148.867V32.9723C0 15.2703 14.3504 0.919922 32.0524 0.919922ZM119.562 82.8879H85.0826C82.4732 82.8879 80.3579 85.0032 80.3579 87.6126V94.2348C80.3579 96.8442 82.4732 98.9595 85.0826 98.9595H119.562C122.171 98.9595 124.286 96.8442 124.286 94.2348V87.6126C124.286 85.0032 122.171 82.8879 119.562 82.8879ZM85.0826 49.1346H127.061C129.67 49.1346 131.785 51.2499 131.785 53.8593V60.4815C131.785 63.0909 129.67 65.2062 127.061 65.2062H85.0826C82.4732 65.2062 80.3579 63.0909 80.3579 60.4815V53.8593C80.3579 51.2499 82.4732 49.1346 85.0826 49.1346ZM131.785 127.981V121.358C131.785 118.75 129.669 116.634 127.061 116.634H76.5706C69.7821 116.634 64.2863 111.138 64.2863 104.349V53.8593C64.2863 51.2513 62.1697 49.1346 59.5616 49.1346H52.9395C50.3314 49.1346 48.2147 51.2513 48.2147 53.8593V114.199C48.8497 124.133 56.7873 132.07 66.7205 132.705H127.061C129.669 132.705 131.785 130.589 131.785 127.981Z'
|
||||||
height='24'
|
fill='#316BFF'
|
||||||
fill='none'
|
/>
|
||||||
>
|
<path
|
||||||
<rect width='24' height='24' rx='4' fill='#316BFF' />
|
d='M85.0826 49.1346H127.061C129.67 49.1346 131.785 51.2499 131.785 53.8593V60.4815C131.785 63.0909 129.67 65.2062 127.061 65.2062H85.0826C82.4732 65.2062 80.3579 63.0909 80.3579 60.4815V53.8593C80.3579 51.2499 82.4732 49.1346 85.0826 49.1346Z'
|
||||||
<path d='M7 6h2v9h5v2H7V6Z' fill='white' />
|
fill='white'
|
||||||
<circle cx='17' cy='8' r='2' fill='white' />
|
/>
|
||||||
|
<path
|
||||||
|
d='M85.0826 82.8879H119.562C122.171 82.8879 124.286 85.0032 124.286 87.6126V94.2348C124.286 96.8442 122.171 98.9595 119.562 98.9595H85.0826C82.4732 98.9595 80.3579 96.8442 80.3579 94.2348V87.6126C80.3579 85.0032 82.4732 82.8879 85.0826 82.8879Z'
|
||||||
|
fill='white'
|
||||||
|
/>
|
||||||
|
<path
|
||||||
|
d='M131.785 121.358V127.981C131.785 130.589 129.669 132.705 127.061 132.705H66.7205C56.7873 132.07 48.8497 124.133 48.2147 114.199V53.8593C48.2147 51.2513 50.3314 49.1346 52.9395 49.1346H59.5616C62.1697 49.1346 64.2863 51.2513 64.2863 53.8593V104.349C64.2863 111.138 69.7821 116.634 76.5706 116.634H127.061C129.669 116.634 131.785 118.75 131.785 121.358Z'
|
||||||
|
fill='white'
|
||||||
|
/>
|
||||||
</svg>
|
</svg>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
@@ -1889,6 +1897,19 @@ export function TelegramIcon(props: SVGProps<SVGSVGElement>) {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function TinybirdIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
|
return (
|
||||||
|
<svg {...props} xmlns='http://www.w3.org/2000/svg' viewBox='0 0 24 24' fill='none'>
|
||||||
|
<rect x='0' y='0' width='24' height='24' fill='#2EF598' rx='6' />
|
||||||
|
<g transform='translate(2, 2) scale(0.833)'>
|
||||||
|
<path d='M25 2.64 17.195.5 14.45 6.635z' fill='#1E7F63' />
|
||||||
|
<path d='M17.535 17.77 10.39 15.215 6.195 25.5z' fill='#1E7F63' />
|
||||||
|
<path d='M0 11.495 17.535 17.77 20.41 4.36z' fill='#1F2437' />
|
||||||
|
</g>
|
||||||
|
</svg>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
export function ClayIcon(props: SVGProps<SVGSVGElement>) {
|
export function ClayIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
return (
|
return (
|
||||||
<svg {...props} xmlns='http://www.w3.org/2000/svg' width='40' height='40' viewBox='0 0 400 400'>
|
<svg {...props} xmlns='http://www.w3.org/2000/svg' width='40' height='40' viewBox='0 0 400 400'>
|
||||||
|
|||||||
@@ -107,6 +107,7 @@ import {
|
|||||||
SupabaseIcon,
|
SupabaseIcon,
|
||||||
TavilyIcon,
|
TavilyIcon,
|
||||||
TelegramIcon,
|
TelegramIcon,
|
||||||
|
TinybirdIcon,
|
||||||
TranslateIcon,
|
TranslateIcon,
|
||||||
TrelloIcon,
|
TrelloIcon,
|
||||||
TTSIcon,
|
TTSIcon,
|
||||||
@@ -230,6 +231,8 @@ export const blockTypeToIconMap: Record<string, IconComponent> = {
|
|||||||
supabase: SupabaseIcon,
|
supabase: SupabaseIcon,
|
||||||
tavily: TavilyIcon,
|
tavily: TavilyIcon,
|
||||||
telegram: TelegramIcon,
|
telegram: TelegramIcon,
|
||||||
|
thinking: BrainIcon,
|
||||||
|
tinybird: TinybirdIcon,
|
||||||
translate: TranslateIcon,
|
translate: TranslateIcon,
|
||||||
trello: TrelloIcon,
|
trello: TrelloIcon,
|
||||||
tts: TTSIcon,
|
tts: TTSIcon,
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ Sim automatically calculates costs for all workflow executions, providing transp
|
|||||||
|
|
||||||
Every workflow execution includes two cost components:
|
Every workflow execution includes two cost components:
|
||||||
|
|
||||||
**Base Execution Charge**: $0.001 per execution
|
**Base Execution Charge**: $0.005 per execution
|
||||||
|
|
||||||
**AI Model Usage**: Variable cost based on token consumption
|
**AI Model Usage**: Variable cost based on token consumption
|
||||||
```javascript
|
```javascript
|
||||||
@@ -48,40 +48,40 @@ The model breakdown shows:
|
|||||||
|
|
||||||
<Tabs items={['Hosted Models', 'Bring Your Own API Key']}>
|
<Tabs items={['Hosted Models', 'Bring Your Own API Key']}>
|
||||||
<Tab>
|
<Tab>
|
||||||
**Hosted Models** - Sim provides API keys with a 1.4x pricing multiplier for Agent blocks:
|
**Hosted Models** - Sim provides API keys with a 1.1x pricing multiplier for Agent blocks:
|
||||||
|
|
||||||
**OpenAI**
|
**OpenAI**
|
||||||
| Model | Base Price (Input/Output) | Hosted Price (Input/Output) |
|
| Model | Base Price (Input/Output) | Hosted Price (Input/Output) |
|
||||||
|-------|---------------------------|----------------------------|
|
|-------|---------------------------|----------------------------|
|
||||||
| GPT-5.1 | $1.25 / $10.00 | $1.75 / $14.00 |
|
| GPT-5.1 | $1.25 / $10.00 | $1.38 / $11.00 |
|
||||||
| GPT-5 | $1.25 / $10.00 | $1.75 / $14.00 |
|
| GPT-5 | $1.25 / $10.00 | $1.38 / $11.00 |
|
||||||
| GPT-5 Mini | $0.25 / $2.00 | $0.35 / $2.80 |
|
| GPT-5 Mini | $0.25 / $2.00 | $0.28 / $2.20 |
|
||||||
| GPT-5 Nano | $0.05 / $0.40 | $0.07 / $0.56 |
|
| GPT-5 Nano | $0.05 / $0.40 | $0.06 / $0.44 |
|
||||||
| GPT-4o | $2.50 / $10.00 | $3.50 / $14.00 |
|
| GPT-4o | $2.50 / $10.00 | $2.75 / $11.00 |
|
||||||
| GPT-4.1 | $2.00 / $8.00 | $2.80 / $11.20 |
|
| GPT-4.1 | $2.00 / $8.00 | $2.20 / $8.80 |
|
||||||
| GPT-4.1 Mini | $0.40 / $1.60 | $0.56 / $2.24 |
|
| GPT-4.1 Mini | $0.40 / $1.60 | $0.44 / $1.76 |
|
||||||
| GPT-4.1 Nano | $0.10 / $0.40 | $0.14 / $0.56 |
|
| GPT-4.1 Nano | $0.10 / $0.40 | $0.11 / $0.44 |
|
||||||
| o1 | $15.00 / $60.00 | $21.00 / $84.00 |
|
| o1 | $15.00 / $60.00 | $16.50 / $66.00 |
|
||||||
| o3 | $2.00 / $8.00 | $2.80 / $11.20 |
|
| o3 | $2.00 / $8.00 | $2.20 / $8.80 |
|
||||||
| o4 Mini | $1.10 / $4.40 | $1.54 / $6.16 |
|
| o4 Mini | $1.10 / $4.40 | $1.21 / $4.84 |
|
||||||
|
|
||||||
**Anthropic**
|
**Anthropic**
|
||||||
| Model | Base Price (Input/Output) | Hosted Price (Input/Output) |
|
| Model | Base Price (Input/Output) | Hosted Price (Input/Output) |
|
||||||
|-------|---------------------------|----------------------------|
|
|-------|---------------------------|----------------------------|
|
||||||
| Claude Opus 4.5 | $5.00 / $25.00 | $7.00 / $35.00 |
|
| Claude Opus 4.5 | $5.00 / $25.00 | $5.50 / $27.50 |
|
||||||
| Claude Opus 4.1 | $15.00 / $75.00 | $21.00 / $105.00 |
|
| Claude Opus 4.1 | $15.00 / $75.00 | $16.50 / $82.50 |
|
||||||
| Claude Sonnet 4.5 | $3.00 / $15.00 | $4.20 / $21.00 |
|
| Claude Sonnet 4.5 | $3.00 / $15.00 | $3.30 / $16.50 |
|
||||||
| Claude Sonnet 4.0 | $3.00 / $15.00 | $4.20 / $21.00 |
|
| Claude Sonnet 4.0 | $3.00 / $15.00 | $3.30 / $16.50 |
|
||||||
| Claude Haiku 4.5 | $1.00 / $5.00 | $1.40 / $7.00 |
|
| Claude Haiku 4.5 | $1.00 / $5.00 | $1.10 / $5.50 |
|
||||||
|
|
||||||
**Google**
|
**Google**
|
||||||
| Model | Base Price (Input/Output) | Hosted Price (Input/Output) |
|
| Model | Base Price (Input/Output) | Hosted Price (Input/Output) |
|
||||||
|-------|---------------------------|----------------------------|
|
|-------|---------------------------|----------------------------|
|
||||||
| Gemini 3 Pro Preview | $2.00 / $12.00 | $2.80 / $16.80 |
|
| Gemini 3 Pro Preview | $2.00 / $12.00 | $2.20 / $13.20 |
|
||||||
| Gemini 2.5 Pro | $1.25 / $10.00 | $1.75 / $14.00 |
|
| Gemini 2.5 Pro | $1.25 / $10.00 | $1.38 / $11.00 |
|
||||||
| Gemini 2.5 Flash | $0.30 / $2.50 | $0.42 / $3.50 |
|
| Gemini 2.5 Flash | $0.30 / $2.50 | $0.33 / $2.75 |
|
||||||
|
|
||||||
*The 1.4x multiplier covers infrastructure and API management costs.*
|
*The 1.1x multiplier covers infrastructure and API management costs.*
|
||||||
</Tab>
|
</Tab>
|
||||||
|
|
||||||
<Tab>
|
<Tab>
|
||||||
|
|||||||
@@ -44,6 +44,8 @@ Send a message to an external A2A-compatible agent.
|
|||||||
| `message` | string | Yes | Message to send to the agent |
|
| `message` | string | Yes | Message to send to the agent |
|
||||||
| `taskId` | string | No | Task ID for continuing an existing task |
|
| `taskId` | string | No | Task ID for continuing an existing task |
|
||||||
| `contextId` | string | No | Context ID for conversation continuity |
|
| `contextId` | string | No | Context ID for conversation continuity |
|
||||||
|
| `data` | string | No | Structured data to include with the message \(JSON string\) |
|
||||||
|
| `files` | array | No | Files to include with the message |
|
||||||
| `apiKey` | string | No | API key for authentication |
|
| `apiKey` | string | No | API key for authentication |
|
||||||
|
|
||||||
#### Output
|
#### Output
|
||||||
@@ -208,8 +210,3 @@ Delete the push notification webhook configuration for a task.
|
|||||||
| `success` | boolean | Whether deletion was successful |
|
| `success` | boolean | Whether deletion was successful |
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
## Notes
|
|
||||||
|
|
||||||
- Category: `tools`
|
|
||||||
- Type: `a2a`
|
|
||||||
|
|||||||
@@ -49,8 +49,7 @@ Retrieves lead information by email address or lead ID.
|
|||||||
| Parameter | Type | Required | Description |
|
| Parameter | Type | Required | Description |
|
||||||
| --------- | ---- | -------- | ----------- |
|
| --------- | ---- | -------- | ----------- |
|
||||||
| `apiKey` | string | Yes | Lemlist API key |
|
| `apiKey` | string | Yes | Lemlist API key |
|
||||||
| `email` | string | No | Lead email address \(use either email or id\) |
|
| `leadIdentifier` | string | Yes | Lead email address or lead ID |
|
||||||
| `id` | string | No | Lead ID \(use either email or id\) |
|
|
||||||
|
|
||||||
#### Output
|
#### Output
|
||||||
|
|
||||||
|
|||||||
@@ -103,6 +103,8 @@
|
|||||||
"supabase",
|
"supabase",
|
||||||
"tavily",
|
"tavily",
|
||||||
"telegram",
|
"telegram",
|
||||||
|
"thinking",
|
||||||
|
"tinybird",
|
||||||
"translate",
|
"translate",
|
||||||
"trello",
|
"trello",
|
||||||
"tts",
|
"tts",
|
||||||
|
|||||||
@@ -124,6 +124,45 @@ Read the latest messages from Slack channels. Retrieve conversation history with
|
|||||||
| --------- | ---- | ----------- |
|
| --------- | ---- | ----------- |
|
||||||
| `messages` | array | Array of message objects from the channel |
|
| `messages` | array | Array of message objects from the channel |
|
||||||
|
|
||||||
|
### `slack_get_message`
|
||||||
|
|
||||||
|
Retrieve a specific message by its timestamp. Useful for getting a thread parent message.
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `authMethod` | string | No | Authentication method: oauth or bot_token |
|
||||||
|
| `botToken` | string | No | Bot token for Custom Bot |
|
||||||
|
| `channel` | string | Yes | Slack channel ID \(e.g., C1234567890\) |
|
||||||
|
| `timestamp` | string | Yes | Message timestamp to retrieve \(e.g., 1405894322.002768\) |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `message` | object | The retrieved message object |
|
||||||
|
|
||||||
|
### `slack_get_thread`
|
||||||
|
|
||||||
|
Retrieve an entire thread including the parent message and all replies. Useful for getting full conversation context.
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `authMethod` | string | No | Authentication method: oauth or bot_token |
|
||||||
|
| `botToken` | string | No | Bot token for Custom Bot |
|
||||||
|
| `channel` | string | Yes | Slack channel ID \(e.g., C1234567890\) |
|
||||||
|
| `threadTs` | string | Yes | Thread timestamp \(thread_ts\) to retrieve \(e.g., 1405894322.002768\) |
|
||||||
|
| `limit` | number | No | Maximum number of messages to return \(default: 100, max: 200\) |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `parentMessage` | object | The thread parent message |
|
||||||
|
|
||||||
### `slack_list_channels`
|
### `slack_list_channels`
|
||||||
|
|
||||||
List all channels in a Slack workspace. Returns public and private channels the bot has access to.
|
List all channels in a Slack workspace. Returns public and private channels the bot has access to.
|
||||||
|
|||||||
70
apps/docs/content/docs/en/tools/tinybird.mdx
Normal file
70
apps/docs/content/docs/en/tools/tinybird.mdx
Normal file
@@ -0,0 +1,70 @@
|
|||||||
|
---
|
||||||
|
title: Tinybird
|
||||||
|
description: Send events and query data with Tinybird
|
||||||
|
---
|
||||||
|
|
||||||
|
import { BlockInfoCard } from "@/components/ui/block-info-card"
|
||||||
|
|
||||||
|
<BlockInfoCard
|
||||||
|
type="tinybird"
|
||||||
|
color="#2EF598"
|
||||||
|
/>
|
||||||
|
|
||||||
|
## Usage Instructions
|
||||||
|
|
||||||
|
Interact with Tinybird using the Events API to stream JSON or NDJSON events, or use the Query API to execute SQL queries against Pipes and Data Sources.
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
## Tools
|
||||||
|
|
||||||
|
### `tinybird_events`
|
||||||
|
|
||||||
|
Send events to a Tinybird Data Source using the Events API. Supports JSON and NDJSON formats with optional gzip compression.
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `base_url` | string | Yes | Tinybird API base URL \(e.g., https://api.tinybird.co or https://api.us-east.tinybird.co\) |
|
||||||
|
| `datasource` | string | Yes | Name of the Tinybird Data Source to send events to |
|
||||||
|
| `data` | string | Yes | Data to send as NDJSON \(newline-delimited JSON\) or JSON string. Each event should be a valid JSON object. |
|
||||||
|
| `wait` | boolean | No | Wait for database acknowledgment before responding. Enables safer retries but introduces latency. Defaults to false. |
|
||||||
|
| `format` | string | No | Format of the events data: "ndjson" \(default\) or "json" |
|
||||||
|
| `compression` | string | No | Compression format: "none" \(default\) or "gzip" |
|
||||||
|
| `token` | string | Yes | Tinybird API Token with DATASOURCE:APPEND or DATASOURCE:CREATE scope |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `successful_rows` | number | Number of rows successfully ingested |
|
||||||
|
| `quarantined_rows` | number | Number of rows quarantined \(failed validation\) |
|
||||||
|
|
||||||
|
### `tinybird_query`
|
||||||
|
|
||||||
|
Execute SQL queries against Tinybird Pipes and Data Sources using the Query API.
|
||||||
|
|
||||||
|
#### Input
|
||||||
|
|
||||||
|
| Parameter | Type | Required | Description |
|
||||||
|
| --------- | ---- | -------- | ----------- |
|
||||||
|
| `base_url` | string | Yes | Tinybird API base URL \(e.g., https://api.tinybird.co\) |
|
||||||
|
| `query` | string | Yes | SQL query to execute. Specify your desired output format \(e.g., FORMAT JSON, FORMAT CSV, FORMAT TSV\). JSON format provides structured data, while other formats return raw text. |
|
||||||
|
| `pipeline` | string | No | Optional pipe name. When provided, enables SELECT * FROM _ syntax |
|
||||||
|
| `token` | string | Yes | Tinybird API Token with PIPE:READ scope |
|
||||||
|
|
||||||
|
#### Output
|
||||||
|
|
||||||
|
| Parameter | Type | Description |
|
||||||
|
| --------- | ---- | ----------- |
|
||||||
|
| `data` | json | Query result data. For FORMAT JSON: array of objects. For other formats \(CSV, TSV, etc.\): raw text string. |
|
||||||
|
| `rows` | number | Number of rows returned \(only available with FORMAT JSON\) |
|
||||||
|
| `statistics` | json | Query execution statistics - elapsed time, rows read, bytes read \(only available with FORMAT JSON\) |
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
## Notes
|
||||||
|
|
||||||
|
- Category: `tools`
|
||||||
|
- Type: `tinybird`
|
||||||
@@ -76,14 +76,6 @@
|
|||||||
pointer-events: none;
|
pointer-events: none;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Suppress the default selection ring for grouped selections
|
|
||||||
* These blocks show a more transparent ring via the component's ring overlay
|
|
||||||
*/
|
|
||||||
.react-flow__node.selected > div[data-grouped-selection="true"] > div::after {
|
|
||||||
box-shadow: none;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Color tokens - single source of truth for all colors
|
* Color tokens - single source of truth for all colors
|
||||||
* Light mode: Warm theme
|
* Light mode: Warm theme
|
||||||
|
|||||||
@@ -1,150 +0,0 @@
|
|||||||
import type {
|
|
||||||
Artifact,
|
|
||||||
Message,
|
|
||||||
Task,
|
|
||||||
TaskArtifactUpdateEvent,
|
|
||||||
TaskState,
|
|
||||||
TaskStatusUpdateEvent,
|
|
||||||
} from '@a2a-js/sdk'
|
|
||||||
import { createLogger } from '@sim/logger'
|
|
||||||
import { type NextRequest, NextResponse } from 'next/server'
|
|
||||||
import { z } from 'zod'
|
|
||||||
import { createA2AClient, extractTextContent, isTerminalState } from '@/lib/a2a/utils'
|
|
||||||
import { checkHybridAuth } from '@/lib/auth/hybrid'
|
|
||||||
import { generateRequestId } from '@/lib/core/utils/request'
|
|
||||||
|
|
||||||
export const dynamic = 'force-dynamic'
|
|
||||||
|
|
||||||
const logger = createLogger('A2ASendMessageStreamAPI')
|
|
||||||
|
|
||||||
const A2ASendMessageStreamSchema = z.object({
|
|
||||||
agentUrl: z.string().min(1, 'Agent URL is required'),
|
|
||||||
message: z.string().min(1, 'Message is required'),
|
|
||||||
taskId: z.string().optional(),
|
|
||||||
contextId: z.string().optional(),
|
|
||||||
apiKey: z.string().optional(),
|
|
||||||
})
|
|
||||||
|
|
||||||
export async function POST(request: NextRequest) {
|
|
||||||
const requestId = generateRequestId()
|
|
||||||
|
|
||||||
try {
|
|
||||||
const authResult = await checkHybridAuth(request, { requireWorkflowId: false })
|
|
||||||
|
|
||||||
if (!authResult.success) {
|
|
||||||
logger.warn(
|
|
||||||
`[${requestId}] Unauthorized A2A send message stream attempt: ${authResult.error}`
|
|
||||||
)
|
|
||||||
return NextResponse.json(
|
|
||||||
{
|
|
||||||
success: false,
|
|
||||||
error: authResult.error || 'Authentication required',
|
|
||||||
},
|
|
||||||
{ status: 401 }
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.info(
|
|
||||||
`[${requestId}] Authenticated A2A send message stream request via ${authResult.authType}`,
|
|
||||||
{
|
|
||||||
userId: authResult.userId,
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
const body = await request.json()
|
|
||||||
const validatedData = A2ASendMessageStreamSchema.parse(body)
|
|
||||||
|
|
||||||
logger.info(`[${requestId}] Sending A2A streaming message`, {
|
|
||||||
agentUrl: validatedData.agentUrl,
|
|
||||||
hasTaskId: !!validatedData.taskId,
|
|
||||||
hasContextId: !!validatedData.contextId,
|
|
||||||
})
|
|
||||||
|
|
||||||
const client = await createA2AClient(validatedData.agentUrl, validatedData.apiKey)
|
|
||||||
|
|
||||||
const message: Message = {
|
|
||||||
kind: 'message',
|
|
||||||
messageId: crypto.randomUUID(),
|
|
||||||
role: 'user',
|
|
||||||
parts: [{ kind: 'text', text: validatedData.message }],
|
|
||||||
...(validatedData.taskId && { taskId: validatedData.taskId }),
|
|
||||||
...(validatedData.contextId && { contextId: validatedData.contextId }),
|
|
||||||
}
|
|
||||||
|
|
||||||
const stream = client.sendMessageStream({ message })
|
|
||||||
|
|
||||||
let taskId = ''
|
|
||||||
let contextId: string | undefined
|
|
||||||
let state: TaskState = 'working'
|
|
||||||
let content = ''
|
|
||||||
let artifacts: Artifact[] = []
|
|
||||||
let history: Message[] = []
|
|
||||||
|
|
||||||
for await (const event of stream) {
|
|
||||||
if (event.kind === 'message') {
|
|
||||||
const msg = event as Message
|
|
||||||
content = extractTextContent(msg)
|
|
||||||
taskId = msg.taskId || taskId
|
|
||||||
contextId = msg.contextId || contextId
|
|
||||||
state = 'completed'
|
|
||||||
} else if (event.kind === 'task') {
|
|
||||||
const task = event as Task
|
|
||||||
taskId = task.id
|
|
||||||
contextId = task.contextId
|
|
||||||
state = task.status.state
|
|
||||||
artifacts = task.artifacts || []
|
|
||||||
history = task.history || []
|
|
||||||
const lastAgentMessage = history.filter((m) => m.role === 'agent').pop()
|
|
||||||
if (lastAgentMessage) {
|
|
||||||
content = extractTextContent(lastAgentMessage)
|
|
||||||
}
|
|
||||||
} else if ('status' in event) {
|
|
||||||
const statusEvent = event as TaskStatusUpdateEvent
|
|
||||||
state = statusEvent.status.state
|
|
||||||
} else if ('artifact' in event) {
|
|
||||||
const artifactEvent = event as TaskArtifactUpdateEvent
|
|
||||||
artifacts.push(artifactEvent.artifact)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.info(`[${requestId}] A2A streaming message completed`, {
|
|
||||||
taskId,
|
|
||||||
state,
|
|
||||||
artifactCount: artifacts.length,
|
|
||||||
})
|
|
||||||
|
|
||||||
return NextResponse.json({
|
|
||||||
success: isTerminalState(state) && state !== 'failed',
|
|
||||||
output: {
|
|
||||||
content,
|
|
||||||
taskId,
|
|
||||||
contextId,
|
|
||||||
state,
|
|
||||||
artifacts,
|
|
||||||
history,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
} catch (error) {
|
|
||||||
if (error instanceof z.ZodError) {
|
|
||||||
logger.warn(`[${requestId}] Invalid request data`, { errors: error.errors })
|
|
||||||
return NextResponse.json(
|
|
||||||
{
|
|
||||||
success: false,
|
|
||||||
error: 'Invalid request data',
|
|
||||||
details: error.errors,
|
|
||||||
},
|
|
||||||
{ status: 400 }
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.error(`[${requestId}] Error in A2A streaming:`, error)
|
|
||||||
|
|
||||||
return NextResponse.json(
|
|
||||||
{
|
|
||||||
success: false,
|
|
||||||
error: error instanceof Error ? error.message : 'Streaming failed',
|
|
||||||
},
|
|
||||||
{ status: 500 }
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
import type { Message, Task } from '@a2a-js/sdk'
|
import type { DataPart, FilePart, Message, Part, Task, TextPart } from '@a2a-js/sdk'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { type NextRequest, NextResponse } from 'next/server'
|
import { type NextRequest, NextResponse } from 'next/server'
|
||||||
import { z } from 'zod'
|
import { z } from 'zod'
|
||||||
@@ -10,11 +10,20 @@ export const dynamic = 'force-dynamic'
|
|||||||
|
|
||||||
const logger = createLogger('A2ASendMessageAPI')
|
const logger = createLogger('A2ASendMessageAPI')
|
||||||
|
|
||||||
|
const FileInputSchema = z.object({
|
||||||
|
type: z.enum(['file', 'url']),
|
||||||
|
data: z.string(),
|
||||||
|
name: z.string(),
|
||||||
|
mime: z.string().optional(),
|
||||||
|
})
|
||||||
|
|
||||||
const A2ASendMessageSchema = z.object({
|
const A2ASendMessageSchema = z.object({
|
||||||
agentUrl: z.string().min(1, 'Agent URL is required'),
|
agentUrl: z.string().min(1, 'Agent URL is required'),
|
||||||
message: z.string().min(1, 'Message is required'),
|
message: z.string().min(1, 'Message is required'),
|
||||||
taskId: z.string().optional(),
|
taskId: z.string().optional(),
|
||||||
contextId: z.string().optional(),
|
contextId: z.string().optional(),
|
||||||
|
data: z.string().optional(),
|
||||||
|
files: z.array(FileInputSchema).optional(),
|
||||||
apiKey: z.string().optional(),
|
apiKey: z.string().optional(),
|
||||||
})
|
})
|
||||||
|
|
||||||
@@ -51,18 +60,100 @@ export async function POST(request: NextRequest) {
|
|||||||
hasContextId: !!validatedData.contextId,
|
hasContextId: !!validatedData.contextId,
|
||||||
})
|
})
|
||||||
|
|
||||||
const client = await createA2AClient(validatedData.agentUrl, validatedData.apiKey)
|
let client
|
||||||
|
try {
|
||||||
|
client = await createA2AClient(validatedData.agentUrl, validatedData.apiKey)
|
||||||
|
logger.info(`[${requestId}] A2A client created successfully`)
|
||||||
|
} catch (clientError) {
|
||||||
|
logger.error(`[${requestId}] Failed to create A2A client:`, clientError)
|
||||||
|
return NextResponse.json(
|
||||||
|
{
|
||||||
|
success: false,
|
||||||
|
error: `Failed to connect to agent: ${clientError instanceof Error ? clientError.message : 'Unknown error'}`,
|
||||||
|
},
|
||||||
|
{ status: 502 }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
const parts: Part[] = []
|
||||||
|
|
||||||
|
const textPart: TextPart = { kind: 'text', text: validatedData.message }
|
||||||
|
parts.push(textPart)
|
||||||
|
|
||||||
|
if (validatedData.data) {
|
||||||
|
try {
|
||||||
|
const parsedData = JSON.parse(validatedData.data)
|
||||||
|
const dataPart: DataPart = { kind: 'data', data: parsedData }
|
||||||
|
parts.push(dataPart)
|
||||||
|
} catch (parseError) {
|
||||||
|
logger.warn(`[${requestId}] Failed to parse data as JSON, skipping DataPart`, {
|
||||||
|
error: parseError instanceof Error ? parseError.message : String(parseError),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (validatedData.files && validatedData.files.length > 0) {
|
||||||
|
for (const file of validatedData.files) {
|
||||||
|
if (file.type === 'url') {
|
||||||
|
const filePart: FilePart = {
|
||||||
|
kind: 'file',
|
||||||
|
file: {
|
||||||
|
name: file.name,
|
||||||
|
mimeType: file.mime,
|
||||||
|
uri: file.data,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
parts.push(filePart)
|
||||||
|
} else if (file.type === 'file') {
|
||||||
|
let bytes = file.data
|
||||||
|
let mimeType = file.mime
|
||||||
|
|
||||||
|
if (file.data.startsWith('data:')) {
|
||||||
|
const match = file.data.match(/^data:([^;]+);base64,(.+)$/)
|
||||||
|
if (match) {
|
||||||
|
mimeType = mimeType || match[1]
|
||||||
|
bytes = match[2]
|
||||||
|
} else {
|
||||||
|
bytes = file.data
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const filePart: FilePart = {
|
||||||
|
kind: 'file',
|
||||||
|
file: {
|
||||||
|
name: file.name,
|
||||||
|
mimeType: mimeType || 'application/octet-stream',
|
||||||
|
bytes,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
parts.push(filePart)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
const message: Message = {
|
const message: Message = {
|
||||||
kind: 'message',
|
kind: 'message',
|
||||||
messageId: crypto.randomUUID(),
|
messageId: crypto.randomUUID(),
|
||||||
role: 'user',
|
role: 'user',
|
||||||
parts: [{ kind: 'text', text: validatedData.message }],
|
parts,
|
||||||
...(validatedData.taskId && { taskId: validatedData.taskId }),
|
...(validatedData.taskId && { taskId: validatedData.taskId }),
|
||||||
...(validatedData.contextId && { contextId: validatedData.contextId }),
|
...(validatedData.contextId && { contextId: validatedData.contextId }),
|
||||||
}
|
}
|
||||||
|
|
||||||
const result = await client.sendMessage({ message })
|
let result
|
||||||
|
try {
|
||||||
|
result = await client.sendMessage({ message })
|
||||||
|
logger.info(`[${requestId}] A2A sendMessage completed`, { resultKind: result?.kind })
|
||||||
|
} catch (sendError) {
|
||||||
|
logger.error(`[${requestId}] Failed to send A2A message:`, sendError)
|
||||||
|
return NextResponse.json(
|
||||||
|
{
|
||||||
|
success: false,
|
||||||
|
error: `Failed to send message: ${sendError instanceof Error ? sendError.message : 'Unknown error'}`,
|
||||||
|
},
|
||||||
|
{ status: 502 }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
if (result.kind === 'message') {
|
if (result.kind === 'message') {
|
||||||
const responseMessage = result as Message
|
const responseMessage = result as Message
|
||||||
|
|||||||
@@ -2,13 +2,6 @@ import { createSession, createWorkspaceRecord, loggerMock } from '@sim/testing'
|
|||||||
import { NextRequest } from 'next/server'
|
import { NextRequest } from 'next/server'
|
||||||
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
import { beforeEach, describe, expect, it, vi } from 'vitest'
|
||||||
|
|
||||||
/**
|
|
||||||
* Tests for workspace invitation by ID API route
|
|
||||||
* Tests GET (details + token acceptance), DELETE (cancellation)
|
|
||||||
*
|
|
||||||
* @vitest-environment node
|
|
||||||
*/
|
|
||||||
|
|
||||||
const mockGetSession = vi.fn()
|
const mockGetSession = vi.fn()
|
||||||
const mockHasWorkspaceAdminAccess = vi.fn()
|
const mockHasWorkspaceAdminAccess = vi.fn()
|
||||||
|
|
||||||
@@ -227,7 +220,7 @@ describe('Workspace Invitation [invitationId] API Route', () => {
|
|||||||
expect(response.headers.get('location')).toBe('https://test.sim.ai/workspace/workspace-456/w')
|
expect(response.headers.get('location')).toBe('https://test.sim.ai/workspace/workspace-456/w')
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should redirect to error page when invitation expired', async () => {
|
it('should redirect to error page with token preserved when invitation expired', async () => {
|
||||||
const session = createSession({
|
const session = createSession({
|
||||||
userId: mockUser.id,
|
userId: mockUser.id,
|
||||||
email: 'invited@example.com',
|
email: 'invited@example.com',
|
||||||
@@ -250,12 +243,13 @@ describe('Workspace Invitation [invitationId] API Route', () => {
|
|||||||
const response = await GET(request, { params })
|
const response = await GET(request, { params })
|
||||||
|
|
||||||
expect(response.status).toBe(307)
|
expect(response.status).toBe(307)
|
||||||
expect(response.headers.get('location')).toBe(
|
const location = response.headers.get('location')
|
||||||
'https://test.sim.ai/invite/invitation-789?error=expired'
|
expect(location).toBe(
|
||||||
|
'https://test.sim.ai/invite/invitation-789?error=expired&token=token-abc123'
|
||||||
)
|
)
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should redirect to error page when email mismatch', async () => {
|
it('should redirect to error page with token preserved when email mismatch', async () => {
|
||||||
const session = createSession({
|
const session = createSession({
|
||||||
userId: mockUser.id,
|
userId: mockUser.id,
|
||||||
email: 'wrong@example.com',
|
email: 'wrong@example.com',
|
||||||
@@ -277,12 +271,13 @@ describe('Workspace Invitation [invitationId] API Route', () => {
|
|||||||
const response = await GET(request, { params })
|
const response = await GET(request, { params })
|
||||||
|
|
||||||
expect(response.status).toBe(307)
|
expect(response.status).toBe(307)
|
||||||
expect(response.headers.get('location')).toBe(
|
const location = response.headers.get('location')
|
||||||
'https://test.sim.ai/invite/invitation-789?error=email-mismatch'
|
expect(location).toBe(
|
||||||
|
'https://test.sim.ai/invite/invitation-789?error=email-mismatch&token=token-abc123'
|
||||||
)
|
)
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should return 404 when invitation not found', async () => {
|
it('should return 404 when invitation not found (without token)', async () => {
|
||||||
const session = createSession({ userId: mockUser.id, email: mockUser.email })
|
const session = createSession({ userId: mockUser.id, email: mockUser.email })
|
||||||
mockGetSession.mockResolvedValue(session)
|
mockGetSession.mockResolvedValue(session)
|
||||||
dbSelectResults = [[]]
|
dbSelectResults = [[]]
|
||||||
@@ -296,6 +291,189 @@ describe('Workspace Invitation [invitationId] API Route', () => {
|
|||||||
expect(response.status).toBe(404)
|
expect(response.status).toBe(404)
|
||||||
expect(data).toEqual({ error: 'Invitation not found or has expired' })
|
expect(data).toEqual({ error: 'Invitation not found or has expired' })
|
||||||
})
|
})
|
||||||
|
|
||||||
|
it('should redirect to error page with token preserved when invitation not found (with token)', async () => {
|
||||||
|
const session = createSession({ userId: mockUser.id, email: mockUser.email })
|
||||||
|
mockGetSession.mockResolvedValue(session)
|
||||||
|
dbSelectResults = [[]]
|
||||||
|
|
||||||
|
const request = new NextRequest(
|
||||||
|
'http://localhost/api/workspaces/invitations/non-existent?token=some-invalid-token'
|
||||||
|
)
|
||||||
|
const params = Promise.resolve({ invitationId: 'non-existent' })
|
||||||
|
|
||||||
|
const response = await GET(request, { params })
|
||||||
|
|
||||||
|
expect(response.status).toBe(307)
|
||||||
|
const location = response.headers.get('location')
|
||||||
|
expect(location).toBe(
|
||||||
|
'https://test.sim.ai/invite/non-existent?error=invalid-token&token=some-invalid-token'
|
||||||
|
)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should redirect to error page with token preserved when invitation already processed', async () => {
|
||||||
|
const session = createSession({
|
||||||
|
userId: mockUser.id,
|
||||||
|
email: 'invited@example.com',
|
||||||
|
name: mockUser.name,
|
||||||
|
})
|
||||||
|
mockGetSession.mockResolvedValue(session)
|
||||||
|
|
||||||
|
const acceptedInvitation = {
|
||||||
|
...mockInvitation,
|
||||||
|
status: 'accepted',
|
||||||
|
}
|
||||||
|
|
||||||
|
dbSelectResults = [[acceptedInvitation], [mockWorkspace]]
|
||||||
|
|
||||||
|
const request = new NextRequest(
|
||||||
|
'http://localhost/api/workspaces/invitations/token-abc123?token=token-abc123'
|
||||||
|
)
|
||||||
|
const params = Promise.resolve({ invitationId: 'token-abc123' })
|
||||||
|
|
||||||
|
const response = await GET(request, { params })
|
||||||
|
|
||||||
|
expect(response.status).toBe(307)
|
||||||
|
const location = response.headers.get('location')
|
||||||
|
expect(location).toBe(
|
||||||
|
'https://test.sim.ai/invite/invitation-789?error=already-processed&token=token-abc123'
|
||||||
|
)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should redirect to error page with token preserved when workspace not found', async () => {
|
||||||
|
const session = createSession({
|
||||||
|
userId: mockUser.id,
|
||||||
|
email: 'invited@example.com',
|
||||||
|
name: mockUser.name,
|
||||||
|
})
|
||||||
|
mockGetSession.mockResolvedValue(session)
|
||||||
|
|
||||||
|
dbSelectResults = [[mockInvitation], []]
|
||||||
|
|
||||||
|
const request = new NextRequest(
|
||||||
|
'http://localhost/api/workspaces/invitations/token-abc123?token=token-abc123'
|
||||||
|
)
|
||||||
|
const params = Promise.resolve({ invitationId: 'token-abc123' })
|
||||||
|
|
||||||
|
const response = await GET(request, { params })
|
||||||
|
|
||||||
|
expect(response.status).toBe(307)
|
||||||
|
const location = response.headers.get('location')
|
||||||
|
expect(location).toBe(
|
||||||
|
'https://test.sim.ai/invite/invitation-789?error=workspace-not-found&token=token-abc123'
|
||||||
|
)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should redirect to error page with token preserved when user not found', async () => {
|
||||||
|
const session = createSession({
|
||||||
|
userId: mockUser.id,
|
||||||
|
email: 'invited@example.com',
|
||||||
|
name: mockUser.name,
|
||||||
|
})
|
||||||
|
mockGetSession.mockResolvedValue(session)
|
||||||
|
|
||||||
|
dbSelectResults = [[mockInvitation], [mockWorkspace], []]
|
||||||
|
|
||||||
|
const request = new NextRequest(
|
||||||
|
'http://localhost/api/workspaces/invitations/token-abc123?token=token-abc123'
|
||||||
|
)
|
||||||
|
const params = Promise.resolve({ invitationId: 'token-abc123' })
|
||||||
|
|
||||||
|
const response = await GET(request, { params })
|
||||||
|
|
||||||
|
expect(response.status).toBe(307)
|
||||||
|
const location = response.headers.get('location')
|
||||||
|
expect(location).toBe(
|
||||||
|
'https://test.sim.ai/invite/invitation-789?error=user-not-found&token=token-abc123'
|
||||||
|
)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should URL encode special characters in token when preserving in error redirects', async () => {
|
||||||
|
const session = createSession({
|
||||||
|
userId: mockUser.id,
|
||||||
|
email: 'wrong@example.com',
|
||||||
|
name: mockUser.name,
|
||||||
|
})
|
||||||
|
mockGetSession.mockResolvedValue(session)
|
||||||
|
|
||||||
|
dbSelectResults = [
|
||||||
|
[mockInvitation],
|
||||||
|
[mockWorkspace],
|
||||||
|
[{ ...mockUser, email: 'wrong@example.com' }],
|
||||||
|
]
|
||||||
|
|
||||||
|
const specialToken = 'token+with/special=chars&more'
|
||||||
|
const request = new NextRequest(
|
||||||
|
`http://localhost/api/workspaces/invitations/token-abc123?token=${encodeURIComponent(specialToken)}`
|
||||||
|
)
|
||||||
|
const params = Promise.resolve({ invitationId: 'token-abc123' })
|
||||||
|
|
||||||
|
const response = await GET(request, { params })
|
||||||
|
|
||||||
|
expect(response.status).toBe(307)
|
||||||
|
const location = response.headers.get('location')
|
||||||
|
expect(location).toContain('error=email-mismatch')
|
||||||
|
expect(location).toContain(`token=${encodeURIComponent(specialToken)}`)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('Token Preservation - Full Flow Scenario', () => {
|
||||||
|
it('should preserve token through email mismatch so user can retry with correct account', async () => {
|
||||||
|
const wrongSession = createSession({
|
||||||
|
userId: 'wrong-user',
|
||||||
|
email: 'wrong@example.com',
|
||||||
|
name: 'Wrong User',
|
||||||
|
})
|
||||||
|
mockGetSession.mockResolvedValue(wrongSession)
|
||||||
|
|
||||||
|
dbSelectResults = [
|
||||||
|
[mockInvitation],
|
||||||
|
[mockWorkspace],
|
||||||
|
[{ id: 'wrong-user', email: 'wrong@example.com' }],
|
||||||
|
]
|
||||||
|
|
||||||
|
const request1 = new NextRequest(
|
||||||
|
'http://localhost/api/workspaces/invitations/token-abc123?token=token-abc123'
|
||||||
|
)
|
||||||
|
const params1 = Promise.resolve({ invitationId: 'token-abc123' })
|
||||||
|
|
||||||
|
const response1 = await GET(request1, { params: params1 })
|
||||||
|
|
||||||
|
expect(response1.status).toBe(307)
|
||||||
|
const location1 = response1.headers.get('location')
|
||||||
|
expect(location1).toBe(
|
||||||
|
'https://test.sim.ai/invite/invitation-789?error=email-mismatch&token=token-abc123'
|
||||||
|
)
|
||||||
|
|
||||||
|
vi.clearAllMocks()
|
||||||
|
dbSelectCallIndex = 0
|
||||||
|
|
||||||
|
const correctSession = createSession({
|
||||||
|
userId: mockUser.id,
|
||||||
|
email: 'invited@example.com',
|
||||||
|
name: mockUser.name,
|
||||||
|
})
|
||||||
|
mockGetSession.mockResolvedValue(correctSession)
|
||||||
|
|
||||||
|
dbSelectResults = [
|
||||||
|
[mockInvitation],
|
||||||
|
[mockWorkspace],
|
||||||
|
[{ ...mockUser, email: 'invited@example.com' }],
|
||||||
|
[],
|
||||||
|
]
|
||||||
|
|
||||||
|
const request2 = new NextRequest(
|
||||||
|
'http://localhost/api/workspaces/invitations/token-abc123?token=token-abc123'
|
||||||
|
)
|
||||||
|
const params2 = Promise.resolve({ invitationId: 'token-abc123' })
|
||||||
|
|
||||||
|
const response2 = await GET(request2, { params: params2 })
|
||||||
|
|
||||||
|
expect(response2.status).toBe(307)
|
||||||
|
expect(response2.headers.get('location')).toBe(
|
||||||
|
'https://test.sim.ai/workspace/workspace-456/w'
|
||||||
|
)
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe('DELETE /api/workspaces/invitations/[invitationId]', () => {
|
describe('DELETE /api/workspaces/invitations/[invitationId]', () => {
|
||||||
|
|||||||
@@ -31,7 +31,6 @@ export async function GET(
|
|||||||
const isAcceptFlow = !!token // If token is provided, this is an acceptance flow
|
const isAcceptFlow = !!token // If token is provided, this is an acceptance flow
|
||||||
|
|
||||||
if (!session?.user?.id) {
|
if (!session?.user?.id) {
|
||||||
// For token-based acceptance flows, redirect to login
|
|
||||||
if (isAcceptFlow) {
|
if (isAcceptFlow) {
|
||||||
return NextResponse.redirect(new URL(`/invite/${invitationId}?token=${token}`, getBaseUrl()))
|
return NextResponse.redirect(new URL(`/invite/${invitationId}?token=${token}`, getBaseUrl()))
|
||||||
}
|
}
|
||||||
@@ -51,8 +50,9 @@ export async function GET(
|
|||||||
|
|
||||||
if (!invitation) {
|
if (!invitation) {
|
||||||
if (isAcceptFlow) {
|
if (isAcceptFlow) {
|
||||||
|
const tokenParam = token ? `&token=${encodeURIComponent(token)}` : ''
|
||||||
return NextResponse.redirect(
|
return NextResponse.redirect(
|
||||||
new URL(`/invite/${invitationId}?error=invalid-token`, getBaseUrl())
|
new URL(`/invite/${invitationId}?error=invalid-token${tokenParam}`, getBaseUrl())
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
return NextResponse.json({ error: 'Invitation not found or has expired' }, { status: 404 })
|
return NextResponse.json({ error: 'Invitation not found or has expired' }, { status: 404 })
|
||||||
@@ -60,8 +60,9 @@ export async function GET(
|
|||||||
|
|
||||||
if (new Date() > new Date(invitation.expiresAt)) {
|
if (new Date() > new Date(invitation.expiresAt)) {
|
||||||
if (isAcceptFlow) {
|
if (isAcceptFlow) {
|
||||||
|
const tokenParam = token ? `&token=${encodeURIComponent(token)}` : ''
|
||||||
return NextResponse.redirect(
|
return NextResponse.redirect(
|
||||||
new URL(`/invite/${invitation.id}?error=expired`, getBaseUrl())
|
new URL(`/invite/${invitation.id}?error=expired${tokenParam}`, getBaseUrl())
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
return NextResponse.json({ error: 'Invitation has expired' }, { status: 400 })
|
return NextResponse.json({ error: 'Invitation has expired' }, { status: 400 })
|
||||||
@@ -75,17 +76,20 @@ export async function GET(
|
|||||||
|
|
||||||
if (!workspaceDetails) {
|
if (!workspaceDetails) {
|
||||||
if (isAcceptFlow) {
|
if (isAcceptFlow) {
|
||||||
|
const tokenParam = token ? `&token=${encodeURIComponent(token)}` : ''
|
||||||
return NextResponse.redirect(
|
return NextResponse.redirect(
|
||||||
new URL(`/invite/${invitation.id}?error=workspace-not-found`, getBaseUrl())
|
new URL(`/invite/${invitation.id}?error=workspace-not-found${tokenParam}`, getBaseUrl())
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
return NextResponse.json({ error: 'Workspace not found' }, { status: 404 })
|
return NextResponse.json({ error: 'Workspace not found' }, { status: 404 })
|
||||||
}
|
}
|
||||||
|
|
||||||
if (isAcceptFlow) {
|
if (isAcceptFlow) {
|
||||||
|
const tokenParam = token ? `&token=${encodeURIComponent(token)}` : ''
|
||||||
|
|
||||||
if (invitation.status !== ('pending' as WorkspaceInvitationStatus)) {
|
if (invitation.status !== ('pending' as WorkspaceInvitationStatus)) {
|
||||||
return NextResponse.redirect(
|
return NextResponse.redirect(
|
||||||
new URL(`/invite/${invitation.id}?error=already-processed`, getBaseUrl())
|
new URL(`/invite/${invitation.id}?error=already-processed${tokenParam}`, getBaseUrl())
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -100,7 +104,7 @@ export async function GET(
|
|||||||
|
|
||||||
if (!userData) {
|
if (!userData) {
|
||||||
return NextResponse.redirect(
|
return NextResponse.redirect(
|
||||||
new URL(`/invite/${invitation.id}?error=user-not-found`, getBaseUrl())
|
new URL(`/invite/${invitation.id}?error=user-not-found${tokenParam}`, getBaseUrl())
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -108,7 +112,7 @@ export async function GET(
|
|||||||
|
|
||||||
if (!isValidMatch) {
|
if (!isValidMatch) {
|
||||||
return NextResponse.redirect(
|
return NextResponse.redirect(
|
||||||
new URL(`/invite/${invitation.id}?error=email-mismatch`, getBaseUrl())
|
new URL(`/invite/${invitation.id}?error=email-mismatch${tokenParam}`, getBaseUrl())
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -178,23 +178,25 @@ export default function Invite() {
|
|||||||
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
const errorReason = searchParams.get('error')
|
const errorReason = searchParams.get('error')
|
||||||
|
const isNew = searchParams.get('new') === 'true'
|
||||||
|
setIsNewUser(isNew)
|
||||||
|
|
||||||
|
const tokenFromQuery = searchParams.get('token')
|
||||||
|
if (tokenFromQuery) {
|
||||||
|
setToken(tokenFromQuery)
|
||||||
|
sessionStorage.setItem('inviteToken', tokenFromQuery)
|
||||||
|
} else {
|
||||||
|
const storedToken = sessionStorage.getItem('inviteToken')
|
||||||
|
if (storedToken && storedToken !== inviteId) {
|
||||||
|
setToken(storedToken)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if (errorReason) {
|
if (errorReason) {
|
||||||
setError(getInviteError(errorReason))
|
setError(getInviteError(errorReason))
|
||||||
setIsLoading(false)
|
setIsLoading(false)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
const isNew = searchParams.get('new') === 'true'
|
|
||||||
setIsNewUser(isNew)
|
|
||||||
|
|
||||||
const tokenFromQuery = searchParams.get('token')
|
|
||||||
const effectiveToken = tokenFromQuery || inviteId
|
|
||||||
|
|
||||||
if (effectiveToken) {
|
|
||||||
setToken(effectiveToken)
|
|
||||||
sessionStorage.setItem('inviteToken', effectiveToken)
|
|
||||||
}
|
|
||||||
}, [searchParams, inviteId])
|
}, [searchParams, inviteId])
|
||||||
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
@@ -203,7 +205,6 @@ export default function Invite() {
|
|||||||
async function fetchInvitationDetails() {
|
async function fetchInvitationDetails() {
|
||||||
setIsLoading(true)
|
setIsLoading(true)
|
||||||
try {
|
try {
|
||||||
// Fetch invitation details using the invitation ID from the URL path
|
|
||||||
const workspaceInviteResponse = await fetch(`/api/workspaces/invitations/${inviteId}`, {
|
const workspaceInviteResponse = await fetch(`/api/workspaces/invitations/${inviteId}`, {
|
||||||
method: 'GET',
|
method: 'GET',
|
||||||
})
|
})
|
||||||
@@ -220,7 +221,6 @@ export default function Invite() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Handle workspace invitation errors with specific status codes
|
|
||||||
if (!workspaceInviteResponse.ok && workspaceInviteResponse.status !== 404) {
|
if (!workspaceInviteResponse.ok && workspaceInviteResponse.status !== 404) {
|
||||||
const errorCode = parseApiError(null, workspaceInviteResponse.status)
|
const errorCode = parseApiError(null, workspaceInviteResponse.status)
|
||||||
const errorData = await workspaceInviteResponse.json().catch(() => ({}))
|
const errorData = await workspaceInviteResponse.json().catch(() => ({}))
|
||||||
@@ -229,7 +229,6 @@ export default function Invite() {
|
|||||||
error: errorData,
|
error: errorData,
|
||||||
})
|
})
|
||||||
|
|
||||||
// Refine error code based on response body if available
|
|
||||||
if (errorData.error) {
|
if (errorData.error) {
|
||||||
const refinedCode = parseApiError(errorData.error, workspaceInviteResponse.status)
|
const refinedCode = parseApiError(errorData.error, workspaceInviteResponse.status)
|
||||||
setError(getInviteError(refinedCode))
|
setError(getInviteError(refinedCode))
|
||||||
@@ -254,13 +253,11 @@ export default function Invite() {
|
|||||||
if (data) {
|
if (data) {
|
||||||
setInvitationType('organization')
|
setInvitationType('organization')
|
||||||
|
|
||||||
// Check if user is already in an organization BEFORE showing the invitation
|
|
||||||
const activeOrgResponse = await client.organization
|
const activeOrgResponse = await client.organization
|
||||||
.getFullOrganization()
|
.getFullOrganization()
|
||||||
.catch(() => ({ data: null }))
|
.catch(() => ({ data: null }))
|
||||||
|
|
||||||
if (activeOrgResponse?.data) {
|
if (activeOrgResponse?.data) {
|
||||||
// User is already in an organization
|
|
||||||
setCurrentOrgName(activeOrgResponse.data.name)
|
setCurrentOrgName(activeOrgResponse.data.name)
|
||||||
setError(getInviteError('already-in-organization'))
|
setError(getInviteError('already-in-organization'))
|
||||||
setIsLoading(false)
|
setIsLoading(false)
|
||||||
@@ -289,7 +286,6 @@ export default function Invite() {
|
|||||||
throw { code: 'invalid-invitation' }
|
throw { code: 'invalid-invitation' }
|
||||||
}
|
}
|
||||||
} catch (orgErr: any) {
|
} catch (orgErr: any) {
|
||||||
// If this is our structured error, use it directly
|
|
||||||
if (orgErr.code) {
|
if (orgErr.code) {
|
||||||
throw orgErr
|
throw orgErr
|
||||||
}
|
}
|
||||||
@@ -316,7 +312,6 @@ export default function Invite() {
|
|||||||
window.location.href = `/api/workspaces/invitations/${encodeURIComponent(inviteId)}?token=${encodeURIComponent(token || '')}`
|
window.location.href = `/api/workspaces/invitations/${encodeURIComponent(inviteId)}?token=${encodeURIComponent(token || '')}`
|
||||||
} else {
|
} else {
|
||||||
try {
|
try {
|
||||||
// Get the organizationId from invitation details
|
|
||||||
const orgId = invitationDetails?.data?.organizationId
|
const orgId = invitationDetails?.data?.organizationId
|
||||||
|
|
||||||
if (!orgId) {
|
if (!orgId) {
|
||||||
@@ -325,7 +320,6 @@ export default function Invite() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Use our custom API endpoint that handles Pro usage snapshot
|
|
||||||
const response = await fetch(`/api/organizations/${orgId}/invitations/${inviteId}`, {
|
const response = await fetch(`/api/organizations/${orgId}/invitations/${inviteId}`, {
|
||||||
method: 'PUT',
|
method: 'PUT',
|
||||||
headers: {
|
headers: {
|
||||||
@@ -347,7 +341,6 @@ export default function Invite() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Set the organization as active
|
|
||||||
await client.organization.setActive({
|
await client.organization.setActive({
|
||||||
organizationId: orgId,
|
organizationId: orgId,
|
||||||
})
|
})
|
||||||
@@ -360,7 +353,6 @@ export default function Invite() {
|
|||||||
} catch (err: any) {
|
} catch (err: any) {
|
||||||
logger.error('Error accepting invitation:', err)
|
logger.error('Error accepting invitation:', err)
|
||||||
|
|
||||||
// Reset accepted state on error
|
|
||||||
setAccepted(false)
|
setAccepted(false)
|
||||||
|
|
||||||
const errorCode = parseApiError(err)
|
const errorCode = parseApiError(err)
|
||||||
@@ -371,7 +363,9 @@ export default function Invite() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const getCallbackUrl = () => {
|
const getCallbackUrl = () => {
|
||||||
return `/invite/${inviteId}${token && token !== inviteId ? `?token=${token}` : ''}`
|
const effectiveToken =
|
||||||
|
token || sessionStorage.getItem('inviteToken') || searchParams.get('token')
|
||||||
|
return `/invite/${inviteId}${effectiveToken && effectiveToken !== inviteId ? `?token=${effectiveToken}` : ''}`
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!session?.user && !isPending) {
|
if (!session?.user && !isPending) {
|
||||||
@@ -435,7 +429,6 @@ export default function Invite() {
|
|||||||
if (error) {
|
if (error) {
|
||||||
const callbackUrl = encodeURIComponent(getCallbackUrl())
|
const callbackUrl = encodeURIComponent(getCallbackUrl())
|
||||||
|
|
||||||
// Special handling for already in organization
|
|
||||||
if (error.code === 'already-in-organization') {
|
if (error.code === 'already-in-organization') {
|
||||||
return (
|
return (
|
||||||
<InviteLayout>
|
<InviteLayout>
|
||||||
@@ -463,7 +456,6 @@ export default function Invite() {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Handle email mismatch - user needs to sign in with a different account
|
|
||||||
if (error.code === 'email-mismatch') {
|
if (error.code === 'email-mismatch') {
|
||||||
return (
|
return (
|
||||||
<InviteLayout>
|
<InviteLayout>
|
||||||
@@ -490,7 +482,6 @@ export default function Invite() {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Handle auth-related errors - prompt user to sign in
|
|
||||||
if (error.requiresAuth) {
|
if (error.requiresAuth) {
|
||||||
return (
|
return (
|
||||||
<InviteLayout>
|
<InviteLayout>
|
||||||
@@ -518,7 +509,6 @@ export default function Invite() {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Handle retryable errors
|
|
||||||
const actions: Array<{
|
const actions: Array<{
|
||||||
label: string
|
label: string
|
||||||
onClick: () => void
|
onClick: () => void
|
||||||
@@ -550,7 +540,6 @@ export default function Invite() {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Show success only if accepted AND no error
|
|
||||||
if (accepted && !error) {
|
if (accepted && !error) {
|
||||||
return (
|
return (
|
||||||
<InviteLayout>
|
<InviteLayout>
|
||||||
|
|||||||
@@ -221,7 +221,9 @@ export function Chat() {
|
|||||||
exportChatCSV,
|
exportChatCSV,
|
||||||
} = useChatStore()
|
} = useChatStore()
|
||||||
|
|
||||||
const { entries } = useTerminalConsoleStore()
|
const hasConsoleHydrated = useTerminalConsoleStore((state) => state._hasHydrated)
|
||||||
|
const entriesFromStore = useTerminalConsoleStore((state) => state.entries)
|
||||||
|
const entries = hasConsoleHydrated ? entriesFromStore : []
|
||||||
const { isExecuting } = useExecutionStore()
|
const { isExecuting } = useExecutionStore()
|
||||||
const { handleRunWorkflow, handleCancelExecution } = useWorkflowExecution()
|
const { handleRunWorkflow, handleCancelExecution } = useWorkflowExecution()
|
||||||
const { data: session } = useSession()
|
const { data: session } = useSession()
|
||||||
@@ -531,35 +533,6 @@ export function Chat() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
if (
|
|
||||||
selectedOutputs.length > 0 &&
|
|
||||||
'logs' in result &&
|
|
||||||
Array.isArray(result.logs) &&
|
|
||||||
activeWorkflowId
|
|
||||||
) {
|
|
||||||
const additionalOutputs: string[] = []
|
|
||||||
|
|
||||||
for (const outputId of selectedOutputs) {
|
|
||||||
const blockId = extractBlockIdFromOutputId(outputId)
|
|
||||||
const path = extractPathFromOutputId(outputId, blockId)
|
|
||||||
|
|
||||||
if (path === 'content') continue
|
|
||||||
|
|
||||||
const outputValue = extractOutputFromLogs(result.logs as BlockLog[], outputId)
|
|
||||||
if (outputValue !== undefined) {
|
|
||||||
const formattedValue =
|
|
||||||
typeof outputValue === 'string' ? outputValue : JSON.stringify(outputValue)
|
|
||||||
if (formattedValue) {
|
|
||||||
additionalOutputs.push(`**${path}:** ${formattedValue}`)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (additionalOutputs.length > 0) {
|
|
||||||
appendMessageContent(responseMessageId, `\n\n${additionalOutputs.join('\n\n')}`)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
finalizeMessageStream(responseMessageId)
|
finalizeMessageStream(responseMessageId)
|
||||||
} else if (contentChunk) {
|
} else if (contentChunk) {
|
||||||
accumulatedContent += contentChunk
|
accumulatedContent += contentChunk
|
||||||
|
|||||||
@@ -29,8 +29,6 @@ export function BlockContextMenu({
|
|||||||
onRemoveFromSubflow,
|
onRemoveFromSubflow,
|
||||||
onOpenEditor,
|
onOpenEditor,
|
||||||
onRename,
|
onRename,
|
||||||
onGroupBlocks,
|
|
||||||
onUngroupBlocks,
|
|
||||||
hasClipboard = false,
|
hasClipboard = false,
|
||||||
showRemoveFromSubflow = false,
|
showRemoveFromSubflow = false,
|
||||||
disableEdit = false,
|
disableEdit = false,
|
||||||
@@ -49,14 +47,6 @@ export function BlockContextMenu({
|
|||||||
|
|
||||||
const canRemoveFromSubflow = showRemoveFromSubflow && !hasStarterBlock
|
const canRemoveFromSubflow = showRemoveFromSubflow && !hasStarterBlock
|
||||||
|
|
||||||
// Check if we can group: need at least 2 blocks selected
|
|
||||||
const canGroup = selectedBlocks.length >= 2
|
|
||||||
|
|
||||||
// Check if we can ungroup: at least one selected block must be in a group
|
|
||||||
// Ungrouping will ungroup all blocks in that group (the entire group, not just selected blocks)
|
|
||||||
const hasGroupedBlock = selectedBlocks.some((b) => !!b.groupId)
|
|
||||||
const canUngroup = hasGroupedBlock
|
|
||||||
|
|
||||||
const getToggleEnabledLabel = () => {
|
const getToggleEnabledLabel = () => {
|
||||||
if (allEnabled) return 'Disable'
|
if (allEnabled) return 'Disable'
|
||||||
if (allDisabled) return 'Enable'
|
if (allDisabled) return 'Enable'
|
||||||
@@ -151,31 +141,6 @@ export function BlockContextMenu({
|
|||||||
</PopoverItem>
|
</PopoverItem>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
{/* Block group actions */}
|
|
||||||
{(canGroup || canUngroup) && <PopoverDivider />}
|
|
||||||
{canGroup && (
|
|
||||||
<PopoverItem
|
|
||||||
disabled={disableEdit}
|
|
||||||
onClick={() => {
|
|
||||||
onGroupBlocks()
|
|
||||||
onClose()
|
|
||||||
}}
|
|
||||||
>
|
|
||||||
Group Blocks
|
|
||||||
</PopoverItem>
|
|
||||||
)}
|
|
||||||
{canUngroup && (
|
|
||||||
<PopoverItem
|
|
||||||
disabled={disableEdit}
|
|
||||||
onClick={() => {
|
|
||||||
onUngroupBlocks()
|
|
||||||
onClose()
|
|
||||||
}}
|
|
||||||
>
|
|
||||||
Ungroup
|
|
||||||
</PopoverItem>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{/* Single block actions */}
|
{/* Single block actions */}
|
||||||
{isSingleBlock && <PopoverDivider />}
|
{isSingleBlock && <PopoverDivider />}
|
||||||
{isSingleBlock && !isSubflow && (
|
{isSingleBlock && !isSubflow && (
|
||||||
|
|||||||
@@ -24,8 +24,6 @@ export interface ContextMenuBlockInfo {
|
|||||||
parentId?: string
|
parentId?: string
|
||||||
/** Parent type ('loop' | 'parallel') if nested */
|
/** Parent type ('loop' | 'parallel') if nested */
|
||||||
parentType?: string
|
parentType?: string
|
||||||
/** Group ID if block is in a group */
|
|
||||||
groupId?: string
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -52,8 +50,6 @@ export interface BlockContextMenuProps {
|
|||||||
onRemoveFromSubflow: () => void
|
onRemoveFromSubflow: () => void
|
||||||
onOpenEditor: () => void
|
onOpenEditor: () => void
|
||||||
onRename: () => void
|
onRename: () => void
|
||||||
onGroupBlocks: () => void
|
|
||||||
onUngroupBlocks: () => void
|
|
||||||
/** Whether clipboard has content for pasting */
|
/** Whether clipboard has content for pasting */
|
||||||
hasClipboard?: boolean
|
hasClipboard?: boolean
|
||||||
/** Whether remove from subflow option should be shown */
|
/** Whether remove from subflow option should be shown */
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
import { memo, useCallback, useMemo } from 'react'
|
import { memo, useCallback, useMemo } from 'react'
|
||||||
import ReactMarkdown from 'react-markdown'
|
import ReactMarkdown from 'react-markdown'
|
||||||
import { type NodeProps, useReactFlow } from 'reactflow'
|
import type { NodeProps } from 'reactflow'
|
||||||
import remarkGfm from 'remark-gfm'
|
import remarkGfm from 'remark-gfm'
|
||||||
import { cn } from '@/lib/core/utils/cn'
|
import { cn } from '@/lib/core/utils/cn'
|
||||||
import { useUserPermissionsContext } from '@/app/workspace/[workspaceId]/providers/workspace-permissions-provider'
|
import { useUserPermissionsContext } from '@/app/workspace/[workspaceId]/providers/workspace-permissions-provider'
|
||||||
@@ -10,7 +10,6 @@ import {
|
|||||||
useBlockDimensions,
|
useBlockDimensions,
|
||||||
} from '@/app/workspace/[workspaceId]/w/[workflowId]/hooks/use-block-dimensions'
|
} from '@/app/workspace/[workspaceId]/w/[workflowId]/hooks/use-block-dimensions'
|
||||||
import { useSubBlockStore } from '@/stores/workflows/subblock/store'
|
import { useSubBlockStore } from '@/stores/workflows/subblock/store'
|
||||||
import { useWorkflowStore } from '@/stores/workflows/workflow/store'
|
|
||||||
import { ActionBar } from '../workflow-block/components'
|
import { ActionBar } from '../workflow-block/components'
|
||||||
import type { WorkflowBlockProps } from '../workflow-block/types'
|
import type { WorkflowBlockProps } from '../workflow-block/types'
|
||||||
|
|
||||||
@@ -199,57 +198,6 @@ export const NoteBlock = memo(function NoteBlock({ id, data }: NodeProps<NoteBlo
|
|||||||
|
|
||||||
const userPermissions = useUserPermissionsContext()
|
const userPermissions = useUserPermissionsContext()
|
||||||
|
|
||||||
// Get React Flow methods for group selection expansion
|
|
||||||
const { getNodes, setNodes } = useReactFlow()
|
|
||||||
const { getGroups } = useWorkflowStore()
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Expands selection to include all group members on mouse down.
|
|
||||||
* This ensures that when a user starts dragging a note in a group,
|
|
||||||
* all other blocks in the group are also selected and will move together.
|
|
||||||
*/
|
|
||||||
const handleGroupMouseDown = useCallback(
|
|
||||||
(e: React.MouseEvent) => {
|
|
||||||
// Only process left mouse button clicks
|
|
||||||
if (e.button !== 0) return
|
|
||||||
|
|
||||||
const groupId = data.groupId
|
|
||||||
if (!groupId) return
|
|
||||||
|
|
||||||
const groups = getGroups()
|
|
||||||
const group = groups[groupId]
|
|
||||||
if (!group || group.blockIds.length <= 1) return
|
|
||||||
|
|
||||||
const groupBlockIds = new Set(group.blockIds)
|
|
||||||
const allNodes = getNodes()
|
|
||||||
|
|
||||||
// Check if all group members are already selected
|
|
||||||
const allSelected = [...groupBlockIds].every((blockId) =>
|
|
||||||
allNodes.find((n) => n.id === blockId && n.selected)
|
|
||||||
)
|
|
||||||
|
|
||||||
if (allSelected) return
|
|
||||||
|
|
||||||
// Expand selection to include all group members
|
|
||||||
setNodes((nodes) =>
|
|
||||||
nodes.map((n) => {
|
|
||||||
const isInGroup = groupBlockIds.has(n.id)
|
|
||||||
const isThisBlock = n.id === id
|
|
||||||
return {
|
|
||||||
...n,
|
|
||||||
selected: isInGroup ? true : n.selected,
|
|
||||||
data: {
|
|
||||||
...n.data,
|
|
||||||
// Mark as grouped selection if in group but not the directly clicked block
|
|
||||||
isGroupedSelection: isInGroup && !isThisBlock && !n.selected,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
})
|
|
||||||
)
|
|
||||||
},
|
|
||||||
[id, data.groupId, getNodes, setNodes, getGroups]
|
|
||||||
)
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Calculate deterministic dimensions based on content structure.
|
* Calculate deterministic dimensions based on content structure.
|
||||||
* Uses fixed width and computed height to avoid ResizeObserver jitter.
|
* Uses fixed width and computed height to avoid ResizeObserver jitter.
|
||||||
@@ -268,14 +216,8 @@ export const NoteBlock = memo(function NoteBlock({ id, data }: NodeProps<NoteBlo
|
|||||||
dependencies: [isEmpty],
|
dependencies: [isEmpty],
|
||||||
})
|
})
|
||||||
|
|
||||||
const isGroupedSelection = data.isGroupedSelection ?? false
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<div
|
<div className='group relative'>
|
||||||
className='group relative'
|
|
||||||
data-grouped-selection={isGroupedSelection ? 'true' : undefined}
|
|
||||||
onMouseDown={handleGroupMouseDown}
|
|
||||||
>
|
|
||||||
<div
|
<div
|
||||||
className={cn(
|
className={cn(
|
||||||
'relative z-[20] w-[250px] cursor-default select-none rounded-[8px] border border-[var(--border)] bg-[var(--surface-2)]'
|
'relative z-[20] w-[250px] cursor-default select-none rounded-[8px] border border-[var(--border)] bg-[var(--surface-2)]'
|
||||||
|
|||||||
@@ -93,10 +93,14 @@ function calculateAdaptiveDelay(displayedLength: number, totalLength: number): n
|
|||||||
*/
|
*/
|
||||||
export const SmoothStreamingText = memo(
|
export const SmoothStreamingText = memo(
|
||||||
({ content, isStreaming }: SmoothStreamingTextProps) => {
|
({ content, isStreaming }: SmoothStreamingTextProps) => {
|
||||||
const [displayedContent, setDisplayedContent] = useState('')
|
// Initialize with full content when not streaming to avoid flash on page load
|
||||||
|
const [displayedContent, setDisplayedContent] = useState(() =>
|
||||||
|
isStreaming ? '' : content
|
||||||
|
)
|
||||||
const contentRef = useRef(content)
|
const contentRef = useRef(content)
|
||||||
const rafRef = useRef<number | null>(null)
|
const rafRef = useRef<number | null>(null)
|
||||||
const indexRef = useRef(0)
|
// Initialize index based on streaming state
|
||||||
|
const indexRef = useRef(isStreaming ? 0 : content.length)
|
||||||
const lastFrameTimeRef = useRef<number>(0)
|
const lastFrameTimeRef = useRef<number>(0)
|
||||||
const isAnimatingRef = useRef(false)
|
const isAnimatingRef = useRef(false)
|
||||||
|
|
||||||
|
|||||||
@@ -46,12 +46,16 @@ interface SmoothThinkingTextProps {
|
|||||||
*/
|
*/
|
||||||
const SmoothThinkingText = memo(
|
const SmoothThinkingText = memo(
|
||||||
({ content, isStreaming }: SmoothThinkingTextProps) => {
|
({ content, isStreaming }: SmoothThinkingTextProps) => {
|
||||||
const [displayedContent, setDisplayedContent] = useState('')
|
// Initialize with full content when not streaming to avoid flash on page load
|
||||||
|
const [displayedContent, setDisplayedContent] = useState(() =>
|
||||||
|
isStreaming ? '' : content
|
||||||
|
)
|
||||||
const [showGradient, setShowGradient] = useState(false)
|
const [showGradient, setShowGradient] = useState(false)
|
||||||
const contentRef = useRef(content)
|
const contentRef = useRef(content)
|
||||||
const textRef = useRef<HTMLDivElement>(null)
|
const textRef = useRef<HTMLDivElement>(null)
|
||||||
const rafRef = useRef<number | null>(null)
|
const rafRef = useRef<number | null>(null)
|
||||||
const indexRef = useRef(0)
|
// Initialize index based on streaming state
|
||||||
|
const indexRef = useRef(isStreaming ? 0 : content.length)
|
||||||
const lastFrameTimeRef = useRef<number>(0)
|
const lastFrameTimeRef = useRef<number>(0)
|
||||||
const isAnimatingRef = useRef(false)
|
const isAnimatingRef = useRef(false)
|
||||||
|
|
||||||
|
|||||||
@@ -26,26 +26,14 @@ function formatTimestamp(iso: string): string {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Common text styling for loading and empty states
|
|
||||||
*/
|
|
||||||
const STATE_TEXT_CLASSES = 'px-[8px] py-[8px] text-[12px] text-[var(--text-muted)]'
|
const STATE_TEXT_CLASSES = 'px-[8px] py-[8px] text-[12px] text-[var(--text-muted)]'
|
||||||
|
|
||||||
/**
|
|
||||||
* Loading state component for mention folders
|
|
||||||
*/
|
|
||||||
const LoadingState = () => <div className={STATE_TEXT_CLASSES}>Loading...</div>
|
const LoadingState = () => <div className={STATE_TEXT_CLASSES}>Loading...</div>
|
||||||
|
|
||||||
/**
|
|
||||||
* Empty state component for mention folders
|
|
||||||
*/
|
|
||||||
const EmptyState = ({ message }: { message: string }) => (
|
const EmptyState = ({ message }: { message: string }) => (
|
||||||
<div className={STATE_TEXT_CLASSES}>{message}</div>
|
<div className={STATE_TEXT_CLASSES}>{message}</div>
|
||||||
)
|
)
|
||||||
|
|
||||||
/**
|
|
||||||
* Aggregated item type for filtered results
|
|
||||||
*/
|
|
||||||
interface AggregatedItem {
|
interface AggregatedItem {
|
||||||
id: string
|
id: string
|
||||||
label: string
|
label: string
|
||||||
@@ -78,14 +66,6 @@ interface MentionMenuProps {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* MentionMenu component for mention menu dropdown.
|
|
||||||
* Handles rendering of mention options, submenus, and aggregated search results.
|
|
||||||
* Manages keyboard navigation and selection of mentions.
|
|
||||||
*
|
|
||||||
* @param props - Component props
|
|
||||||
* @returns Rendered mention menu
|
|
||||||
*/
|
|
||||||
export function MentionMenu({
|
export function MentionMenu({
|
||||||
mentionMenu,
|
mentionMenu,
|
||||||
mentionData,
|
mentionData,
|
||||||
@@ -100,6 +80,7 @@ export function MentionMenu({
|
|||||||
submenuActiveIndex,
|
submenuActiveIndex,
|
||||||
mentionActiveIndex,
|
mentionActiveIndex,
|
||||||
openSubmenuFor,
|
openSubmenuFor,
|
||||||
|
setOpenSubmenuFor,
|
||||||
} = mentionMenu
|
} = mentionMenu
|
||||||
|
|
||||||
const {
|
const {
|
||||||
@@ -308,72 +289,55 @@ export function MentionMenu({
|
|||||||
'Docs', // 7
|
'Docs', // 7
|
||||||
] as const
|
] as const
|
||||||
|
|
||||||
// Get active folder based on navigation when not in submenu and no query
|
|
||||||
const isInFolderNavigationMode = !openSubmenuFor && !showAggregatedView
|
const isInFolderNavigationMode = !openSubmenuFor && !showAggregatedView
|
||||||
|
|
||||||
// Compute caret viewport position via mirror technique for precise anchoring
|
|
||||||
const textareaEl = mentionMenu.textareaRef.current
|
const textareaEl = mentionMenu.textareaRef.current
|
||||||
if (!textareaEl) return null
|
if (!textareaEl) return null
|
||||||
|
|
||||||
const getCaretViewport = (textarea: HTMLTextAreaElement, caretPosition: number, text: string) => {
|
const caretPos = getCaretPos()
|
||||||
const textareaRect = textarea.getBoundingClientRect()
|
const textareaRect = textareaEl.getBoundingClientRect()
|
||||||
const style = window.getComputedStyle(textarea)
|
const style = window.getComputedStyle(textareaEl)
|
||||||
|
|
||||||
const mirrorDiv = document.createElement('div')
|
const mirrorDiv = document.createElement('div')
|
||||||
mirrorDiv.style.position = 'absolute'
|
mirrorDiv.style.position = 'absolute'
|
||||||
mirrorDiv.style.visibility = 'hidden'
|
mirrorDiv.style.visibility = 'hidden'
|
||||||
mirrorDiv.style.whiteSpace = 'pre-wrap'
|
mirrorDiv.style.whiteSpace = 'pre-wrap'
|
||||||
mirrorDiv.style.wordWrap = 'break-word'
|
mirrorDiv.style.wordWrap = 'break-word'
|
||||||
mirrorDiv.style.font = style.font
|
mirrorDiv.style.font = style.font
|
||||||
mirrorDiv.style.padding = style.padding
|
mirrorDiv.style.padding = style.padding
|
||||||
mirrorDiv.style.border = style.border
|
mirrorDiv.style.border = style.border
|
||||||
mirrorDiv.style.width = style.width
|
mirrorDiv.style.width = style.width
|
||||||
mirrorDiv.style.lineHeight = style.lineHeight
|
mirrorDiv.style.lineHeight = style.lineHeight
|
||||||
mirrorDiv.style.boxSizing = style.boxSizing
|
mirrorDiv.style.boxSizing = style.boxSizing
|
||||||
mirrorDiv.style.letterSpacing = style.letterSpacing
|
mirrorDiv.style.letterSpacing = style.letterSpacing
|
||||||
mirrorDiv.style.textTransform = style.textTransform
|
mirrorDiv.style.textTransform = style.textTransform
|
||||||
mirrorDiv.style.textIndent = style.textIndent
|
mirrorDiv.style.textIndent = style.textIndent
|
||||||
mirrorDiv.style.textAlign = style.textAlign
|
mirrorDiv.style.textAlign = style.textAlign
|
||||||
|
mirrorDiv.textContent = message.substring(0, caretPos)
|
||||||
|
|
||||||
mirrorDiv.textContent = text.substring(0, caretPosition)
|
const caretMarker = document.createElement('span')
|
||||||
|
caretMarker.style.display = 'inline-block'
|
||||||
|
caretMarker.style.width = '0px'
|
||||||
|
caretMarker.style.padding = '0'
|
||||||
|
caretMarker.style.border = '0'
|
||||||
|
mirrorDiv.appendChild(caretMarker)
|
||||||
|
|
||||||
const caretMarker = document.createElement('span')
|
document.body.appendChild(mirrorDiv)
|
||||||
caretMarker.style.display = 'inline-block'
|
const markerRect = caretMarker.getBoundingClientRect()
|
||||||
caretMarker.style.width = '0px'
|
const mirrorRect = mirrorDiv.getBoundingClientRect()
|
||||||
caretMarker.style.padding = '0'
|
document.body.removeChild(mirrorDiv)
|
||||||
caretMarker.style.border = '0'
|
|
||||||
mirrorDiv.appendChild(caretMarker)
|
|
||||||
|
|
||||||
document.body.appendChild(mirrorDiv)
|
const caretViewport = {
|
||||||
const markerRect = caretMarker.getBoundingClientRect()
|
left: textareaRect.left + (markerRect.left - mirrorRect.left) - textareaEl.scrollLeft,
|
||||||
const mirrorRect = mirrorDiv.getBoundingClientRect()
|
top: textareaRect.top + (markerRect.top - mirrorRect.top) - textareaEl.scrollTop,
|
||||||
document.body.removeChild(mirrorDiv)
|
|
||||||
|
|
||||||
const leftOffset = markerRect.left - mirrorRect.left - textarea.scrollLeft
|
|
||||||
const topOffset = markerRect.top - mirrorRect.top - textarea.scrollTop
|
|
||||||
|
|
||||||
return {
|
|
||||||
left: textareaRect.left + leftOffset,
|
|
||||||
top: textareaRect.top + topOffset,
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const caretPos = getCaretPos()
|
|
||||||
const caretViewport = getCaretViewport(textareaEl, caretPos, message)
|
|
||||||
|
|
||||||
// Decide preferred side based on available space
|
|
||||||
const margin = 8
|
const margin = 8
|
||||||
const spaceAbove = caretViewport.top - margin
|
|
||||||
const spaceBelow = window.innerHeight - caretViewport.top - margin
|
const spaceBelow = window.innerHeight - caretViewport.top - margin
|
||||||
const side: 'top' | 'bottom' = spaceBelow >= spaceAbove ? 'bottom' : 'top'
|
const side: 'top' | 'bottom' = spaceBelow >= caretViewport.top - margin ? 'bottom' : 'top'
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Popover
|
<Popover open={open} onOpenChange={() => {}}>
|
||||||
open={open}
|
|
||||||
onOpenChange={() => {
|
|
||||||
/* controlled by mentionMenu */
|
|
||||||
}}
|
|
||||||
>
|
|
||||||
<PopoverAnchor asChild>
|
<PopoverAnchor asChild>
|
||||||
<div
|
<div
|
||||||
style={{
|
style={{
|
||||||
@@ -399,7 +363,7 @@ export function MentionMenu({
|
|||||||
onOpenAutoFocus={(e) => e.preventDefault()}
|
onOpenAutoFocus={(e) => e.preventDefault()}
|
||||||
onCloseAutoFocus={(e) => e.preventDefault()}
|
onCloseAutoFocus={(e) => e.preventDefault()}
|
||||||
>
|
>
|
||||||
<PopoverBackButton />
|
<PopoverBackButton onClick={() => setOpenSubmenuFor(null)} />
|
||||||
<PopoverScrollArea ref={menuListRef} className='space-y-[2px]'>
|
<PopoverScrollArea ref={menuListRef} className='space-y-[2px]'>
|
||||||
{openSubmenuFor ? (
|
{openSubmenuFor ? (
|
||||||
// Submenu view - showing contents of a specific folder
|
// Submenu view - showing contents of a specific folder
|
||||||
|
|||||||
@@ -12,31 +12,19 @@ import {
|
|||||||
} from '@/components/emcn'
|
} from '@/components/emcn'
|
||||||
import type { useMentionMenu } from '../../hooks/use-mention-menu'
|
import type { useMentionMenu } from '../../hooks/use-mention-menu'
|
||||||
|
|
||||||
/**
|
|
||||||
* Top-level slash command options
|
|
||||||
*/
|
|
||||||
const TOP_LEVEL_COMMANDS = [
|
const TOP_LEVEL_COMMANDS = [
|
||||||
{ id: 'fast', label: 'fast' },
|
{ id: 'fast', label: 'Fast' },
|
||||||
{ id: 'plan', label: 'plan' },
|
{ id: 'research', label: 'Research' },
|
||||||
{ id: 'debug', label: 'debug' },
|
{ id: 'superagent', label: 'Actions' },
|
||||||
{ id: 'research', label: 'research' },
|
|
||||||
{ id: 'deploy', label: 'deploy' },
|
|
||||||
{ id: 'superagent', label: 'superagent' },
|
|
||||||
] as const
|
] as const
|
||||||
|
|
||||||
/**
|
|
||||||
* Web submenu commands
|
|
||||||
*/
|
|
||||||
const WEB_COMMANDS = [
|
const WEB_COMMANDS = [
|
||||||
{ id: 'search', label: 'search' },
|
{ id: 'search', label: 'Search' },
|
||||||
{ id: 'read', label: 'read' },
|
{ id: 'read', label: 'Read' },
|
||||||
{ id: 'scrape', label: 'scrape' },
|
{ id: 'scrape', label: 'Scrape' },
|
||||||
{ id: 'crawl', label: 'crawl' },
|
{ id: 'crawl', label: 'Crawl' },
|
||||||
] as const
|
] as const
|
||||||
|
|
||||||
/**
|
|
||||||
* All command labels for filtering
|
|
||||||
*/
|
|
||||||
const ALL_COMMANDS = [...TOP_LEVEL_COMMANDS, ...WEB_COMMANDS]
|
const ALL_COMMANDS = [...TOP_LEVEL_COMMANDS, ...WEB_COMMANDS]
|
||||||
|
|
||||||
interface SlashMenuProps {
|
interface SlashMenuProps {
|
||||||
@@ -45,13 +33,6 @@ interface SlashMenuProps {
|
|||||||
onSelectCommand: (command: string) => void
|
onSelectCommand: (command: string) => void
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* SlashMenu component for slash command dropdown.
|
|
||||||
* Shows command options when user types '/'.
|
|
||||||
*
|
|
||||||
* @param props - Component props
|
|
||||||
* @returns Rendered slash menu
|
|
||||||
*/
|
|
||||||
export function SlashMenu({ mentionMenu, message, onSelectCommand }: SlashMenuProps) {
|
export function SlashMenu({ mentionMenu, message, onSelectCommand }: SlashMenuProps) {
|
||||||
const {
|
const {
|
||||||
mentionMenuRef,
|
mentionMenuRef,
|
||||||
@@ -64,92 +45,71 @@ export function SlashMenu({ mentionMenu, message, onSelectCommand }: SlashMenuPr
|
|||||||
setOpenSubmenuFor,
|
setOpenSubmenuFor,
|
||||||
} = mentionMenu
|
} = mentionMenu
|
||||||
|
|
||||||
/**
|
|
||||||
* Get the current query string after /
|
|
||||||
*/
|
|
||||||
const currentQuery = useMemo(() => {
|
const currentQuery = useMemo(() => {
|
||||||
const caretPos = getCaretPos()
|
const caretPos = getCaretPos()
|
||||||
const active = getActiveSlashQueryAtPosition(caretPos, message)
|
const active = getActiveSlashQueryAtPosition(caretPos, message)
|
||||||
return active?.query.trim().toLowerCase() || ''
|
return active?.query.trim().toLowerCase() || ''
|
||||||
}, [message, getCaretPos, getActiveSlashQueryAtPosition])
|
}, [message, getCaretPos, getActiveSlashQueryAtPosition])
|
||||||
|
|
||||||
/**
|
|
||||||
* Filter commands based on query (search across all commands when there's a query)
|
|
||||||
*/
|
|
||||||
const filteredCommands = useMemo(() => {
|
const filteredCommands = useMemo(() => {
|
||||||
if (!currentQuery) return null // Show folder view when no query
|
if (!currentQuery) return null
|
||||||
return ALL_COMMANDS.filter((cmd) => cmd.label.toLowerCase().includes(currentQuery))
|
return ALL_COMMANDS.filter(
|
||||||
|
(cmd) =>
|
||||||
|
cmd.id.toLowerCase().includes(currentQuery) ||
|
||||||
|
cmd.label.toLowerCase().includes(currentQuery)
|
||||||
|
)
|
||||||
}, [currentQuery])
|
}, [currentQuery])
|
||||||
|
|
||||||
// Show aggregated view when there's a query
|
|
||||||
const showAggregatedView = currentQuery.length > 0
|
const showAggregatedView = currentQuery.length > 0
|
||||||
|
const isInFolderNavigationMode = !openSubmenuFor && !showAggregatedView
|
||||||
|
|
||||||
// Compute caret viewport position via mirror technique for precise anchoring
|
|
||||||
const textareaEl = mentionMenu.textareaRef.current
|
const textareaEl = mentionMenu.textareaRef.current
|
||||||
if (!textareaEl) return null
|
if (!textareaEl) return null
|
||||||
|
|
||||||
const getCaretViewport = (textarea: HTMLTextAreaElement, caretPosition: number, text: string) => {
|
const caretPos = getCaretPos()
|
||||||
const textareaRect = textarea.getBoundingClientRect()
|
const textareaRect = textareaEl.getBoundingClientRect()
|
||||||
const style = window.getComputedStyle(textarea)
|
const style = window.getComputedStyle(textareaEl)
|
||||||
|
|
||||||
const mirrorDiv = document.createElement('div')
|
const mirrorDiv = document.createElement('div')
|
||||||
mirrorDiv.style.position = 'absolute'
|
mirrorDiv.style.position = 'absolute'
|
||||||
mirrorDiv.style.visibility = 'hidden'
|
mirrorDiv.style.visibility = 'hidden'
|
||||||
mirrorDiv.style.whiteSpace = 'pre-wrap'
|
mirrorDiv.style.whiteSpace = 'pre-wrap'
|
||||||
mirrorDiv.style.wordWrap = 'break-word'
|
mirrorDiv.style.wordWrap = 'break-word'
|
||||||
mirrorDiv.style.font = style.font
|
mirrorDiv.style.font = style.font
|
||||||
mirrorDiv.style.padding = style.padding
|
mirrorDiv.style.padding = style.padding
|
||||||
mirrorDiv.style.border = style.border
|
mirrorDiv.style.border = style.border
|
||||||
mirrorDiv.style.width = style.width
|
mirrorDiv.style.width = style.width
|
||||||
mirrorDiv.style.lineHeight = style.lineHeight
|
mirrorDiv.style.lineHeight = style.lineHeight
|
||||||
mirrorDiv.style.boxSizing = style.boxSizing
|
mirrorDiv.style.boxSizing = style.boxSizing
|
||||||
mirrorDiv.style.letterSpacing = style.letterSpacing
|
mirrorDiv.style.letterSpacing = style.letterSpacing
|
||||||
mirrorDiv.style.textTransform = style.textTransform
|
mirrorDiv.style.textTransform = style.textTransform
|
||||||
mirrorDiv.style.textIndent = style.textIndent
|
mirrorDiv.style.textIndent = style.textIndent
|
||||||
mirrorDiv.style.textAlign = style.textAlign
|
mirrorDiv.style.textAlign = style.textAlign
|
||||||
|
mirrorDiv.textContent = message.substring(0, caretPos)
|
||||||
|
|
||||||
mirrorDiv.textContent = text.substring(0, caretPosition)
|
const caretMarker = document.createElement('span')
|
||||||
|
caretMarker.style.display = 'inline-block'
|
||||||
|
caretMarker.style.width = '0px'
|
||||||
|
caretMarker.style.padding = '0'
|
||||||
|
caretMarker.style.border = '0'
|
||||||
|
mirrorDiv.appendChild(caretMarker)
|
||||||
|
|
||||||
const caretMarker = document.createElement('span')
|
document.body.appendChild(mirrorDiv)
|
||||||
caretMarker.style.display = 'inline-block'
|
const markerRect = caretMarker.getBoundingClientRect()
|
||||||
caretMarker.style.width = '0px'
|
const mirrorRect = mirrorDiv.getBoundingClientRect()
|
||||||
caretMarker.style.padding = '0'
|
document.body.removeChild(mirrorDiv)
|
||||||
caretMarker.style.border = '0'
|
|
||||||
mirrorDiv.appendChild(caretMarker)
|
|
||||||
|
|
||||||
document.body.appendChild(mirrorDiv)
|
const caretViewport = {
|
||||||
const markerRect = caretMarker.getBoundingClientRect()
|
left: textareaRect.left + (markerRect.left - mirrorRect.left) - textareaEl.scrollLeft,
|
||||||
const mirrorRect = mirrorDiv.getBoundingClientRect()
|
top: textareaRect.top + (markerRect.top - mirrorRect.top) - textareaEl.scrollTop,
|
||||||
document.body.removeChild(mirrorDiv)
|
|
||||||
|
|
||||||
const leftOffset = markerRect.left - mirrorRect.left - textarea.scrollLeft
|
|
||||||
const topOffset = markerRect.top - mirrorRect.top - textarea.scrollTop
|
|
||||||
|
|
||||||
return {
|
|
||||||
left: textareaRect.left + leftOffset,
|
|
||||||
top: textareaRect.top + topOffset,
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const caretPos = getCaretPos()
|
|
||||||
const caretViewport = getCaretViewport(textareaEl, caretPos, message)
|
|
||||||
|
|
||||||
// Decide preferred side based on available space
|
|
||||||
const margin = 8
|
const margin = 8
|
||||||
const spaceAbove = caretViewport.top - margin
|
|
||||||
const spaceBelow = window.innerHeight - caretViewport.top - margin
|
const spaceBelow = window.innerHeight - caretViewport.top - margin
|
||||||
const side: 'top' | 'bottom' = spaceBelow >= spaceAbove ? 'bottom' : 'top'
|
const side: 'top' | 'bottom' = spaceBelow >= caretViewport.top - margin ? 'bottom' : 'top'
|
||||||
|
|
||||||
// Check if we're in folder navigation mode (no query, not in submenu)
|
|
||||||
const isInFolderNavigationMode = !openSubmenuFor && !showAggregatedView
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Popover
|
<Popover open={true} onOpenChange={() => {}}>
|
||||||
open={true}
|
|
||||||
onOpenChange={() => {
|
|
||||||
/* controlled externally */
|
|
||||||
}}
|
|
||||||
>
|
|
||||||
<PopoverAnchor asChild>
|
<PopoverAnchor asChild>
|
||||||
<div
|
<div
|
||||||
style={{
|
style={{
|
||||||
@@ -175,24 +135,22 @@ export function SlashMenu({ mentionMenu, message, onSelectCommand }: SlashMenuPr
|
|||||||
onOpenAutoFocus={(e) => e.preventDefault()}
|
onOpenAutoFocus={(e) => e.preventDefault()}
|
||||||
onCloseAutoFocus={(e) => e.preventDefault()}
|
onCloseAutoFocus={(e) => e.preventDefault()}
|
||||||
>
|
>
|
||||||
<PopoverBackButton />
|
<PopoverBackButton onClick={() => setOpenSubmenuFor(null)} />
|
||||||
<PopoverScrollArea ref={menuListRef} className='space-y-[2px]'>
|
<PopoverScrollArea ref={menuListRef} className='space-y-[2px]'>
|
||||||
{openSubmenuFor === 'Web' ? (
|
{openSubmenuFor === 'Web' ? (
|
||||||
// Web submenu view
|
|
||||||
<>
|
<>
|
||||||
{WEB_COMMANDS.map((cmd, index) => (
|
{WEB_COMMANDS.map((cmd, index) => (
|
||||||
<PopoverItem
|
<PopoverItem
|
||||||
key={cmd.id}
|
key={cmd.id}
|
||||||
onClick={() => onSelectCommand(cmd.label)}
|
onClick={() => onSelectCommand(cmd.id)}
|
||||||
data-idx={index}
|
data-idx={index}
|
||||||
active={index === submenuActiveIndex}
|
active={index === submenuActiveIndex}
|
||||||
>
|
>
|
||||||
<span className='truncate capitalize'>{cmd.label}</span>
|
<span className='truncate'>{cmd.label}</span>
|
||||||
</PopoverItem>
|
</PopoverItem>
|
||||||
))}
|
))}
|
||||||
</>
|
</>
|
||||||
) : showAggregatedView ? (
|
) : showAggregatedView ? (
|
||||||
// Aggregated filtered view
|
|
||||||
<>
|
<>
|
||||||
{filteredCommands && filteredCommands.length === 0 ? (
|
{filteredCommands && filteredCommands.length === 0 ? (
|
||||||
<div className='px-[8px] py-[8px] text-[12px] text-[var(--text-muted)]'>
|
<div className='px-[8px] py-[8px] text-[12px] text-[var(--text-muted)]'>
|
||||||
@@ -202,26 +160,25 @@ export function SlashMenu({ mentionMenu, message, onSelectCommand }: SlashMenuPr
|
|||||||
filteredCommands?.map((cmd, index) => (
|
filteredCommands?.map((cmd, index) => (
|
||||||
<PopoverItem
|
<PopoverItem
|
||||||
key={cmd.id}
|
key={cmd.id}
|
||||||
onClick={() => onSelectCommand(cmd.label)}
|
onClick={() => onSelectCommand(cmd.id)}
|
||||||
data-idx={index}
|
data-idx={index}
|
||||||
active={index === submenuActiveIndex}
|
active={index === submenuActiveIndex}
|
||||||
>
|
>
|
||||||
<span className='truncate capitalize'>{cmd.label}</span>
|
<span className='truncate'>{cmd.label}</span>
|
||||||
</PopoverItem>
|
</PopoverItem>
|
||||||
))
|
))
|
||||||
)}
|
)}
|
||||||
</>
|
</>
|
||||||
) : (
|
) : (
|
||||||
// Folder navigation view
|
|
||||||
<>
|
<>
|
||||||
{TOP_LEVEL_COMMANDS.map((cmd, index) => (
|
{TOP_LEVEL_COMMANDS.map((cmd, index) => (
|
||||||
<PopoverItem
|
<PopoverItem
|
||||||
key={cmd.id}
|
key={cmd.id}
|
||||||
onClick={() => onSelectCommand(cmd.label)}
|
onClick={() => onSelectCommand(cmd.id)}
|
||||||
data-idx={index}
|
data-idx={index}
|
||||||
active={isInFolderNavigationMode && index === mentionActiveIndex}
|
active={isInFolderNavigationMode && index === mentionActiveIndex}
|
||||||
>
|
>
|
||||||
<span className='truncate capitalize'>{cmd.label}</span>
|
<span className='truncate'>{cmd.label}</span>
|
||||||
</PopoverItem>
|
</PopoverItem>
|
||||||
))}
|
))}
|
||||||
|
|
||||||
@@ -235,8 +192,8 @@ export function SlashMenu({ mentionMenu, message, onSelectCommand }: SlashMenuPr
|
|||||||
data-idx={TOP_LEVEL_COMMANDS.length}
|
data-idx={TOP_LEVEL_COMMANDS.length}
|
||||||
>
|
>
|
||||||
{WEB_COMMANDS.map((cmd) => (
|
{WEB_COMMANDS.map((cmd) => (
|
||||||
<PopoverItem key={cmd.id} onClick={() => onSelectCommand(cmd.label)}>
|
<PopoverItem key={cmd.id} onClick={() => onSelectCommand(cmd.id)}>
|
||||||
<span className='truncate capitalize'>{cmd.label}</span>
|
<span className='truncate'>{cmd.label}</span>
|
||||||
</PopoverItem>
|
</PopoverItem>
|
||||||
))}
|
))}
|
||||||
</PopoverFolder>
|
</PopoverFolder>
|
||||||
|
|||||||
@@ -40,6 +40,24 @@ import { useCopilotStore } from '@/stores/panel'
|
|||||||
|
|
||||||
const logger = createLogger('CopilotUserInput')
|
const logger = createLogger('CopilotUserInput')
|
||||||
|
|
||||||
|
const TOP_LEVEL_COMMANDS = ['fast', 'research', 'superagent'] as const
|
||||||
|
const WEB_COMMANDS = ['search', 'read', 'scrape', 'crawl'] as const
|
||||||
|
const ALL_COMMANDS = [...TOP_LEVEL_COMMANDS, ...WEB_COMMANDS]
|
||||||
|
|
||||||
|
const COMMAND_DISPLAY_LABELS: Record<string, string> = {
|
||||||
|
superagent: 'Actions',
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Calculates the next index for circular navigation (wraps around at bounds)
|
||||||
|
*/
|
||||||
|
function getNextIndex(current: number, direction: 'up' | 'down', maxIndex: number): number {
|
||||||
|
if (direction === 'down') {
|
||||||
|
return current >= maxIndex ? 0 : current + 1
|
||||||
|
}
|
||||||
|
return current <= 0 ? maxIndex : current - 1
|
||||||
|
}
|
||||||
|
|
||||||
interface UserInputProps {
|
interface UserInputProps {
|
||||||
onSubmit: (
|
onSubmit: (
|
||||||
message: string,
|
message: string,
|
||||||
@@ -110,7 +128,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
},
|
},
|
||||||
ref
|
ref
|
||||||
) => {
|
) => {
|
||||||
// Refs and external hooks
|
|
||||||
const { data: session } = useSession()
|
const { data: session } = useSession()
|
||||||
const params = useParams()
|
const params = useParams()
|
||||||
const workspaceId = params.workspaceId as string
|
const workspaceId = params.workspaceId as string
|
||||||
@@ -122,19 +139,16 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
selectedModelOverride !== undefined ? selectedModelOverride : copilotStore.selectedModel
|
selectedModelOverride !== undefined ? selectedModelOverride : copilotStore.selectedModel
|
||||||
const setSelectedModel = onModelChangeOverride || copilotStore.setSelectedModel
|
const setSelectedModel = onModelChangeOverride || copilotStore.setSelectedModel
|
||||||
|
|
||||||
// Internal state
|
|
||||||
const [internalMessage, setInternalMessage] = useState('')
|
const [internalMessage, setInternalMessage] = useState('')
|
||||||
const [isNearTop, setIsNearTop] = useState(false)
|
const [isNearTop, setIsNearTop] = useState(false)
|
||||||
const [containerRef, setContainerRef] = useState<HTMLDivElement | null>(null)
|
const [containerRef, setContainerRef] = useState<HTMLDivElement | null>(null)
|
||||||
const [inputContainerRef, setInputContainerRef] = useState<HTMLDivElement | null>(null)
|
const [inputContainerRef, setInputContainerRef] = useState<HTMLDivElement | null>(null)
|
||||||
const [showSlashMenu, setShowSlashMenu] = useState(false)
|
const [showSlashMenu, setShowSlashMenu] = useState(false)
|
||||||
|
|
||||||
// Controlled vs uncontrolled message state
|
|
||||||
const message = controlledValue !== undefined ? controlledValue : internalMessage
|
const message = controlledValue !== undefined ? controlledValue : internalMessage
|
||||||
const setMessage =
|
const setMessage =
|
||||||
controlledValue !== undefined ? onControlledChange || (() => {}) : setInternalMessage
|
controlledValue !== undefined ? onControlledChange || (() => {}) : setInternalMessage
|
||||||
|
|
||||||
// Effective placeholder
|
|
||||||
const effectivePlaceholder =
|
const effectivePlaceholder =
|
||||||
placeholder ||
|
placeholder ||
|
||||||
(mode === 'ask'
|
(mode === 'ask'
|
||||||
@@ -143,11 +157,8 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
? 'Plan your workflow'
|
? 'Plan your workflow'
|
||||||
: 'Plan, search, build anything')
|
: 'Plan, search, build anything')
|
||||||
|
|
||||||
// Custom hooks - order matters for ref sharing
|
|
||||||
// Context management (manages selectedContexts state)
|
|
||||||
const contextManagement = useContextManagement({ message, initialContexts })
|
const contextManagement = useContextManagement({ message, initialContexts })
|
||||||
|
|
||||||
// Mention menu
|
|
||||||
const mentionMenu = useMentionMenu({
|
const mentionMenu = useMentionMenu({
|
||||||
message,
|
message,
|
||||||
selectedContexts: contextManagement.selectedContexts,
|
selectedContexts: contextManagement.selectedContexts,
|
||||||
@@ -155,7 +166,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
onMessageChange: setMessage,
|
onMessageChange: setMessage,
|
||||||
})
|
})
|
||||||
|
|
||||||
// Mention token utilities
|
|
||||||
const mentionTokensWithContext = useMentionTokens({
|
const mentionTokensWithContext = useMentionTokens({
|
||||||
message,
|
message,
|
||||||
selectedContexts: contextManagement.selectedContexts,
|
selectedContexts: contextManagement.selectedContexts,
|
||||||
@@ -183,7 +193,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
isLoading,
|
isLoading,
|
||||||
})
|
})
|
||||||
|
|
||||||
// Insert mention handlers
|
|
||||||
const insertHandlers = useMentionInsertHandlers({
|
const insertHandlers = useMentionInsertHandlers({
|
||||||
mentionMenu,
|
mentionMenu,
|
||||||
workflowId: workflowId || null,
|
workflowId: workflowId || null,
|
||||||
@@ -191,14 +200,12 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
onContextAdd: contextManagement.addContext,
|
onContextAdd: contextManagement.addContext,
|
||||||
})
|
})
|
||||||
|
|
||||||
// Keyboard navigation hook
|
|
||||||
const mentionKeyboard = useMentionKeyboard({
|
const mentionKeyboard = useMentionKeyboard({
|
||||||
mentionMenu,
|
mentionMenu,
|
||||||
mentionData,
|
mentionData,
|
||||||
insertHandlers,
|
insertHandlers,
|
||||||
})
|
})
|
||||||
|
|
||||||
// Expose focus method to parent
|
|
||||||
useImperativeHandle(
|
useImperativeHandle(
|
||||||
ref,
|
ref,
|
||||||
() => ({
|
() => ({
|
||||||
@@ -215,9 +222,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
[mentionMenu.textareaRef]
|
[mentionMenu.textareaRef]
|
||||||
)
|
)
|
||||||
|
|
||||||
// Note: textarea auto-resize is handled by the useTextareaAutoResize hook
|
|
||||||
|
|
||||||
// Load workflows on mount if we have a workflowId
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
if (workflowId) {
|
if (workflowId) {
|
||||||
void mentionData.ensureWorkflowsLoaded()
|
void mentionData.ensureWorkflowsLoaded()
|
||||||
@@ -225,7 +229,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
// eslint-disable-next-line react-hooks/exhaustive-deps
|
// eslint-disable-next-line react-hooks/exhaustive-deps
|
||||||
}, [workflowId])
|
}, [workflowId])
|
||||||
|
|
||||||
// Detect if input is near top of screen
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
const checkPosition = () => {
|
const checkPosition = () => {
|
||||||
if (containerRef) {
|
if (containerRef) {
|
||||||
@@ -253,7 +256,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
}
|
}
|
||||||
}, [containerRef])
|
}, [containerRef])
|
||||||
|
|
||||||
// Also check position when mention menu opens
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
if (mentionMenu.showMentionMenu && containerRef) {
|
if (mentionMenu.showMentionMenu && containerRef) {
|
||||||
const rect = containerRef.getBoundingClientRect()
|
const rect = containerRef.getBoundingClientRect()
|
||||||
@@ -261,7 +263,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
}
|
}
|
||||||
}, [mentionMenu.showMentionMenu, containerRef])
|
}, [mentionMenu.showMentionMenu, containerRef])
|
||||||
|
|
||||||
// Preload mention data when query is active
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
if (!mentionMenu.showMentionMenu || mentionMenu.openSubmenuFor) {
|
if (!mentionMenu.showMentionMenu || mentionMenu.openSubmenuFor) {
|
||||||
return
|
return
|
||||||
@@ -273,7 +274,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
.toLowerCase()
|
.toLowerCase()
|
||||||
|
|
||||||
if (q && q.length > 0) {
|
if (q && q.length > 0) {
|
||||||
// Prefetch all lists when there's any query for instant filtering
|
|
||||||
void mentionData.ensurePastChatsLoaded()
|
void mentionData.ensurePastChatsLoaded()
|
||||||
void mentionData.ensureWorkflowsLoaded()
|
void mentionData.ensureWorkflowsLoaded()
|
||||||
void mentionData.ensureWorkflowBlocksLoaded()
|
void mentionData.ensureWorkflowBlocksLoaded()
|
||||||
@@ -282,15 +282,12 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
void mentionData.ensureTemplatesLoaded()
|
void mentionData.ensureTemplatesLoaded()
|
||||||
void mentionData.ensureLogsLoaded()
|
void mentionData.ensureLogsLoaded()
|
||||||
|
|
||||||
// Reset to first item when query changes
|
|
||||||
mentionMenu.setSubmenuActiveIndex(0)
|
mentionMenu.setSubmenuActiveIndex(0)
|
||||||
requestAnimationFrame(() => mentionMenu.scrollActiveItemIntoView(0))
|
requestAnimationFrame(() => mentionMenu.scrollActiveItemIntoView(0))
|
||||||
}
|
}
|
||||||
// Only depend on values that trigger data loading, not the entire objects
|
|
||||||
// eslint-disable-next-line react-hooks/exhaustive-deps
|
// eslint-disable-next-line react-hooks/exhaustive-deps
|
||||||
}, [mentionMenu.showMentionMenu, mentionMenu.openSubmenuFor, message])
|
}, [mentionMenu.showMentionMenu, mentionMenu.openSubmenuFor, message])
|
||||||
|
|
||||||
// When switching into a submenu, select the first item and scroll to it
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
if (mentionMenu.openSubmenuFor) {
|
if (mentionMenu.openSubmenuFor) {
|
||||||
mentionMenu.setSubmenuActiveIndex(0)
|
mentionMenu.setSubmenuActiveIndex(0)
|
||||||
@@ -299,12 +296,10 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
// eslint-disable-next-line react-hooks/exhaustive-deps
|
// eslint-disable-next-line react-hooks/exhaustive-deps
|
||||||
}, [mentionMenu.openSubmenuFor])
|
}, [mentionMenu.openSubmenuFor])
|
||||||
|
|
||||||
// Handlers
|
|
||||||
const handleSubmit = useCallback(
|
const handleSubmit = useCallback(
|
||||||
async (overrideMessage?: string, options: { preserveInput?: boolean } = {}) => {
|
async (overrideMessage?: string, options: { preserveInput?: boolean } = {}) => {
|
||||||
const targetMessage = overrideMessage ?? message
|
const targetMessage = overrideMessage ?? message
|
||||||
const trimmedMessage = targetMessage.trim()
|
const trimmedMessage = targetMessage.trim()
|
||||||
// Allow submission even when isLoading - store will queue the message
|
|
||||||
if (!trimmedMessage || disabled) return
|
if (!trimmedMessage || disabled) return
|
||||||
|
|
||||||
const failedUploads = fileAttachments.attachedFiles.filter((f) => !f.uploading && !f.key)
|
const failedUploads = fileAttachments.attachedFiles.filter((f) => !f.uploading && !f.key)
|
||||||
@@ -377,17 +372,13 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
|
|
||||||
const handleSlashCommandSelect = useCallback(
|
const handleSlashCommandSelect = useCallback(
|
||||||
(command: string) => {
|
(command: string) => {
|
||||||
// Capitalize the command for display
|
const displayLabel =
|
||||||
const capitalizedCommand = command.charAt(0).toUpperCase() + command.slice(1)
|
COMMAND_DISPLAY_LABELS[command] || command.charAt(0).toUpperCase() + command.slice(1)
|
||||||
|
mentionMenu.replaceActiveSlashWith(displayLabel)
|
||||||
// Replace the active slash query with the capitalized command
|
|
||||||
mentionMenu.replaceActiveSlashWith(capitalizedCommand)
|
|
||||||
|
|
||||||
// Add as a context so it gets highlighted
|
|
||||||
contextManagement.addContext({
|
contextManagement.addContext({
|
||||||
kind: 'slash_command',
|
kind: 'slash_command',
|
||||||
command,
|
command,
|
||||||
label: capitalizedCommand,
|
label: displayLabel,
|
||||||
})
|
})
|
||||||
|
|
||||||
setShowSlashMenu(false)
|
setShowSlashMenu(false)
|
||||||
@@ -398,7 +389,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
|
|
||||||
const handleKeyDown = useCallback(
|
const handleKeyDown = useCallback(
|
||||||
(e: KeyboardEvent<HTMLTextAreaElement>) => {
|
(e: KeyboardEvent<HTMLTextAreaElement>) => {
|
||||||
// Escape key handling
|
|
||||||
if (e.key === 'Escape' && (mentionMenu.showMentionMenu || showSlashMenu)) {
|
if (e.key === 'Escape' && (mentionMenu.showMentionMenu || showSlashMenu)) {
|
||||||
e.preventDefault()
|
e.preventDefault()
|
||||||
if (mentionMenu.openSubmenuFor) {
|
if (mentionMenu.openSubmenuFor) {
|
||||||
@@ -411,65 +401,33 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Arrow navigation in slash menu
|
|
||||||
if (showSlashMenu) {
|
if (showSlashMenu) {
|
||||||
const TOP_LEVEL_COMMANDS = ['fast', 'plan', 'debug', 'research', 'deploy', 'superagent']
|
|
||||||
const WEB_COMMANDS = ['search', 'read', 'scrape', 'crawl']
|
|
||||||
const ALL_COMMANDS = [...TOP_LEVEL_COMMANDS, ...WEB_COMMANDS]
|
|
||||||
|
|
||||||
const caretPos = mentionMenu.getCaretPos()
|
const caretPos = mentionMenu.getCaretPos()
|
||||||
const activeSlash = mentionMenu.getActiveSlashQueryAtPosition(caretPos, message)
|
const activeSlash = mentionMenu.getActiveSlashQueryAtPosition(caretPos, message)
|
||||||
const query = activeSlash?.query.trim().toLowerCase() || ''
|
const query = activeSlash?.query.trim().toLowerCase() || ''
|
||||||
const showAggregatedView = query.length > 0
|
const showAggregatedView = query.length > 0
|
||||||
|
const direction = e.key === 'ArrowDown' ? 'down' : 'up'
|
||||||
|
|
||||||
if (e.key === 'ArrowDown' || e.key === 'ArrowUp') {
|
if (e.key === 'ArrowDown' || e.key === 'ArrowUp') {
|
||||||
e.preventDefault()
|
e.preventDefault()
|
||||||
|
|
||||||
if (mentionMenu.openSubmenuFor === 'Web') {
|
if (mentionMenu.openSubmenuFor === 'Web') {
|
||||||
// Navigate in Web submenu
|
|
||||||
const last = WEB_COMMANDS.length - 1
|
|
||||||
mentionMenu.setSubmenuActiveIndex((prev) => {
|
mentionMenu.setSubmenuActiveIndex((prev) => {
|
||||||
const next =
|
const next = getNextIndex(prev, direction, WEB_COMMANDS.length - 1)
|
||||||
e.key === 'ArrowDown'
|
|
||||||
? prev >= last
|
|
||||||
? 0
|
|
||||||
: prev + 1
|
|
||||||
: prev <= 0
|
|
||||||
? last
|
|
||||||
: prev - 1
|
|
||||||
requestAnimationFrame(() => mentionMenu.scrollActiveItemIntoView(next))
|
requestAnimationFrame(() => mentionMenu.scrollActiveItemIntoView(next))
|
||||||
return next
|
return next
|
||||||
})
|
})
|
||||||
} else if (showAggregatedView) {
|
} else if (showAggregatedView) {
|
||||||
// Navigate in filtered view
|
|
||||||
const filtered = ALL_COMMANDS.filter((cmd) => cmd.includes(query))
|
const filtered = ALL_COMMANDS.filter((cmd) => cmd.includes(query))
|
||||||
const last = Math.max(0, filtered.length - 1)
|
|
||||||
mentionMenu.setSubmenuActiveIndex((prev) => {
|
mentionMenu.setSubmenuActiveIndex((prev) => {
|
||||||
if (filtered.length === 0) return 0
|
if (filtered.length === 0) return 0
|
||||||
const next =
|
const next = getNextIndex(prev, direction, filtered.length - 1)
|
||||||
e.key === 'ArrowDown'
|
|
||||||
? prev >= last
|
|
||||||
? 0
|
|
||||||
: prev + 1
|
|
||||||
: prev <= 0
|
|
||||||
? last
|
|
||||||
: prev - 1
|
|
||||||
requestAnimationFrame(() => mentionMenu.scrollActiveItemIntoView(next))
|
requestAnimationFrame(() => mentionMenu.scrollActiveItemIntoView(next))
|
||||||
return next
|
return next
|
||||||
})
|
})
|
||||||
} else {
|
} else {
|
||||||
// Navigate in folder view (top-level + Web folder)
|
|
||||||
const totalItems = TOP_LEVEL_COMMANDS.length + 1 // +1 for Web folder
|
|
||||||
const last = totalItems - 1
|
|
||||||
mentionMenu.setMentionActiveIndex((prev) => {
|
mentionMenu.setMentionActiveIndex((prev) => {
|
||||||
const next =
|
const next = getNextIndex(prev, direction, TOP_LEVEL_COMMANDS.length)
|
||||||
e.key === 'ArrowDown'
|
|
||||||
? prev >= last
|
|
||||||
? 0
|
|
||||||
: prev + 1
|
|
||||||
: prev <= 0
|
|
||||||
? last
|
|
||||||
: prev - 1
|
|
||||||
requestAnimationFrame(() => mentionMenu.scrollActiveItemIntoView(next))
|
requestAnimationFrame(() => mentionMenu.scrollActiveItemIntoView(next))
|
||||||
return next
|
return next
|
||||||
})
|
})
|
||||||
@@ -477,11 +435,9 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Arrow right to enter Web submenu
|
|
||||||
if (e.key === 'ArrowRight') {
|
if (e.key === 'ArrowRight') {
|
||||||
e.preventDefault()
|
e.preventDefault()
|
||||||
if (!showAggregatedView && !mentionMenu.openSubmenuFor) {
|
if (!showAggregatedView && !mentionMenu.openSubmenuFor) {
|
||||||
// Check if Web folder is selected (it's after all top-level commands)
|
|
||||||
if (mentionMenu.mentionActiveIndex === TOP_LEVEL_COMMANDS.length) {
|
if (mentionMenu.mentionActiveIndex === TOP_LEVEL_COMMANDS.length) {
|
||||||
mentionMenu.setOpenSubmenuFor('Web')
|
mentionMenu.setOpenSubmenuFor('Web')
|
||||||
mentionMenu.setSubmenuActiveIndex(0)
|
mentionMenu.setSubmenuActiveIndex(0)
|
||||||
@@ -490,7 +446,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Arrow left to exit submenu
|
|
||||||
if (e.key === 'ArrowLeft') {
|
if (e.key === 'ArrowLeft') {
|
||||||
e.preventDefault()
|
e.preventDefault()
|
||||||
if (mentionMenu.openSubmenuFor) {
|
if (mentionMenu.openSubmenuFor) {
|
||||||
@@ -500,44 +455,33 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Arrow navigation in mention menu
|
|
||||||
if (mentionKeyboard.handleArrowNavigation(e)) return
|
if (mentionKeyboard.handleArrowNavigation(e)) return
|
||||||
if (mentionKeyboard.handleArrowRight(e)) return
|
if (mentionKeyboard.handleArrowRight(e)) return
|
||||||
if (mentionKeyboard.handleArrowLeft(e)) return
|
if (mentionKeyboard.handleArrowLeft(e)) return
|
||||||
|
|
||||||
// Enter key handling
|
|
||||||
if (e.key === 'Enter' && !e.shiftKey && !e.nativeEvent.isComposing) {
|
if (e.key === 'Enter' && !e.shiftKey && !e.nativeEvent.isComposing) {
|
||||||
e.preventDefault()
|
e.preventDefault()
|
||||||
if (showSlashMenu) {
|
if (showSlashMenu) {
|
||||||
const TOP_LEVEL_COMMANDS = ['fast', 'plan', 'debug', 'research', 'deploy', 'superagent']
|
|
||||||
const WEB_COMMANDS = ['search', 'read', 'scrape', 'crawl']
|
|
||||||
const ALL_COMMANDS = [...TOP_LEVEL_COMMANDS, ...WEB_COMMANDS]
|
|
||||||
|
|
||||||
const caretPos = mentionMenu.getCaretPos()
|
const caretPos = mentionMenu.getCaretPos()
|
||||||
const activeSlash = mentionMenu.getActiveSlashQueryAtPosition(caretPos, message)
|
const activeSlash = mentionMenu.getActiveSlashQueryAtPosition(caretPos, message)
|
||||||
const query = activeSlash?.query.trim().toLowerCase() || ''
|
const query = activeSlash?.query.trim().toLowerCase() || ''
|
||||||
const showAggregatedView = query.length > 0
|
const showAggregatedView = query.length > 0
|
||||||
|
|
||||||
if (mentionMenu.openSubmenuFor === 'Web') {
|
if (mentionMenu.openSubmenuFor === 'Web') {
|
||||||
// Select from Web submenu
|
|
||||||
const selectedCommand =
|
const selectedCommand =
|
||||||
WEB_COMMANDS[mentionMenu.submenuActiveIndex] || WEB_COMMANDS[0]
|
WEB_COMMANDS[mentionMenu.submenuActiveIndex] || WEB_COMMANDS[0]
|
||||||
handleSlashCommandSelect(selectedCommand)
|
handleSlashCommandSelect(selectedCommand)
|
||||||
} else if (showAggregatedView) {
|
} else if (showAggregatedView) {
|
||||||
// Select from filtered view
|
|
||||||
const filtered = ALL_COMMANDS.filter((cmd) => cmd.includes(query))
|
const filtered = ALL_COMMANDS.filter((cmd) => cmd.includes(query))
|
||||||
if (filtered.length > 0) {
|
if (filtered.length > 0) {
|
||||||
const selectedCommand = filtered[mentionMenu.submenuActiveIndex] || filtered[0]
|
const selectedCommand = filtered[mentionMenu.submenuActiveIndex] || filtered[0]
|
||||||
handleSlashCommandSelect(selectedCommand)
|
handleSlashCommandSelect(selectedCommand)
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
// Folder navigation view
|
|
||||||
const selectedIndex = mentionMenu.mentionActiveIndex
|
const selectedIndex = mentionMenu.mentionActiveIndex
|
||||||
if (selectedIndex < TOP_LEVEL_COMMANDS.length) {
|
if (selectedIndex < TOP_LEVEL_COMMANDS.length) {
|
||||||
// Top-level command selected
|
|
||||||
handleSlashCommandSelect(TOP_LEVEL_COMMANDS[selectedIndex])
|
handleSlashCommandSelect(TOP_LEVEL_COMMANDS[selectedIndex])
|
||||||
} else if (selectedIndex === TOP_LEVEL_COMMANDS.length) {
|
} else if (selectedIndex === TOP_LEVEL_COMMANDS.length) {
|
||||||
// Web folder selected - open it
|
|
||||||
mentionMenu.setOpenSubmenuFor('Web')
|
mentionMenu.setOpenSubmenuFor('Web')
|
||||||
mentionMenu.setSubmenuActiveIndex(0)
|
mentionMenu.setSubmenuActiveIndex(0)
|
||||||
}
|
}
|
||||||
@@ -552,7 +496,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Handle mention token behavior (backspace, delete, arrow keys) when menu is closed
|
|
||||||
if (!mentionMenu.showMentionMenu) {
|
if (!mentionMenu.showMentionMenu) {
|
||||||
const textarea = mentionMenu.textareaRef.current
|
const textarea = mentionMenu.textareaRef.current
|
||||||
const selStart = textarea?.selectionStart ?? 0
|
const selStart = textarea?.selectionStart ?? 0
|
||||||
@@ -561,11 +504,8 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
|
|
||||||
if (e.key === 'Backspace' || e.key === 'Delete') {
|
if (e.key === 'Backspace' || e.key === 'Delete') {
|
||||||
if (selectionLength > 0) {
|
if (selectionLength > 0) {
|
||||||
// Multi-character selection: Clean up contexts for any overlapping mentions
|
|
||||||
// but let the default behavior handle the actual text deletion
|
|
||||||
mentionTokensWithContext.removeContextsInSelection(selStart, selEnd)
|
mentionTokensWithContext.removeContextsInSelection(selStart, selEnd)
|
||||||
} else {
|
} else {
|
||||||
// Single character delete - check if cursor is inside/at a mention token
|
|
||||||
const ranges = mentionTokensWithContext.computeMentionRanges()
|
const ranges = mentionTokensWithContext.computeMentionRanges()
|
||||||
const target =
|
const target =
|
||||||
e.key === 'Backspace'
|
e.key === 'Backspace'
|
||||||
@@ -604,7 +544,6 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Prevent typing inside token
|
|
||||||
if (e.key.length === 1 || e.key === 'Space') {
|
if (e.key.length === 1 || e.key === 'Space') {
|
||||||
const blocked =
|
const blocked =
|
||||||
selectionLength === 0 && !!mentionTokensWithContext.findRangeContaining(selStart)
|
selectionLength === 0 && !!mentionTokensWithContext.findRangeContaining(selStart)
|
||||||
@@ -637,14 +576,10 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
const newValue = e.target.value
|
const newValue = e.target.value
|
||||||
setMessage(newValue)
|
setMessage(newValue)
|
||||||
|
|
||||||
// Skip mention menu logic if mentions are disabled
|
|
||||||
if (disableMentions) return
|
if (disableMentions) return
|
||||||
|
|
||||||
const caret = e.target.selectionStart ?? newValue.length
|
const caret = e.target.selectionStart ?? newValue.length
|
||||||
|
|
||||||
// Check for @ mention trigger
|
|
||||||
const activeMention = mentionMenu.getActiveMentionQueryAtPosition(caret, newValue)
|
const activeMention = mentionMenu.getActiveMentionQueryAtPosition(caret, newValue)
|
||||||
// Check for / slash command trigger
|
|
||||||
const activeSlash = mentionMenu.getActiveSlashQueryAtPosition(caret, newValue)
|
const activeSlash = mentionMenu.getActiveSlashQueryAtPosition(caret, newValue)
|
||||||
|
|
||||||
if (activeMention) {
|
if (activeMention) {
|
||||||
@@ -686,84 +621,66 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
}
|
}
|
||||||
}, [mentionMenu.textareaRef, mentionTokensWithContext])
|
}, [mentionMenu.textareaRef, mentionTokensWithContext])
|
||||||
|
|
||||||
const handleOpenMentionMenuWithAt = useCallback(() => {
|
const insertTriggerAndOpenMenu = useCallback(
|
||||||
if (disabled || isLoading) return
|
(trigger: '@' | '/') => {
|
||||||
const textarea = mentionMenu.textareaRef.current
|
if (disabled || isLoading) return
|
||||||
if (!textarea) return
|
const textarea = mentionMenu.textareaRef.current
|
||||||
textarea.focus()
|
if (!textarea) return
|
||||||
const pos = textarea.selectionStart ?? message.length
|
|
||||||
const needsSpaceBefore = pos > 0 && !/\s/.test(message.charAt(pos - 1))
|
|
||||||
|
|
||||||
const insertText = needsSpaceBefore ? ' @' : '@'
|
|
||||||
const start = textarea.selectionStart ?? message.length
|
|
||||||
const end = textarea.selectionEnd ?? message.length
|
|
||||||
const before = message.slice(0, start)
|
|
||||||
const after = message.slice(end)
|
|
||||||
const next = `${before}${insertText}${after}`
|
|
||||||
setMessage(next)
|
|
||||||
|
|
||||||
setTimeout(() => {
|
|
||||||
const newPos = before.length + insertText.length
|
|
||||||
textarea.setSelectionRange(newPos, newPos)
|
|
||||||
textarea.focus()
|
textarea.focus()
|
||||||
}, 0)
|
const start = textarea.selectionStart ?? message.length
|
||||||
|
const end = textarea.selectionEnd ?? message.length
|
||||||
|
const needsSpaceBefore = start > 0 && !/\s/.test(message.charAt(start - 1))
|
||||||
|
|
||||||
mentionMenu.setShowMentionMenu(true)
|
const insertText = needsSpaceBefore ? ` ${trigger}` : trigger
|
||||||
mentionMenu.setOpenSubmenuFor(null)
|
const before = message.slice(0, start)
|
||||||
mentionMenu.setMentionActiveIndex(0)
|
const after = message.slice(end)
|
||||||
mentionMenu.setSubmenuActiveIndex(0)
|
setMessage(`${before}${insertText}${after}`)
|
||||||
}, [disabled, isLoading, mentionMenu, message, setMessage])
|
|
||||||
|
|
||||||
const handleOpenSlashMenu = useCallback(() => {
|
setTimeout(() => {
|
||||||
if (disabled || isLoading) return
|
const newPos = before.length + insertText.length
|
||||||
const textarea = mentionMenu.textareaRef.current
|
textarea.setSelectionRange(newPos, newPos)
|
||||||
if (!textarea) return
|
textarea.focus()
|
||||||
textarea.focus()
|
}, 0)
|
||||||
const pos = textarea.selectionStart ?? message.length
|
|
||||||
const needsSpaceBefore = pos > 0 && !/\s/.test(message.charAt(pos - 1))
|
|
||||||
|
|
||||||
const insertText = needsSpaceBefore ? ' /' : '/'
|
if (trigger === '@') {
|
||||||
const start = textarea.selectionStart ?? message.length
|
mentionMenu.setShowMentionMenu(true)
|
||||||
const end = textarea.selectionEnd ?? message.length
|
mentionMenu.setOpenSubmenuFor(null)
|
||||||
const before = message.slice(0, start)
|
mentionMenu.setMentionActiveIndex(0)
|
||||||
const after = message.slice(end)
|
} else {
|
||||||
const next = `${before}${insertText}${after}`
|
setShowSlashMenu(true)
|
||||||
setMessage(next)
|
}
|
||||||
|
mentionMenu.setSubmenuActiveIndex(0)
|
||||||
|
},
|
||||||
|
[disabled, isLoading, mentionMenu, message, setMessage]
|
||||||
|
)
|
||||||
|
|
||||||
setTimeout(() => {
|
const handleOpenMentionMenuWithAt = useCallback(
|
||||||
const newPos = before.length + insertText.length
|
() => insertTriggerAndOpenMenu('@'),
|
||||||
textarea.setSelectionRange(newPos, newPos)
|
[insertTriggerAndOpenMenu]
|
||||||
textarea.focus()
|
)
|
||||||
}, 0)
|
|
||||||
|
|
||||||
setShowSlashMenu(true)
|
const handleOpenSlashMenu = useCallback(
|
||||||
mentionMenu.setSubmenuActiveIndex(0)
|
() => insertTriggerAndOpenMenu('/'),
|
||||||
}, [disabled, isLoading, mentionMenu, message, setMessage])
|
[insertTriggerAndOpenMenu]
|
||||||
|
)
|
||||||
|
|
||||||
const canSubmit = message.trim().length > 0 && !disabled && !isLoading
|
const canSubmit = message.trim().length > 0 && !disabled && !isLoading
|
||||||
const showAbortButton = isLoading && onAbort
|
const showAbortButton = isLoading && onAbort
|
||||||
|
|
||||||
// Render overlay content with highlighted mentions
|
|
||||||
const renderOverlayContent = useCallback(() => {
|
const renderOverlayContent = useCallback(() => {
|
||||||
const contexts = contextManagement.selectedContexts
|
const contexts = contextManagement.selectedContexts
|
||||||
|
|
||||||
// Handle empty message
|
|
||||||
if (!message) {
|
if (!message) {
|
||||||
return <span>{'\u00A0'}</span>
|
return <span>{'\u00A0'}</span>
|
||||||
}
|
}
|
||||||
|
|
||||||
// If no contexts, render the message directly with proper newline handling
|
|
||||||
if (contexts.length === 0) {
|
if (contexts.length === 0) {
|
||||||
// Add a zero-width space at the end if message ends with newline
|
|
||||||
// This ensures the newline is rendered and height is calculated correctly
|
|
||||||
const displayText = message.endsWith('\n') ? `${message}\u200B` : message
|
const displayText = message.endsWith('\n') ? `${message}\u200B` : message
|
||||||
return <span>{displayText}</span>
|
return <span>{displayText}</span>
|
||||||
}
|
}
|
||||||
|
|
||||||
const elements: React.ReactNode[] = []
|
const elements: React.ReactNode[] = []
|
||||||
const labels = contexts.map((c) => c.label).filter(Boolean)
|
|
||||||
|
|
||||||
// Build ranges for all mentions to highlight them including spaces
|
|
||||||
const ranges = mentionTokensWithContext.computeMentionRanges()
|
const ranges = mentionTokensWithContext.computeMentionRanges()
|
||||||
|
|
||||||
if (ranges.length === 0) {
|
if (ranges.length === 0) {
|
||||||
@@ -775,14 +692,11 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
for (let i = 0; i < ranges.length; i++) {
|
for (let i = 0; i < ranges.length; i++) {
|
||||||
const range = ranges[i]
|
const range = ranges[i]
|
||||||
|
|
||||||
// Add text before mention
|
|
||||||
if (range.start > lastIndex) {
|
if (range.start > lastIndex) {
|
||||||
const before = message.slice(lastIndex, range.start)
|
const before = message.slice(lastIndex, range.start)
|
||||||
elements.push(<span key={`text-${i}-${lastIndex}-${range.start}`}>{before}</span>)
|
elements.push(<span key={`text-${i}-${lastIndex}-${range.start}`}>{before}</span>)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add highlighted mention (including spaces)
|
|
||||||
// Use index + start + end to ensure unique keys even with duplicate contexts
|
|
||||||
const mentionText = message.slice(range.start, range.end)
|
const mentionText = message.slice(range.start, range.end)
|
||||||
elements.push(
|
elements.push(
|
||||||
<span
|
<span
|
||||||
@@ -797,12 +711,10 @@ const UserInput = forwardRef<UserInputRef, UserInputProps>(
|
|||||||
|
|
||||||
const tail = message.slice(lastIndex)
|
const tail = message.slice(lastIndex)
|
||||||
if (tail) {
|
if (tail) {
|
||||||
// Add a zero-width space at the end if tail ends with newline
|
|
||||||
const displayTail = tail.endsWith('\n') ? `${tail}\u200B` : tail
|
const displayTail = tail.endsWith('\n') ? `${tail}\u200B` : tail
|
||||||
elements.push(<span key={`tail-${lastIndex}`}>{displayTail}</span>)
|
elements.push(<span key={`tail-${lastIndex}`}>{displayTail}</span>)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Ensure there's always something to render for height calculation
|
|
||||||
return elements.length > 0 ? elements : <span>{'\u00A0'}</span>
|
return elements.length > 0 ? elements : <span>{'\u00A0'}</span>
|
||||||
}, [message, contextManagement.selectedContexts, mentionTokensWithContext])
|
}, [message, contextManagement.selectedContexts, mentionTokensWithContext])
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { memo, useCallback, useMemo, useRef } from 'react'
|
import { memo, useMemo, useRef } from 'react'
|
||||||
import { RepeatIcon, SplitIcon } from 'lucide-react'
|
import { RepeatIcon, SplitIcon } from 'lucide-react'
|
||||||
import { Handle, type NodeProps, Position, useReactFlow } from 'reactflow'
|
import { Handle, type NodeProps, Position, useReactFlow } from 'reactflow'
|
||||||
import { Button, Trash } from '@/components/emcn'
|
import { Button, Trash } from '@/components/emcn'
|
||||||
@@ -8,7 +8,6 @@ import { type DiffStatus, hasDiffStatus } from '@/lib/workflows/diff/types'
|
|||||||
import { useCurrentWorkflow } from '@/app/workspace/[workspaceId]/w/[workflowId]/hooks'
|
import { useCurrentWorkflow } from '@/app/workspace/[workspaceId]/w/[workflowId]/hooks'
|
||||||
import { useCollaborativeWorkflow } from '@/hooks/use-collaborative-workflow'
|
import { useCollaborativeWorkflow } from '@/hooks/use-collaborative-workflow'
|
||||||
import { usePanelEditorStore } from '@/stores/panel'
|
import { usePanelEditorStore } from '@/stores/panel'
|
||||||
import { useWorkflowStore } from '@/stores/workflows/workflow/store'
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Global styles for subflow nodes (loop and parallel containers).
|
* Global styles for subflow nodes (loop and parallel containers).
|
||||||
@@ -52,8 +51,6 @@ export interface SubflowNodeData {
|
|||||||
isPreviewSelected?: boolean
|
isPreviewSelected?: boolean
|
||||||
kind: 'loop' | 'parallel'
|
kind: 'loop' | 'parallel'
|
||||||
name?: string
|
name?: string
|
||||||
/** The ID of the group this subflow belongs to */
|
|
||||||
groupId?: string
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -65,9 +62,8 @@ export interface SubflowNodeData {
|
|||||||
* @returns Rendered subflow node component
|
* @returns Rendered subflow node component
|
||||||
*/
|
*/
|
||||||
export const SubflowNodeComponent = memo(({ data, id }: NodeProps<SubflowNodeData>) => {
|
export const SubflowNodeComponent = memo(({ data, id }: NodeProps<SubflowNodeData>) => {
|
||||||
const { getNodes, setNodes } = useReactFlow()
|
const { getNodes } = useReactFlow()
|
||||||
const { collaborativeBatchRemoveBlocks } = useCollaborativeWorkflow()
|
const { collaborativeBatchRemoveBlocks } = useCollaborativeWorkflow()
|
||||||
const { getGroups } = useWorkflowStore()
|
|
||||||
const blockRef = useRef<HTMLDivElement>(null)
|
const blockRef = useRef<HTMLDivElement>(null)
|
||||||
|
|
||||||
const currentWorkflow = useCurrentWorkflow()
|
const currentWorkflow = useCurrentWorkflow()
|
||||||
@@ -144,57 +140,10 @@ export const SubflowNodeComponent = memo(({ data, id }: NodeProps<SubflowNodeDat
|
|||||||
diffStatus === 'edited' && 'ring-[var(--warning)]'
|
diffStatus === 'edited' && 'ring-[var(--warning)]'
|
||||||
)
|
)
|
||||||
|
|
||||||
/**
|
|
||||||
* Expands selection to include all group members on mouse down.
|
|
||||||
* This ensures that when a user starts dragging a subflow in a group,
|
|
||||||
* all other blocks in the group are also selected and will move together.
|
|
||||||
*/
|
|
||||||
const handleGroupMouseDown = useCallback(
|
|
||||||
(e: React.MouseEvent) => {
|
|
||||||
// Only process left mouse button clicks
|
|
||||||
if (e.button !== 0) return
|
|
||||||
|
|
||||||
const groupId = data.groupId
|
|
||||||
if (!groupId) return
|
|
||||||
|
|
||||||
const groups = getGroups()
|
|
||||||
const group = groups[groupId]
|
|
||||||
if (!group || group.blockIds.length <= 1) return
|
|
||||||
|
|
||||||
const groupBlockIds = new Set(group.blockIds)
|
|
||||||
const allNodes = getNodes()
|
|
||||||
|
|
||||||
// Check if all group members are already selected
|
|
||||||
const allSelected = [...groupBlockIds].every((blockId) =>
|
|
||||||
allNodes.find((n) => n.id === blockId && n.selected)
|
|
||||||
)
|
|
||||||
|
|
||||||
if (allSelected) return
|
|
||||||
|
|
||||||
// Expand selection to include all group members
|
|
||||||
setNodes((nodes) =>
|
|
||||||
nodes.map((n) => {
|
|
||||||
const isInGroup = groupBlockIds.has(n.id)
|
|
||||||
const isThisBlock = n.id === id
|
|
||||||
return {
|
|
||||||
...n,
|
|
||||||
selected: isInGroup ? true : n.selected,
|
|
||||||
data: {
|
|
||||||
...n.data,
|
|
||||||
// Mark as grouped selection if in group but not the directly clicked block
|
|
||||||
isGroupedSelection: isInGroup && !isThisBlock && !n.selected,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
})
|
|
||||||
)
|
|
||||||
},
|
|
||||||
[id, data.groupId, getNodes, setNodes, getGroups]
|
|
||||||
)
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
<SubflowNodeStyles />
|
<SubflowNodeStyles />
|
||||||
<div className='group relative' onMouseDown={handleGroupMouseDown}>
|
<div className='group relative'>
|
||||||
<div
|
<div
|
||||||
ref={blockRef}
|
ref={blockRef}
|
||||||
onClick={() => setCurrentBlockId(id)}
|
onClick={() => setCurrentBlockId(id)}
|
||||||
|
|||||||
@@ -320,12 +320,14 @@ export function Terminal() {
|
|||||||
} = useTerminalStore()
|
} = useTerminalStore()
|
||||||
const isExpanded = useTerminalStore((state) => state.terminalHeight > NEAR_MIN_THRESHOLD)
|
const isExpanded = useTerminalStore((state) => state.terminalHeight > NEAR_MIN_THRESHOLD)
|
||||||
const { activeWorkflowId } = useWorkflowRegistry()
|
const { activeWorkflowId } = useWorkflowRegistry()
|
||||||
|
const hasConsoleHydrated = useTerminalConsoleStore((state) => state._hasHydrated)
|
||||||
const workflowEntriesSelector = useCallback(
|
const workflowEntriesSelector = useCallback(
|
||||||
(state: { entries: ConsoleEntry[] }) =>
|
(state: { entries: ConsoleEntry[] }) =>
|
||||||
state.entries.filter((entry) => entry.workflowId === activeWorkflowId),
|
state.entries.filter((entry) => entry.workflowId === activeWorkflowId),
|
||||||
[activeWorkflowId]
|
[activeWorkflowId]
|
||||||
)
|
)
|
||||||
const entries = useTerminalConsoleStore(useShallow(workflowEntriesSelector))
|
const entriesFromStore = useTerminalConsoleStore(useShallow(workflowEntriesSelector))
|
||||||
|
const entries = hasConsoleHydrated ? entriesFromStore : []
|
||||||
const clearWorkflowConsole = useTerminalConsoleStore((state) => state.clearWorkflowConsole)
|
const clearWorkflowConsole = useTerminalConsoleStore((state) => state.clearWorkflowConsole)
|
||||||
const exportConsoleCSV = useTerminalConsoleStore((state) => state.exportConsoleCSV)
|
const exportConsoleCSV = useTerminalConsoleStore((state) => state.exportConsoleCSV)
|
||||||
const [selectedEntry, setSelectedEntry] = useState<ConsoleEntry | null>(null)
|
const [selectedEntry, setSelectedEntry] = useState<ConsoleEntry | null>(null)
|
||||||
|
|||||||
@@ -107,7 +107,7 @@ export const ActionBar = memo(
|
|||||||
return (
|
return (
|
||||||
<div
|
<div
|
||||||
className={cn(
|
className={cn(
|
||||||
'-top-[46px] absolute right-0 z-[100]',
|
'-top-[46px] absolute right-0',
|
||||||
'flex flex-row items-center',
|
'flex flex-row items-center',
|
||||||
'opacity-0 transition-opacity duration-200 group-hover:opacity-100',
|
'opacity-0 transition-opacity duration-200 group-hover:opacity-100',
|
||||||
'gap-[5px] rounded-[10px] bg-[var(--surface-4)] p-[5px]'
|
'gap-[5px] rounded-[10px] bg-[var(--surface-4)] p-[5px]'
|
||||||
|
|||||||
@@ -12,10 +12,6 @@ export interface WorkflowBlockProps {
|
|||||||
isPreview?: boolean
|
isPreview?: boolean
|
||||||
/** Whether this block is selected in preview mode */
|
/** Whether this block is selected in preview mode */
|
||||||
isPreviewSelected?: boolean
|
isPreviewSelected?: boolean
|
||||||
/** Whether this block is selected as part of a group (not directly clicked) */
|
|
||||||
isGroupedSelection?: boolean
|
|
||||||
/** The ID of the group this block belongs to */
|
|
||||||
groupId?: string
|
|
||||||
subBlockValues?: Record<string, any>
|
subBlockValues?: Record<string, any>
|
||||||
blockState?: any
|
blockState?: any
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
import { memo, useCallback, useEffect, useMemo, useRef, useState } from 'react'
|
import { memo, useCallback, useEffect, useMemo, useRef, useState } from 'react'
|
||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { useParams } from 'next/navigation'
|
import { useParams } from 'next/navigation'
|
||||||
import { Handle, type NodeProps, Position, useReactFlow, useUpdateNodeInternals } from 'reactflow'
|
import { Handle, type NodeProps, Position, useUpdateNodeInternals } from 'reactflow'
|
||||||
import { Badge, Tooltip } from '@/components/emcn'
|
import { Badge, Tooltip } from '@/components/emcn'
|
||||||
import { getEnv, isTruthy } from '@/lib/core/config/env'
|
import { getEnv, isTruthy } from '@/lib/core/config/env'
|
||||||
import { cn } from '@/lib/core/utils/cn'
|
import { cn } from '@/lib/core/utils/cn'
|
||||||
@@ -915,65 +915,8 @@ export const WorkflowBlock = memo(function WorkflowBlock({
|
|||||||
const userPermissions = useUserPermissionsContext()
|
const userPermissions = useUserPermissionsContext()
|
||||||
const isWorkflowSelector = type === 'workflow' || type === 'workflow_input'
|
const isWorkflowSelector = type === 'workflow' || type === 'workflow_input'
|
||||||
|
|
||||||
const isGroupedSelection = data.isGroupedSelection ?? false
|
|
||||||
|
|
||||||
// Get React Flow methods for group selection expansion
|
|
||||||
const { getNodes, setNodes } = useReactFlow()
|
|
||||||
const { getGroups } = useWorkflowStore()
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Expands selection to include all group members on mouse down.
|
|
||||||
* This ensures that when a user starts dragging a block in a group,
|
|
||||||
* all other blocks in the group are also selected and will move together.
|
|
||||||
*/
|
|
||||||
const handleGroupMouseDown = useCallback(
|
|
||||||
(e: React.MouseEvent) => {
|
|
||||||
// Only process left mouse button clicks
|
|
||||||
if (e.button !== 0) return
|
|
||||||
|
|
||||||
const groupId = data.groupId
|
|
||||||
if (!groupId) return
|
|
||||||
|
|
||||||
const groups = getGroups()
|
|
||||||
const group = groups[groupId]
|
|
||||||
if (!group || group.blockIds.length <= 1) return
|
|
||||||
|
|
||||||
const groupBlockIds = new Set(group.blockIds)
|
|
||||||
const allNodes = getNodes()
|
|
||||||
|
|
||||||
// Check if all group members are already selected
|
|
||||||
const allSelected = [...groupBlockIds].every((blockId) =>
|
|
||||||
allNodes.find((n) => n.id === blockId && n.selected)
|
|
||||||
)
|
|
||||||
|
|
||||||
if (allSelected) return
|
|
||||||
|
|
||||||
// Expand selection to include all group members
|
|
||||||
setNodes((nodes) =>
|
|
||||||
nodes.map((n) => {
|
|
||||||
const isInGroup = groupBlockIds.has(n.id)
|
|
||||||
const isThisBlock = n.id === id
|
|
||||||
return {
|
|
||||||
...n,
|
|
||||||
selected: isInGroup ? true : n.selected,
|
|
||||||
data: {
|
|
||||||
...n.data,
|
|
||||||
// Mark as grouped selection if in group but not the directly clicked block
|
|
||||||
isGroupedSelection: isInGroup && !isThisBlock && !n.selected,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
})
|
|
||||||
)
|
|
||||||
},
|
|
||||||
[id, data.groupId, getNodes, setNodes, getGroups]
|
|
||||||
)
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<div
|
<div className='group relative'>
|
||||||
className='group relative'
|
|
||||||
data-grouped-selection={isGroupedSelection ? 'true' : undefined}
|
|
||||||
onMouseDown={handleGroupMouseDown}
|
|
||||||
>
|
|
||||||
<div
|
<div
|
||||||
ref={contentRef}
|
ref={contentRef}
|
||||||
onClick={handleClick}
|
onClick={handleClick}
|
||||||
|
|||||||
@@ -30,7 +30,6 @@ interface UseBlockVisualProps {
|
|||||||
export function useBlockVisual({ blockId, data, isPending = false }: UseBlockVisualProps) {
|
export function useBlockVisual({ blockId, data, isPending = false }: UseBlockVisualProps) {
|
||||||
const isPreview = data.isPreview ?? false
|
const isPreview = data.isPreview ?? false
|
||||||
const isPreviewSelected = data.isPreviewSelected ?? false
|
const isPreviewSelected = data.isPreviewSelected ?? false
|
||||||
const isGroupedSelection = data.isGroupedSelection ?? false
|
|
||||||
|
|
||||||
const currentWorkflow = useCurrentWorkflow()
|
const currentWorkflow = useCurrentWorkflow()
|
||||||
const activeWorkflowId = useWorkflowRegistry((state) => state.activeWorkflowId)
|
const activeWorkflowId = useWorkflowRegistry((state) => state.activeWorkflowId)
|
||||||
@@ -65,18 +64,8 @@ export function useBlockVisual({ blockId, data, isPending = false }: UseBlockVis
|
|||||||
diffStatus: isPreview ? undefined : diffStatus,
|
diffStatus: isPreview ? undefined : diffStatus,
|
||||||
runPathStatus,
|
runPathStatus,
|
||||||
isPreviewSelection: isPreview && isPreviewSelected,
|
isPreviewSelection: isPreview && isPreviewSelected,
|
||||||
isGroupedSelection: !isPreview && isGroupedSelection,
|
|
||||||
}),
|
}),
|
||||||
[
|
[isActive, isPending, isDeletedBlock, diffStatus, runPathStatus, isPreview, isPreviewSelected]
|
||||||
isActive,
|
|
||||||
isPending,
|
|
||||||
isDeletedBlock,
|
|
||||||
diffStatus,
|
|
||||||
runPathStatus,
|
|
||||||
isPreview,
|
|
||||||
isPreviewSelected,
|
|
||||||
isGroupedSelection,
|
|
||||||
]
|
|
||||||
)
|
)
|
||||||
|
|
||||||
return {
|
return {
|
||||||
|
|||||||
@@ -35,7 +35,6 @@ export function useCanvasContextMenu({ blocks, getNodes }: UseCanvasContextMenuP
|
|||||||
const block = blocks[n.id]
|
const block = blocks[n.id]
|
||||||
const parentId = block?.data?.parentId
|
const parentId = block?.data?.parentId
|
||||||
const parentType = parentId ? blocks[parentId]?.type : undefined
|
const parentType = parentId ? blocks[parentId]?.type : undefined
|
||||||
const groupId = block?.data?.groupId
|
|
||||||
return {
|
return {
|
||||||
id: n.id,
|
id: n.id,
|
||||||
type: block?.type || '',
|
type: block?.type || '',
|
||||||
@@ -43,7 +42,6 @@ export function useCanvasContextMenu({ blocks, getNodes }: UseCanvasContextMenuP
|
|||||||
horizontalHandles: block?.horizontalHandles ?? false,
|
horizontalHandles: block?.horizontalHandles ?? false,
|
||||||
parentId,
|
parentId,
|
||||||
parentType,
|
parentType,
|
||||||
groupId,
|
|
||||||
}
|
}
|
||||||
}),
|
}),
|
||||||
[blocks]
|
[blocks]
|
||||||
@@ -51,22 +49,14 @@ export function useCanvasContextMenu({ blocks, getNodes }: UseCanvasContextMenuP
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Handle right-click on a node (block)
|
* Handle right-click on a node (block)
|
||||||
* If the node is part of a multiselection, include all selected nodes.
|
|
||||||
* If the node is not selected, just use that node.
|
|
||||||
*/
|
*/
|
||||||
const handleNodeContextMenu = useCallback(
|
const handleNodeContextMenu = useCallback(
|
||||||
(event: React.MouseEvent, node: Node) => {
|
(event: React.MouseEvent, node: Node) => {
|
||||||
event.preventDefault()
|
event.preventDefault()
|
||||||
event.stopPropagation()
|
event.stopPropagation()
|
||||||
|
|
||||||
// Get all currently selected nodes
|
const selectedNodes = getNodes().filter((n) => n.selected)
|
||||||
const allNodes = getNodes()
|
const nodesToUse = selectedNodes.some((n) => n.id === node.id) ? selectedNodes : [node]
|
||||||
const selectedNodes = allNodes.filter((n) => n.selected)
|
|
||||||
|
|
||||||
// If the right-clicked node is already selected, use all selected nodes
|
|
||||||
// Otherwise, just use the right-clicked node
|
|
||||||
const isNodeSelected = selectedNodes.some((n) => n.id === node.id)
|
|
||||||
const nodesToUse = isNodeSelected && selectedNodes.length > 0 ? selectedNodes : [node]
|
|
||||||
|
|
||||||
setPosition({ x: event.clientX, y: event.clientY })
|
setPosition({ x: event.clientX, y: event.clientY })
|
||||||
setSelectedBlocks(nodesToBlockInfos(nodesToUse))
|
setSelectedBlocks(nodesToBlockInfos(nodesToUse))
|
||||||
|
|||||||
@@ -11,7 +11,6 @@ export interface BlockRingOptions {
|
|||||||
diffStatus: BlockDiffStatus
|
diffStatus: BlockDiffStatus
|
||||||
runPathStatus: BlockRunPathStatus
|
runPathStatus: BlockRunPathStatus
|
||||||
isPreviewSelection?: boolean
|
isPreviewSelection?: boolean
|
||||||
isGroupedSelection?: boolean
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -22,15 +21,8 @@ export function getBlockRingStyles(options: BlockRingOptions): {
|
|||||||
hasRing: boolean
|
hasRing: boolean
|
||||||
ringClassName: string
|
ringClassName: string
|
||||||
} {
|
} {
|
||||||
const {
|
const { isActive, isPending, isDeletedBlock, diffStatus, runPathStatus, isPreviewSelection } =
|
||||||
isActive,
|
options
|
||||||
isPending,
|
|
||||||
isDeletedBlock,
|
|
||||||
diffStatus,
|
|
||||||
runPathStatus,
|
|
||||||
isPreviewSelection,
|
|
||||||
isGroupedSelection,
|
|
||||||
} = options
|
|
||||||
|
|
||||||
const hasRing =
|
const hasRing =
|
||||||
isActive ||
|
isActive ||
|
||||||
@@ -38,24 +30,17 @@ export function getBlockRingStyles(options: BlockRingOptions): {
|
|||||||
diffStatus === 'new' ||
|
diffStatus === 'new' ||
|
||||||
diffStatus === 'edited' ||
|
diffStatus === 'edited' ||
|
||||||
isDeletedBlock ||
|
isDeletedBlock ||
|
||||||
!!runPathStatus ||
|
!!runPathStatus
|
||||||
!!isGroupedSelection
|
|
||||||
|
|
||||||
const ringClassName = cn(
|
const ringClassName = cn(
|
||||||
// Grouped selection: more transparent ring for blocks selected as part of a group
|
|
||||||
// Using rgba with the brand-secondary color (#33b4ff) at 40% opacity
|
|
||||||
isGroupedSelection &&
|
|
||||||
!isActive &&
|
|
||||||
'ring-[2px] ring-[rgba(51,180,255,0.4)]',
|
|
||||||
// Preview selection: static blue ring (standard thickness, no animation)
|
// Preview selection: static blue ring (standard thickness, no animation)
|
||||||
isActive && isPreviewSelection && 'ring-[1.75px] ring-[var(--brand-secondary)]',
|
isActive && isPreviewSelection && 'ring-[1.75px] ring-[var(--brand-secondary)]',
|
||||||
// Executing block: pulsing success ring with prominent thickness
|
// Executing block: pulsing success ring with prominent thickness
|
||||||
isActive &&
|
isActive &&
|
||||||
!isPreviewSelection &&
|
!isPreviewSelection &&
|
||||||
!isGroupedSelection &&
|
|
||||||
'ring-[3.5px] ring-[var(--border-success)] animate-ring-pulse',
|
'ring-[3.5px] ring-[var(--border-success)] animate-ring-pulse',
|
||||||
// Non-active states use standard ring utilities (except grouped selection which has its own)
|
// Non-active states use standard ring utilities
|
||||||
!isActive && hasRing && !isGroupedSelection && 'ring-[1.75px]',
|
!isActive && hasRing && 'ring-[1.75px]',
|
||||||
// Pending state: warning ring
|
// Pending state: warning ring
|
||||||
!isActive && isPending && 'ring-[var(--warning)]',
|
!isActive && isPending && 'ring-[var(--warning)]',
|
||||||
// Deleted state (highest priority after active/pending)
|
// Deleted state (highest priority after active/pending)
|
||||||
|
|||||||
@@ -264,14 +264,12 @@ const WorkflowContent = React.memo(() => {
|
|||||||
const canUndo = undoRedoStack.undo.length > 0
|
const canUndo = undoRedoStack.undo.length > 0
|
||||||
const canRedo = undoRedoStack.redo.length > 0
|
const canRedo = undoRedoStack.redo.length > 0
|
||||||
|
|
||||||
const { updateNodeDimensions, setDragStartPosition, getDragStartPosition, getGroups } =
|
const { updateNodeDimensions, setDragStartPosition, getDragStartPosition } = useWorkflowStore(
|
||||||
useWorkflowStore(
|
useShallow((state) => ({
|
||||||
useShallow((state) => ({
|
updateNodeDimensions: state.updateNodeDimensions,
|
||||||
updateNodeDimensions: state.updateNodeDimensions,
|
setDragStartPosition: state.setDragStartPosition,
|
||||||
setDragStartPosition: state.setDragStartPosition,
|
getDragStartPosition: state.getDragStartPosition,
|
||||||
getDragStartPosition: state.getDragStartPosition,
|
}))
|
||||||
getGroups: state.getGroups,
|
|
||||||
}))
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const copilotCleanup = useCopilotStore((state) => state.cleanup)
|
const copilotCleanup = useCopilotStore((state) => state.cleanup)
|
||||||
@@ -358,19 +356,14 @@ const WorkflowContent = React.memo(() => {
|
|||||||
/** Stores source node/handle info when a connection drag starts for drop-on-block detection. */
|
/** Stores source node/handle info when a connection drag starts for drop-on-block detection. */
|
||||||
const connectionSourceRef = useRef<{ nodeId: string; handleId: string } | null>(null)
|
const connectionSourceRef = useRef<{ nodeId: string; handleId: string } | null>(null)
|
||||||
|
|
||||||
|
/** Tracks whether onConnect successfully handled the connection (ReactFlow pattern). */
|
||||||
|
const connectionCompletedRef = useRef(false)
|
||||||
|
|
||||||
/** Stores start positions for multi-node drag undo/redo recording. */
|
/** Stores start positions for multi-node drag undo/redo recording. */
|
||||||
const multiNodeDragStartRef = useRef<Map<string, { x: number; y: number; parentId?: string }>>(
|
const multiNodeDragStartRef = useRef<Map<string, { x: number; y: number; parentId?: string }>>(
|
||||||
new Map()
|
new Map()
|
||||||
)
|
)
|
||||||
|
|
||||||
/**
|
|
||||||
* Stores original positions and parentIds for nodes temporarily parented during group drag.
|
|
||||||
* Key: node ID, Value: { originalPosition, originalParentId }
|
|
||||||
*/
|
|
||||||
const groupDragTempParentsRef = useRef<
|
|
||||||
Map<string, { originalPosition: { x: number; y: number }; originalParentId?: string }>
|
|
||||||
>(new Map())
|
|
||||||
|
|
||||||
/** Stores node IDs to select on next derivedNodes sync (for paste/duplicate operations). */
|
/** Stores node IDs to select on next derivedNodes sync (for paste/duplicate operations). */
|
||||||
const pendingSelectionRef = useRef<Set<string> | null>(null)
|
const pendingSelectionRef = useRef<Set<string> | null>(null)
|
||||||
|
|
||||||
@@ -468,8 +461,6 @@ const WorkflowContent = React.memo(() => {
|
|||||||
collaborativeBatchRemoveBlocks,
|
collaborativeBatchRemoveBlocks,
|
||||||
collaborativeBatchToggleBlockEnabled,
|
collaborativeBatchToggleBlockEnabled,
|
||||||
collaborativeBatchToggleBlockHandles,
|
collaborativeBatchToggleBlockHandles,
|
||||||
collaborativeGroupBlocks,
|
|
||||||
collaborativeUngroupBlocks,
|
|
||||||
undo,
|
undo,
|
||||||
redo,
|
redo,
|
||||||
} = useCollaborativeWorkflow()
|
} = useCollaborativeWorkflow()
|
||||||
@@ -794,35 +785,6 @@ const WorkflowContent = React.memo(() => {
|
|||||||
collaborativeBatchToggleBlockHandles(blockIds)
|
collaborativeBatchToggleBlockHandles(blockIds)
|
||||||
}, [contextMenuBlocks, collaborativeBatchToggleBlockHandles])
|
}, [contextMenuBlocks, collaborativeBatchToggleBlockHandles])
|
||||||
|
|
||||||
const handleContextGroupBlocks = useCallback(() => {
|
|
||||||
const blockIds = contextMenuBlocks.map((block) => block.id)
|
|
||||||
if (blockIds.length >= 2) {
|
|
||||||
// Validate that all blocks share the same parent (or all have no parent)
|
|
||||||
// Blocks inside a subflow cannot be grouped with blocks outside that subflow
|
|
||||||
const parentIds = contextMenuBlocks.map((block) => block.parentId || null)
|
|
||||||
const uniqueParentIds = new Set(parentIds)
|
|
||||||
if (uniqueParentIds.size > 1) {
|
|
||||||
addNotification({
|
|
||||||
level: 'error',
|
|
||||||
message: 'Cannot group blocks from different subflows',
|
|
||||||
})
|
|
||||||
return
|
|
||||||
}
|
|
||||||
collaborativeGroupBlocks(blockIds)
|
|
||||||
}
|
|
||||||
}, [contextMenuBlocks, collaborativeGroupBlocks, addNotification])
|
|
||||||
|
|
||||||
const handleContextUngroupBlocks = useCallback(() => {
|
|
||||||
// Find the first block with a groupId
|
|
||||||
const groupedBlock = contextMenuBlocks.find((block) => block.groupId)
|
|
||||||
if (!groupedBlock?.groupId) return
|
|
||||||
|
|
||||||
// The block's groupId is the group we want to ungroup
|
|
||||||
// This is the direct group the block belongs to, which is the "top level" from the user's perspective
|
|
||||||
// (the most recently created group that contains this block)
|
|
||||||
collaborativeUngroupBlocks(groupedBlock.groupId)
|
|
||||||
}, [contextMenuBlocks, collaborativeUngroupBlocks])
|
|
||||||
|
|
||||||
const handleContextRemoveFromSubflow = useCallback(() => {
|
const handleContextRemoveFromSubflow = useCallback(() => {
|
||||||
const blocksToRemove = contextMenuBlocks.filter(
|
const blocksToRemove = contextMenuBlocks.filter(
|
||||||
(block) => block.parentId && (block.parentType === 'loop' || block.parentType === 'parallel')
|
(block) => block.parentId && (block.parentType === 'loop' || block.parentType === 'parallel')
|
||||||
@@ -1947,7 +1909,6 @@ const WorkflowContent = React.memo(() => {
|
|||||||
name: block.name,
|
name: block.name,
|
||||||
isActive,
|
isActive,
|
||||||
isPending,
|
isPending,
|
||||||
groupId: block.data?.groupId,
|
|
||||||
},
|
},
|
||||||
// Include dynamic dimensions for container resizing calculations (must match rendered size)
|
// Include dynamic dimensions for container resizing calculations (must match rendered size)
|
||||||
// Both note and workflow blocks calculate dimensions deterministically via useBlockDimensions
|
// Both note and workflow blocks calculate dimensions deterministically via useBlockDimensions
|
||||||
@@ -2102,56 +2063,16 @@ const WorkflowContent = React.memo(() => {
|
|||||||
window.removeEventListener('remove-from-subflow', handleRemoveFromSubflow as EventListener)
|
window.removeEventListener('remove-from-subflow', handleRemoveFromSubflow as EventListener)
|
||||||
}, [blocks, edgesForDisplay, getNodeAbsolutePosition, collaborativeBatchUpdateParent])
|
}, [blocks, edgesForDisplay, getNodeAbsolutePosition, collaborativeBatchUpdateParent])
|
||||||
|
|
||||||
/** Handles node changes - applies changes and resolves parent-child selection conflicts.
|
/** Handles node changes - applies changes and resolves parent-child selection conflicts. */
|
||||||
* Also expands selection to include all group members when a grouped block is selected.
|
|
||||||
*/
|
|
||||||
const onNodesChange = useCallback(
|
const onNodesChange = useCallback(
|
||||||
(changes: NodeChange[]) => {
|
(changes: NodeChange[]) => {
|
||||||
setDisplayNodes((nds) => {
|
setDisplayNodes((nds) => {
|
||||||
let updated = applyNodeChanges(changes, nds)
|
const updated = applyNodeChanges(changes, nds)
|
||||||
const hasSelectionChange = changes.some((c) => c.type === 'select')
|
const hasSelectionChange = changes.some((c) => c.type === 'select')
|
||||||
|
return hasSelectionChange ? resolveParentChildSelectionConflicts(updated, blocks) : updated
|
||||||
if (hasSelectionChange) {
|
|
||||||
// Expand selection to include all group members
|
|
||||||
const groups = getGroups()
|
|
||||||
const selectedNodeIds = new Set(updated.filter((n) => n.selected).map((n) => n.id))
|
|
||||||
const groupsToInclude = new Set<string>()
|
|
||||||
|
|
||||||
// Find all groups that have at least one selected member
|
|
||||||
selectedNodeIds.forEach((nodeId) => {
|
|
||||||
const groupId = blocks[nodeId]?.data?.groupId
|
|
||||||
if (groupId && groups[groupId]) {
|
|
||||||
groupsToInclude.add(groupId)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
// Add all blocks from those groups to the selection
|
|
||||||
if (groupsToInclude.size > 0) {
|
|
||||||
const expandedNodeIds = new Set(selectedNodeIds)
|
|
||||||
groupsToInclude.forEach((groupId) => {
|
|
||||||
const group = groups[groupId]
|
|
||||||
if (group) {
|
|
||||||
group.blockIds.forEach((blockId) => expandedNodeIds.add(blockId))
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
// Update nodes to include expanded selection
|
|
||||||
if (expandedNodeIds.size > selectedNodeIds.size) {
|
|
||||||
updated = updated.map((n) => ({
|
|
||||||
...n,
|
|
||||||
selected: expandedNodeIds.has(n.id) ? true : n.selected,
|
|
||||||
}))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Resolve parent-child conflicts
|
|
||||||
updated = resolveParentChildSelectionConflicts(updated, blocks)
|
|
||||||
}
|
|
||||||
|
|
||||||
return updated
|
|
||||||
})
|
})
|
||||||
},
|
},
|
||||||
[blocks, getGroups]
|
[blocks]
|
||||||
)
|
)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -2296,7 +2217,8 @@ const WorkflowContent = React.memo(() => {
|
|||||||
)
|
)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Captures the source handle when a connection drag starts
|
* Captures the source handle when a connection drag starts.
|
||||||
|
* Resets connectionCompletedRef to track if onConnect handles this connection.
|
||||||
*/
|
*/
|
||||||
const onConnectStart = useCallback((_event: any, params: any) => {
|
const onConnectStart = useCallback((_event: any, params: any) => {
|
||||||
const handleId: string | undefined = params?.handleId
|
const handleId: string | undefined = params?.handleId
|
||||||
@@ -2305,6 +2227,7 @@ const WorkflowContent = React.memo(() => {
|
|||||||
nodeId: params?.nodeId,
|
nodeId: params?.nodeId,
|
||||||
handleId: params?.handleId,
|
handleId: params?.handleId,
|
||||||
}
|
}
|
||||||
|
connectionCompletedRef.current = false
|
||||||
}, [])
|
}, [])
|
||||||
|
|
||||||
/** Handles new edge connections with container boundary validation. */
|
/** Handles new edge connections with container boundary validation. */
|
||||||
@@ -2365,6 +2288,7 @@ const WorkflowContent = React.memo(() => {
|
|||||||
isInsideContainer: true,
|
isInsideContainer: true,
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
connectionCompletedRef.current = true
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2393,6 +2317,7 @@ const WorkflowContent = React.memo(() => {
|
|||||||
}
|
}
|
||||||
: undefined,
|
: undefined,
|
||||||
})
|
})
|
||||||
|
connectionCompletedRef.current = true
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
[addEdge, getNodes, blocks]
|
[addEdge, getNodes, blocks]
|
||||||
@@ -2401,8 +2326,9 @@ const WorkflowContent = React.memo(() => {
|
|||||||
/**
|
/**
|
||||||
* Handles connection drag end. Detects if the edge was dropped over a block
|
* Handles connection drag end. Detects if the edge was dropped over a block
|
||||||
* and automatically creates a connection to that block's target handle.
|
* and automatically creates a connection to that block's target handle.
|
||||||
* Only creates a connection if ReactFlow didn't already handle it (e.g., when
|
*
|
||||||
* dropping on the block body instead of a handle).
|
* Uses connectionCompletedRef to check if onConnect already handled this connection
|
||||||
|
* (ReactFlow pattern for distinguishing handle-to-handle vs handle-to-body drops).
|
||||||
*/
|
*/
|
||||||
const onConnectEnd = useCallback(
|
const onConnectEnd = useCallback(
|
||||||
(event: MouseEvent | TouchEvent) => {
|
(event: MouseEvent | TouchEvent) => {
|
||||||
@@ -2414,6 +2340,12 @@ const WorkflowContent = React.memo(() => {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// If onConnect already handled this connection, skip (handle-to-handle case)
|
||||||
|
if (connectionCompletedRef.current) {
|
||||||
|
connectionSourceRef.current = null
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
// Get cursor position in flow coordinates
|
// Get cursor position in flow coordinates
|
||||||
const clientPos = 'changedTouches' in event ? event.changedTouches[0] : event
|
const clientPos = 'changedTouches' in event ? event.changedTouches[0] : event
|
||||||
const flowPosition = screenToFlowPosition({
|
const flowPosition = screenToFlowPosition({
|
||||||
@@ -2424,25 +2356,14 @@ const WorkflowContent = React.memo(() => {
|
|||||||
// Find node under cursor
|
// Find node under cursor
|
||||||
const targetNode = findNodeAtPosition(flowPosition)
|
const targetNode = findNodeAtPosition(flowPosition)
|
||||||
|
|
||||||
// Create connection if valid target found AND edge doesn't already exist
|
// Create connection if valid target found (handle-to-body case)
|
||||||
// ReactFlow's onConnect fires first when dropping on a handle, so we check
|
|
||||||
// if that connection already exists to avoid creating duplicates.
|
|
||||||
// IMPORTANT: We must read directly from the store (not React state) because
|
|
||||||
// the store update from ReactFlow's onConnect may not have triggered a
|
|
||||||
// React re-render yet when this callback runs (typically 1-2ms later).
|
|
||||||
if (targetNode && targetNode.id !== source.nodeId) {
|
if (targetNode && targetNode.id !== source.nodeId) {
|
||||||
const currentEdges = useWorkflowStore.getState().edges
|
onConnect({
|
||||||
const edgeAlreadyExists = currentEdges.some(
|
source: source.nodeId,
|
||||||
(e) => e.source === source.nodeId && e.target === targetNode.id
|
sourceHandle: source.handleId,
|
||||||
)
|
target: targetNode.id,
|
||||||
if (!edgeAlreadyExists) {
|
targetHandle: 'target',
|
||||||
onConnect({
|
})
|
||||||
source: source.nodeId,
|
|
||||||
sourceHandle: source.handleId,
|
|
||||||
target: targetNode.id,
|
|
||||||
targetHandle: 'target',
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
connectionSourceRef.current = null
|
connectionSourceRef.current = null
|
||||||
@@ -2612,55 +2533,9 @@ const WorkflowContent = React.memo(() => {
|
|||||||
parentId: currentParentId,
|
parentId: currentParentId,
|
||||||
})
|
})
|
||||||
|
|
||||||
// Expand selection to include all group members before capturing positions
|
|
||||||
const groups = getGroups()
|
|
||||||
const allNodes = getNodes()
|
|
||||||
|
|
||||||
// Find the group of the dragged node
|
|
||||||
const draggedBlockGroupId = blocks[node.id]?.data?.groupId
|
|
||||||
|
|
||||||
// If the dragged node is in a group, expand selection to include all group members
|
|
||||||
if (draggedBlockGroupId && groups[draggedBlockGroupId]) {
|
|
||||||
const group = groups[draggedBlockGroupId]
|
|
||||||
const groupBlockIds = new Set(group.blockIds)
|
|
||||||
|
|
||||||
// Check if we need to expand selection
|
|
||||||
const currentSelectedIds = new Set(allNodes.filter((n) => n.selected).map((n) => n.id))
|
|
||||||
const needsExpansion = [...groupBlockIds].some((id) => !currentSelectedIds.has(id))
|
|
||||||
|
|
||||||
if (needsExpansion) {
|
|
||||||
setNodes((nodes) =>
|
|
||||||
nodes.map((n) => {
|
|
||||||
const isInGroup = groupBlockIds.has(n.id)
|
|
||||||
const isDirectlyDragged = n.id === node.id
|
|
||||||
return {
|
|
||||||
...n,
|
|
||||||
selected: isInGroup ? true : n.selected,
|
|
||||||
data: {
|
|
||||||
...n.data,
|
|
||||||
// Mark as grouped selection if in group but not the directly dragged node
|
|
||||||
isGroupedSelection: isInGroup && !isDirectlyDragged && !n.selected,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
})
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Capture all selected nodes' positions for multi-node undo/redo
|
// Capture all selected nodes' positions for multi-node undo/redo
|
||||||
// Re-get nodes after potential selection expansion
|
const allNodes = getNodes()
|
||||||
const updatedNodes = getNodes()
|
const selectedNodes = allNodes.filter((n) => n.selected)
|
||||||
const selectedNodes = updatedNodes.filter((n) => {
|
|
||||||
// Always include the dragged node
|
|
||||||
if (n.id === node.id) return true
|
|
||||||
// Include node if it's selected OR if it's in the same group as the dragged node
|
|
||||||
if (n.selected) return true
|
|
||||||
if (draggedBlockGroupId && groups[draggedBlockGroupId]) {
|
|
||||||
return groups[draggedBlockGroupId].blockIds.includes(n.id)
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
})
|
|
||||||
|
|
||||||
multiNodeDragStartRef.current.clear()
|
multiNodeDragStartRef.current.clear()
|
||||||
selectedNodes.forEach((n) => {
|
selectedNodes.forEach((n) => {
|
||||||
const block = blocks[n.id]
|
const block = blocks[n.id]
|
||||||
@@ -2672,63 +2547,8 @@ const WorkflowContent = React.memo(() => {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
// Set up temporary parent-child relationships for group members
|
|
||||||
// This leverages React Flow's built-in parent-child drag behavior
|
|
||||||
// BUT: Only do this if NOT all group members are already selected
|
|
||||||
// If all are selected, React Flow's native multiselect drag will handle it
|
|
||||||
groupDragTempParentsRef.current.clear()
|
|
||||||
if (draggedBlockGroupId && groups[draggedBlockGroupId]) {
|
|
||||||
const group = groups[draggedBlockGroupId]
|
|
||||||
if (group.blockIds.length > 1) {
|
|
||||||
// Check if all group members are already selected
|
|
||||||
const allGroupMembersSelected = group.blockIds.every((blockId) =>
|
|
||||||
updatedNodes.find((n) => n.id === blockId && n.selected)
|
|
||||||
)
|
|
||||||
|
|
||||||
// Only use temporary parent approach if NOT all members are selected
|
|
||||||
// (i.e., when click-and-dragging on an unselected grouped block)
|
|
||||||
if (!allGroupMembersSelected) {
|
|
||||||
// Get the dragged node's absolute position for calculating relative positions
|
|
||||||
const draggedNodeAbsPos = getNodeAbsolutePosition(node.id)
|
|
||||||
|
|
||||||
setNodes((nodes) =>
|
|
||||||
nodes.map((n) => {
|
|
||||||
// Skip the dragged node - it becomes the temporary parent
|
|
||||||
if (n.id === node.id) return n
|
|
||||||
|
|
||||||
// Only process nodes in the same group
|
|
||||||
if (group.blockIds.includes(n.id)) {
|
|
||||||
// Store original position and parentId for restoration later
|
|
||||||
groupDragTempParentsRef.current.set(n.id, {
|
|
||||||
originalPosition: { ...n.position },
|
|
||||||
originalParentId: n.parentId,
|
|
||||||
})
|
|
||||||
|
|
||||||
// Get this node's absolute position
|
|
||||||
const nodeAbsPos = getNodeAbsolutePosition(n.id)
|
|
||||||
|
|
||||||
// Calculate position relative to the dragged node
|
|
||||||
const relativePosition = {
|
|
||||||
x: nodeAbsPos.x - draggedNodeAbsPos.x,
|
|
||||||
y: nodeAbsPos.y - draggedNodeAbsPos.y,
|
|
||||||
}
|
|
||||||
|
|
||||||
return {
|
|
||||||
...n,
|
|
||||||
parentId: node.id, // Temporarily make this a child of the dragged node
|
|
||||||
position: relativePosition,
|
|
||||||
extent: undefined, // Remove extent constraint during drag
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return n
|
|
||||||
})
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
[blocks, setDragStartPosition, getNodes, setNodes, getGroups, potentialParentId, setPotentialParentId, getNodeAbsolutePosition]
|
[blocks, setDragStartPosition, getNodes, potentialParentId, setPotentialParentId]
|
||||||
)
|
)
|
||||||
|
|
||||||
/** Handles node drag stop to establish parent-child relationships. */
|
/** Handles node drag stop to establish parent-child relationships. */
|
||||||
@@ -2736,93 +2556,13 @@ const WorkflowContent = React.memo(() => {
|
|||||||
(_event: React.MouseEvent, node: any) => {
|
(_event: React.MouseEvent, node: any) => {
|
||||||
clearDragHighlights()
|
clearDragHighlights()
|
||||||
|
|
||||||
// Compute absolute positions for group members before restoring parentIds
|
|
||||||
// We need to do this first because getNodes() will return stale data after setNodes
|
|
||||||
const computedGroupPositions = new Map<string, { x: number; y: number }>()
|
|
||||||
const draggedBlockGroupId = blocks[node.id]?.data?.groupId
|
|
||||||
|
|
||||||
if (groupDragTempParentsRef.current.size > 0) {
|
|
||||||
const draggedNodeAbsPos = getNodeAbsolutePosition(node.id)
|
|
||||||
const currentNodes = getNodes()
|
|
||||||
|
|
||||||
// Compute absolute positions for all temporarily parented nodes
|
|
||||||
for (const [nodeId, _tempData] of groupDragTempParentsRef.current) {
|
|
||||||
const nodeData = currentNodes.find((n) => n.id === nodeId)
|
|
||||||
if (nodeData) {
|
|
||||||
// The node's current position is relative to the dragged node
|
|
||||||
computedGroupPositions.set(nodeId, {
|
|
||||||
x: draggedNodeAbsPos.x + nodeData.position.x,
|
|
||||||
y: draggedNodeAbsPos.y + nodeData.position.y,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Also store the dragged node's absolute position
|
|
||||||
computedGroupPositions.set(node.id, draggedNodeAbsPos)
|
|
||||||
|
|
||||||
// Restore temporary parent-child relationships
|
|
||||||
setNodes((nodes) =>
|
|
||||||
nodes.map((n) => {
|
|
||||||
const tempData = groupDragTempParentsRef.current.get(n.id)
|
|
||||||
if (tempData) {
|
|
||||||
const absolutePosition = computedGroupPositions.get(n.id) || n.position
|
|
||||||
|
|
||||||
return {
|
|
||||||
...n,
|
|
||||||
parentId: tempData.originalParentId,
|
|
||||||
position: absolutePosition,
|
|
||||||
extent: tempData.originalParentId ? ('parent' as const) : undefined,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return n
|
|
||||||
})
|
|
||||||
)
|
|
||||||
groupDragTempParentsRef.current.clear()
|
|
||||||
}
|
|
||||||
|
|
||||||
// Get all selected nodes to update their positions too
|
// Get all selected nodes to update their positions too
|
||||||
const allNodes = getNodes()
|
const allNodes = getNodes()
|
||||||
let selectedNodes = allNodes.filter((n) => n.selected)
|
const selectedNodes = allNodes.filter((n) => n.selected)
|
||||||
|
|
||||||
// If the dragged node is in a group, include all group members
|
// If multiple nodes are selected, update all their positions
|
||||||
if (draggedBlockGroupId) {
|
|
||||||
const groups = getGroups()
|
|
||||||
const group = groups[draggedBlockGroupId]
|
|
||||||
if (group && group.blockIds.length > 1) {
|
|
||||||
const groupBlockIds = new Set(group.blockIds)
|
|
||||||
// Include the dragged node and all group members that aren't already selected
|
|
||||||
const groupNodes = allNodes.filter(
|
|
||||||
(n) => groupBlockIds.has(n.id) && !selectedNodes.some((sn) => sn.id === n.id)
|
|
||||||
)
|
|
||||||
selectedNodes = [...selectedNodes, ...groupNodes]
|
|
||||||
// Also ensure the dragged node is included
|
|
||||||
if (!selectedNodes.some((n) => n.id === node.id)) {
|
|
||||||
const draggedNode = allNodes.find((n) => n.id === node.id)
|
|
||||||
if (draggedNode) {
|
|
||||||
selectedNodes = [...selectedNodes, draggedNode]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// If multiple nodes are selected (or in a group), update all their positions
|
|
||||||
if (selectedNodes.length > 1) {
|
if (selectedNodes.length > 1) {
|
||||||
// Use pre-computed positions for group members, otherwise use computeClampedPositionUpdates
|
const positionUpdates = computeClampedPositionUpdates(selectedNodes, blocks, allNodes)
|
||||||
let positionUpdates: Array<{ id: string; position: { x: number; y: number } }>
|
|
||||||
|
|
||||||
if (computedGroupPositions.size > 0) {
|
|
||||||
// For group drags, use the pre-computed absolute positions
|
|
||||||
positionUpdates = selectedNodes.map((n) => {
|
|
||||||
const precomputedPos = computedGroupPositions.get(n.id)
|
|
||||||
if (precomputedPos) {
|
|
||||||
return { id: n.id, position: precomputedPos }
|
|
||||||
}
|
|
||||||
// For non-group members, use current position
|
|
||||||
return { id: n.id, position: n.position }
|
|
||||||
})
|
|
||||||
} else {
|
|
||||||
positionUpdates = computeClampedPositionUpdates(selectedNodes, blocks, allNodes)
|
|
||||||
}
|
|
||||||
collaborativeBatchUpdatePositions(positionUpdates, {
|
collaborativeBatchUpdatePositions(positionUpdates, {
|
||||||
previousPositions: multiNodeDragStartRef.current,
|
previousPositions: multiNodeDragStartRef.current,
|
||||||
})
|
})
|
||||||
@@ -3106,7 +2846,6 @@ const WorkflowContent = React.memo(() => {
|
|||||||
},
|
},
|
||||||
[
|
[
|
||||||
getNodes,
|
getNodes,
|
||||||
setNodes,
|
|
||||||
dragStartParentId,
|
dragStartParentId,
|
||||||
potentialParentId,
|
potentialParentId,
|
||||||
updateNodeParent,
|
updateNodeParent,
|
||||||
@@ -3125,7 +2864,6 @@ const WorkflowContent = React.memo(() => {
|
|||||||
activeWorkflowId,
|
activeWorkflowId,
|
||||||
collaborativeBatchUpdatePositions,
|
collaborativeBatchUpdatePositions,
|
||||||
collaborativeBatchUpdateParent,
|
collaborativeBatchUpdateParent,
|
||||||
getGroups,
|
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -3433,81 +3171,19 @@ const WorkflowContent = React.memo(() => {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Handles node click to select the node in ReactFlow.
|
* Handles node click to select the node in ReactFlow.
|
||||||
* When clicking on a grouped block, also selects all other blocks in the group.
|
|
||||||
* Grouped blocks are marked with isGroupedSelection for different visual styling.
|
|
||||||
* Parent-child conflict resolution happens automatically in onNodesChange.
|
* Parent-child conflict resolution happens automatically in onNodesChange.
|
||||||
*/
|
*/
|
||||||
const handleNodeClick = useCallback(
|
const handleNodeClick = useCallback(
|
||||||
(event: React.MouseEvent, node: Node) => {
|
(event: React.MouseEvent, node: Node) => {
|
||||||
const isMultiSelect = event.shiftKey || event.metaKey || event.ctrlKey
|
const isMultiSelect = event.shiftKey || event.metaKey || event.ctrlKey
|
||||||
const groups = getGroups()
|
setNodes((nodes) =>
|
||||||
|
nodes.map((n) => ({
|
||||||
// Track which nodes are directly clicked vs. group-expanded
|
...n,
|
||||||
const directlySelectedIds = new Set<string>()
|
selected: isMultiSelect ? (n.id === node.id ? true : n.selected) : n.id === node.id,
|
||||||
|
}))
|
||||||
setNodes((nodes) => {
|
)
|
||||||
// First, calculate the base selection
|
|
||||||
let updatedNodes = nodes.map((n) => {
|
|
||||||
const isDirectlySelected = isMultiSelect
|
|
||||||
? n.id === node.id
|
|
||||||
? true
|
|
||||||
: n.selected
|
|
||||||
: n.id === node.id
|
|
||||||
if (isDirectlySelected) {
|
|
||||||
directlySelectedIds.add(n.id)
|
|
||||||
}
|
|
||||||
return {
|
|
||||||
...n,
|
|
||||||
selected: isDirectlySelected,
|
|
||||||
data: {
|
|
||||||
...n.data,
|
|
||||||
isGroupedSelection: false, // Reset grouped selection flag
|
|
||||||
},
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
// Expand selection to include all group members
|
|
||||||
const selectedNodeIds = new Set(updatedNodes.filter((n) => n.selected).map((n) => n.id))
|
|
||||||
const groupsToInclude = new Set<string>()
|
|
||||||
|
|
||||||
// Find all groups that have at least one selected member
|
|
||||||
selectedNodeIds.forEach((nodeId) => {
|
|
||||||
const groupId = blocks[nodeId]?.data?.groupId
|
|
||||||
if (groupId && groups[groupId]) {
|
|
||||||
groupsToInclude.add(groupId)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
// Add all blocks from those groups to the selection
|
|
||||||
if (groupsToInclude.size > 0) {
|
|
||||||
const expandedNodeIds = new Set(selectedNodeIds)
|
|
||||||
groupsToInclude.forEach((groupId) => {
|
|
||||||
const group = groups[groupId]
|
|
||||||
if (group) {
|
|
||||||
group.blockIds.forEach((blockId) => expandedNodeIds.add(blockId))
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
// Update nodes with expanded selection, marking group-expanded nodes
|
|
||||||
if (expandedNodeIds.size > selectedNodeIds.size) {
|
|
||||||
updatedNodes = updatedNodes.map((n) => {
|
|
||||||
const isGroupExpanded = expandedNodeIds.has(n.id) && !directlySelectedIds.has(n.id)
|
|
||||||
return {
|
|
||||||
...n,
|
|
||||||
selected: expandedNodeIds.has(n.id) ? true : n.selected,
|
|
||||||
data: {
|
|
||||||
...n.data,
|
|
||||||
isGroupedSelection: isGroupExpanded,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return updatedNodes
|
|
||||||
})
|
|
||||||
},
|
},
|
||||||
[setNodes, blocks, getGroups]
|
[setNodes]
|
||||||
)
|
)
|
||||||
|
|
||||||
/** Handles edge selection with container context tracking and Shift-click multi-selection. */
|
/** Handles edge selection with container context tracking and Shift-click multi-selection. */
|
||||||
@@ -3742,8 +3418,6 @@ const WorkflowContent = React.memo(() => {
|
|||||||
onRemoveFromSubflow={handleContextRemoveFromSubflow}
|
onRemoveFromSubflow={handleContextRemoveFromSubflow}
|
||||||
onOpenEditor={handleContextOpenEditor}
|
onOpenEditor={handleContextOpenEditor}
|
||||||
onRename={handleContextRename}
|
onRename={handleContextRename}
|
||||||
onGroupBlocks={handleContextGroupBlocks}
|
|
||||||
onUngroupBlocks={handleContextUngroupBlocks}
|
|
||||||
hasClipboard={hasClipboard()}
|
hasClipboard={hasClipboard()}
|
||||||
showRemoveFromSubflow={contextMenuBlocks.some(
|
showRemoveFromSubflow={contextMenuBlocks.some(
|
||||||
(b) => b.parentId && (b.parentType === 'loop' || b.parentType === 'parallel')
|
(b) => b.parentId && (b.parentType === 'loop' || b.parentType === 'parallel')
|
||||||
|
|||||||
@@ -98,6 +98,23 @@ export const A2ABlock: BlockConfig<A2AResponse> = {
|
|||||||
condition: { field: 'operation', value: 'a2a_send_message' },
|
condition: { field: 'operation', value: 'a2a_send_message' },
|
||||||
required: true,
|
required: true,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
id: 'data',
|
||||||
|
title: 'Data (JSON)',
|
||||||
|
type: 'code',
|
||||||
|
placeholder: '{\n "key": "value"\n}',
|
||||||
|
description: 'Structured data to include with the message (DataPart)',
|
||||||
|
condition: { field: 'operation', value: 'a2a_send_message' },
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'files',
|
||||||
|
title: 'Files',
|
||||||
|
type: 'file-upload',
|
||||||
|
placeholder: 'Upload files to send',
|
||||||
|
description: 'Files to include with the message (FilePart)',
|
||||||
|
condition: { field: 'operation', value: 'a2a_send_message' },
|
||||||
|
multiple: true,
|
||||||
|
},
|
||||||
{
|
{
|
||||||
id: 'taskId',
|
id: 'taskId',
|
||||||
title: 'Task ID',
|
title: 'Task ID',
|
||||||
@@ -208,6 +225,14 @@ export const A2ABlock: BlockConfig<A2AResponse> = {
|
|||||||
type: 'string',
|
type: 'string',
|
||||||
description: 'Context ID for conversation continuity',
|
description: 'Context ID for conversation continuity',
|
||||||
},
|
},
|
||||||
|
data: {
|
||||||
|
type: 'json',
|
||||||
|
description: 'Structured data to include with the message',
|
||||||
|
},
|
||||||
|
files: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Files to include with the message',
|
||||||
|
},
|
||||||
historyLength: {
|
historyLength: {
|
||||||
type: 'number',
|
type: 'number',
|
||||||
description: 'Number of history messages to include',
|
description: 'Number of history messages to include',
|
||||||
|
|||||||
@@ -26,6 +26,8 @@ export const SlackBlock: BlockConfig<SlackResponse> = {
|
|||||||
{ label: 'Send Message', id: 'send' },
|
{ label: 'Send Message', id: 'send' },
|
||||||
{ label: 'Create Canvas', id: 'canvas' },
|
{ label: 'Create Canvas', id: 'canvas' },
|
||||||
{ label: 'Read Messages', id: 'read' },
|
{ label: 'Read Messages', id: 'read' },
|
||||||
|
{ label: 'Get Message', id: 'get_message' },
|
||||||
|
{ label: 'Get Thread', id: 'get_thread' },
|
||||||
{ label: 'List Channels', id: 'list_channels' },
|
{ label: 'List Channels', id: 'list_channels' },
|
||||||
{ label: 'List Channel Members', id: 'list_members' },
|
{ label: 'List Channel Members', id: 'list_members' },
|
||||||
{ label: 'List Users', id: 'list_users' },
|
{ label: 'List Users', id: 'list_users' },
|
||||||
@@ -316,6 +318,68 @@ export const SlackBlock: BlockConfig<SlackResponse> = {
|
|||||||
},
|
},
|
||||||
required: true,
|
required: true,
|
||||||
},
|
},
|
||||||
|
// Get Message specific fields
|
||||||
|
{
|
||||||
|
id: 'getMessageTimestamp',
|
||||||
|
title: 'Message Timestamp',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Message timestamp (e.g., 1405894322.002768)',
|
||||||
|
condition: {
|
||||||
|
field: 'operation',
|
||||||
|
value: 'get_message',
|
||||||
|
},
|
||||||
|
required: true,
|
||||||
|
wandConfig: {
|
||||||
|
enabled: true,
|
||||||
|
prompt: `Extract or generate a Slack message timestamp from the user's input.
|
||||||
|
Slack message timestamps are in the format: XXXXXXXXXX.XXXXXX (seconds.microseconds since Unix epoch).
|
||||||
|
Examples:
|
||||||
|
- "1405894322.002768" -> 1405894322.002768 (already a valid timestamp)
|
||||||
|
- "thread_ts from the trigger" -> The user wants to reference a variable, output the original text
|
||||||
|
- A URL like "https://slack.com/archives/C123/p1405894322002768" -> Extract 1405894322.002768 (remove 'p' prefix, add decimal after 10th digit)
|
||||||
|
|
||||||
|
If the input looks like a reference to another block's output (contains < and >) or a variable, return it as-is.
|
||||||
|
Return ONLY the timestamp string - no explanations, no quotes, no extra text.`,
|
||||||
|
placeholder: 'Paste a Slack message URL or timestamp...',
|
||||||
|
generationType: 'timestamp',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
// Get Thread specific fields
|
||||||
|
{
|
||||||
|
id: 'getThreadTimestamp',
|
||||||
|
title: 'Thread Timestamp',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Thread timestamp (thread_ts, e.g., 1405894322.002768)',
|
||||||
|
condition: {
|
||||||
|
field: 'operation',
|
||||||
|
value: 'get_thread',
|
||||||
|
},
|
||||||
|
required: true,
|
||||||
|
wandConfig: {
|
||||||
|
enabled: true,
|
||||||
|
prompt: `Extract or generate a Slack thread timestamp from the user's input.
|
||||||
|
Slack thread timestamps (thread_ts) are in the format: XXXXXXXXXX.XXXXXX (seconds.microseconds since Unix epoch).
|
||||||
|
Examples:
|
||||||
|
- "1405894322.002768" -> 1405894322.002768 (already a valid timestamp)
|
||||||
|
- "thread_ts from the trigger" -> The user wants to reference a variable, output the original text
|
||||||
|
- A URL like "https://slack.com/archives/C123/p1405894322002768" -> Extract 1405894322.002768 (remove 'p' prefix, add decimal after 10th digit)
|
||||||
|
|
||||||
|
If the input looks like a reference to another block's output (contains < and >) or a variable, return it as-is.
|
||||||
|
Return ONLY the timestamp string - no explanations, no quotes, no extra text.`,
|
||||||
|
placeholder: 'Paste a Slack thread URL or thread_ts...',
|
||||||
|
generationType: 'timestamp',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'threadLimit',
|
||||||
|
title: 'Message Limit',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: '100',
|
||||||
|
condition: {
|
||||||
|
field: 'operation',
|
||||||
|
value: 'get_thread',
|
||||||
|
},
|
||||||
|
},
|
||||||
{
|
{
|
||||||
id: 'oldest',
|
id: 'oldest',
|
||||||
title: 'Oldest Timestamp',
|
title: 'Oldest Timestamp',
|
||||||
@@ -430,6 +494,8 @@ Return ONLY the timestamp string - no explanations, no quotes, no extra text.`,
|
|||||||
'slack_message',
|
'slack_message',
|
||||||
'slack_canvas',
|
'slack_canvas',
|
||||||
'slack_message_reader',
|
'slack_message_reader',
|
||||||
|
'slack_get_message',
|
||||||
|
'slack_get_thread',
|
||||||
'slack_list_channels',
|
'slack_list_channels',
|
||||||
'slack_list_members',
|
'slack_list_members',
|
||||||
'slack_list_users',
|
'slack_list_users',
|
||||||
@@ -448,6 +514,10 @@ Return ONLY the timestamp string - no explanations, no quotes, no extra text.`,
|
|||||||
return 'slack_canvas'
|
return 'slack_canvas'
|
||||||
case 'read':
|
case 'read':
|
||||||
return 'slack_message_reader'
|
return 'slack_message_reader'
|
||||||
|
case 'get_message':
|
||||||
|
return 'slack_get_message'
|
||||||
|
case 'get_thread':
|
||||||
|
return 'slack_get_thread'
|
||||||
case 'list_channels':
|
case 'list_channels':
|
||||||
return 'slack_list_channels'
|
return 'slack_list_channels'
|
||||||
case 'list_members':
|
case 'list_members':
|
||||||
@@ -498,6 +568,9 @@ Return ONLY the timestamp string - no explanations, no quotes, no extra text.`,
|
|||||||
includeDeleted,
|
includeDeleted,
|
||||||
userLimit,
|
userLimit,
|
||||||
userId,
|
userId,
|
||||||
|
getMessageTimestamp,
|
||||||
|
getThreadTimestamp,
|
||||||
|
threadLimit,
|
||||||
...rest
|
...rest
|
||||||
} = params
|
} = params
|
||||||
|
|
||||||
@@ -574,6 +647,27 @@ Return ONLY the timestamp string - no explanations, no quotes, no extra text.`,
|
|||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
|
case 'get_message':
|
||||||
|
if (!getMessageTimestamp) {
|
||||||
|
throw new Error('Message timestamp is required for get message operation')
|
||||||
|
}
|
||||||
|
baseParams.timestamp = getMessageTimestamp
|
||||||
|
break
|
||||||
|
|
||||||
|
case 'get_thread': {
|
||||||
|
if (!getThreadTimestamp) {
|
||||||
|
throw new Error('Thread timestamp is required for get thread operation')
|
||||||
|
}
|
||||||
|
baseParams.threadTs = getThreadTimestamp
|
||||||
|
if (threadLimit) {
|
||||||
|
const parsedLimit = Number.parseInt(threadLimit, 10)
|
||||||
|
if (!Number.isNaN(parsedLimit) && parsedLimit > 0) {
|
||||||
|
baseParams.limit = Math.min(parsedLimit, 200)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
case 'list_channels': {
|
case 'list_channels': {
|
||||||
baseParams.includePrivate = includePrivate !== 'false'
|
baseParams.includePrivate = includePrivate !== 'false'
|
||||||
baseParams.excludeArchived = true
|
baseParams.excludeArchived = true
|
||||||
@@ -679,6 +773,14 @@ Return ONLY the timestamp string - no explanations, no quotes, no extra text.`,
|
|||||||
userLimit: { type: 'string', description: 'Maximum number of users to return' },
|
userLimit: { type: 'string', description: 'Maximum number of users to return' },
|
||||||
// Get User inputs
|
// Get User inputs
|
||||||
userId: { type: 'string', description: 'User ID to look up' },
|
userId: { type: 'string', description: 'User ID to look up' },
|
||||||
|
// Get Message inputs
|
||||||
|
getMessageTimestamp: { type: 'string', description: 'Message timestamp to retrieve' },
|
||||||
|
// Get Thread inputs
|
||||||
|
getThreadTimestamp: { type: 'string', description: 'Thread timestamp to retrieve' },
|
||||||
|
threadLimit: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'Maximum number of messages to return from thread',
|
||||||
|
},
|
||||||
},
|
},
|
||||||
outputs: {
|
outputs: {
|
||||||
// slack_message outputs (send operation)
|
// slack_message outputs (send operation)
|
||||||
@@ -706,6 +808,24 @@ Return ONLY the timestamp string - no explanations, no quotes, no extra text.`,
|
|||||||
'Array of message objects with comprehensive properties: text, user, timestamp, reactions, threads, files, attachments, blocks, stars, pins, and edit history',
|
'Array of message objects with comprehensive properties: text, user, timestamp, reactions, threads, files, attachments, blocks, stars, pins, and edit history',
|
||||||
},
|
},
|
||||||
|
|
||||||
|
// slack_get_thread outputs (get_thread operation)
|
||||||
|
parentMessage: {
|
||||||
|
type: 'json',
|
||||||
|
description: 'The thread parent message with all properties',
|
||||||
|
},
|
||||||
|
replies: {
|
||||||
|
type: 'json',
|
||||||
|
description: 'Array of reply messages in the thread (excluding the parent)',
|
||||||
|
},
|
||||||
|
replyCount: {
|
||||||
|
type: 'number',
|
||||||
|
description: 'Number of replies returned in this response',
|
||||||
|
},
|
||||||
|
hasMore: {
|
||||||
|
type: 'boolean',
|
||||||
|
description: 'Whether there are more messages in the thread',
|
||||||
|
},
|
||||||
|
|
||||||
// slack_list_channels outputs (list_channels operation)
|
// slack_list_channels outputs (list_channels operation)
|
||||||
channels: {
|
channels: {
|
||||||
type: 'json',
|
type: 'json',
|
||||||
|
|||||||
207
apps/sim/blocks/blocks/tinybird.ts
Normal file
207
apps/sim/blocks/blocks/tinybird.ts
Normal file
@@ -0,0 +1,207 @@
|
|||||||
|
import { TinybirdIcon } from '@/components/icons'
|
||||||
|
import type { BlockConfig } from '@/blocks/types'
|
||||||
|
import { AuthMode } from '@/blocks/types'
|
||||||
|
import type { TinybirdResponse } from '@/tools/tinybird/types'
|
||||||
|
|
||||||
|
export const TinybirdBlock: BlockConfig<TinybirdResponse> = {
|
||||||
|
type: 'tinybird',
|
||||||
|
name: 'Tinybird',
|
||||||
|
description: 'Send events and query data with Tinybird',
|
||||||
|
authMode: AuthMode.ApiKey,
|
||||||
|
longDescription:
|
||||||
|
'Interact with Tinybird using the Events API to stream JSON or NDJSON events, or use the Query API to execute SQL queries against Pipes and Data Sources.',
|
||||||
|
docsLink: 'https://www.tinybird.co/docs/api-reference',
|
||||||
|
category: 'tools',
|
||||||
|
bgColor: '#2EF598',
|
||||||
|
icon: TinybirdIcon,
|
||||||
|
subBlocks: [
|
||||||
|
{
|
||||||
|
id: 'operation',
|
||||||
|
title: 'Operation',
|
||||||
|
type: 'dropdown',
|
||||||
|
options: [
|
||||||
|
{ label: 'Send Events', id: 'tinybird_events' },
|
||||||
|
{ label: 'Query', id: 'tinybird_query' },
|
||||||
|
],
|
||||||
|
value: () => 'tinybird_events',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'base_url',
|
||||||
|
title: 'Base URL',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'https://api.tinybird.co',
|
||||||
|
required: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'token',
|
||||||
|
title: 'API Token',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'Enter your Tinybird API token',
|
||||||
|
password: true,
|
||||||
|
required: true,
|
||||||
|
},
|
||||||
|
// Send Events operation inputs
|
||||||
|
{
|
||||||
|
id: 'datasource',
|
||||||
|
title: 'Data Source',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'my_events_datasource',
|
||||||
|
condition: { field: 'operation', value: 'tinybird_events' },
|
||||||
|
required: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'data',
|
||||||
|
title: 'Data',
|
||||||
|
type: 'code',
|
||||||
|
placeholder:
|
||||||
|
'{"event": "click", "timestamp": "2024-01-01T12:00:00Z"}\n{"event": "view", "timestamp": "2024-01-01T12:00:01Z"}',
|
||||||
|
condition: { field: 'operation', value: 'tinybird_events' },
|
||||||
|
required: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'format',
|
||||||
|
title: 'Format',
|
||||||
|
type: 'dropdown',
|
||||||
|
options: [
|
||||||
|
{ label: 'NDJSON (Newline-delimited JSON)', id: 'ndjson' },
|
||||||
|
{ label: 'JSON', id: 'json' },
|
||||||
|
],
|
||||||
|
value: () => 'ndjson',
|
||||||
|
condition: { field: 'operation', value: 'tinybird_events' },
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'compression',
|
||||||
|
title: 'Compression',
|
||||||
|
type: 'dropdown',
|
||||||
|
options: [
|
||||||
|
{ label: 'None', id: 'none' },
|
||||||
|
{ label: 'Gzip', id: 'gzip' },
|
||||||
|
],
|
||||||
|
value: () => 'none',
|
||||||
|
mode: 'advanced',
|
||||||
|
condition: { field: 'operation', value: 'tinybird_events' },
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'wait',
|
||||||
|
title: 'Wait for Acknowledgment',
|
||||||
|
type: 'switch',
|
||||||
|
value: () => 'false',
|
||||||
|
mode: 'advanced',
|
||||||
|
condition: { field: 'operation', value: 'tinybird_events' },
|
||||||
|
},
|
||||||
|
// Query operation inputs
|
||||||
|
{
|
||||||
|
id: 'query',
|
||||||
|
title: 'SQL Query',
|
||||||
|
type: 'code',
|
||||||
|
placeholder: 'SELECT * FROM my_pipe FORMAT JSON\nOR\nSELECT * FROM my_pipe FORMAT CSV',
|
||||||
|
condition: { field: 'operation', value: 'tinybird_query' },
|
||||||
|
required: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 'pipeline',
|
||||||
|
title: 'Pipeline Name',
|
||||||
|
type: 'short-input',
|
||||||
|
placeholder: 'my_pipe (optional)',
|
||||||
|
condition: { field: 'operation', value: 'tinybird_query' },
|
||||||
|
},
|
||||||
|
],
|
||||||
|
tools: {
|
||||||
|
access: ['tinybird_events', 'tinybird_query'],
|
||||||
|
config: {
|
||||||
|
tool: (params) => params.operation || 'tinybird_events',
|
||||||
|
params: (params) => {
|
||||||
|
const operation = params.operation || 'tinybird_events'
|
||||||
|
const result: Record<string, any> = {
|
||||||
|
base_url: params.base_url,
|
||||||
|
token: params.token,
|
||||||
|
}
|
||||||
|
|
||||||
|
if (operation === 'tinybird_events') {
|
||||||
|
// Send Events operation
|
||||||
|
if (!params.datasource) {
|
||||||
|
throw new Error('Data Source is required for Send Events operation')
|
||||||
|
}
|
||||||
|
if (!params.data) {
|
||||||
|
throw new Error('Data is required for Send Events operation')
|
||||||
|
}
|
||||||
|
|
||||||
|
result.datasource = params.datasource
|
||||||
|
result.data = params.data
|
||||||
|
result.format = params.format || 'ndjson'
|
||||||
|
result.compression = params.compression || 'none'
|
||||||
|
|
||||||
|
// Convert wait from string to boolean
|
||||||
|
// Convert wait from string to boolean
|
||||||
|
if (params.wait !== undefined) {
|
||||||
|
const waitValue =
|
||||||
|
typeof params.wait === 'string' ? params.wait.toLowerCase() : params.wait
|
||||||
|
result.wait = waitValue === 'true' || waitValue === true
|
||||||
|
}
|
||||||
|
} else if (operation === 'tinybird_query') {
|
||||||
|
// Query operation
|
||||||
|
if (!params.query) {
|
||||||
|
throw new Error('SQL Query is required for Query operation')
|
||||||
|
}
|
||||||
|
|
||||||
|
result.query = params.query
|
||||||
|
if (params.pipeline) {
|
||||||
|
result.pipeline = params.pipeline
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return result
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
inputs: {
|
||||||
|
operation: { type: 'string', description: 'Operation to perform' },
|
||||||
|
base_url: { type: 'string', description: 'Tinybird API base URL' },
|
||||||
|
// Send Events inputs
|
||||||
|
datasource: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'Name of the Tinybird Data Source',
|
||||||
|
},
|
||||||
|
data: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'Data to send as JSON or NDJSON string',
|
||||||
|
},
|
||||||
|
wait: { type: 'boolean', description: 'Wait for database acknowledgment' },
|
||||||
|
format: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'Format of the events (ndjson or json)',
|
||||||
|
},
|
||||||
|
compression: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'Compression format (none or gzip)',
|
||||||
|
},
|
||||||
|
// Query inputs
|
||||||
|
query: { type: 'string', description: 'SQL query to execute' },
|
||||||
|
pipeline: { type: 'string', description: 'Optional pipeline name' },
|
||||||
|
// Common
|
||||||
|
token: { type: 'string', description: 'Tinybird API Token' },
|
||||||
|
},
|
||||||
|
outputs: {
|
||||||
|
// Send Events outputs
|
||||||
|
successful_rows: {
|
||||||
|
type: 'number',
|
||||||
|
description: 'Number of rows successfully ingested',
|
||||||
|
},
|
||||||
|
quarantined_rows: {
|
||||||
|
type: 'number',
|
||||||
|
description: 'Number of rows quarantined (failed validation)',
|
||||||
|
},
|
||||||
|
// Query outputs
|
||||||
|
data: {
|
||||||
|
type: 'json',
|
||||||
|
description:
|
||||||
|
'Query result data. FORMAT JSON: array of objects. Other formats (CSV, TSV, etc.): raw text string.',
|
||||||
|
},
|
||||||
|
rows: { type: 'number', description: 'Number of rows returned (only with FORMAT JSON)' },
|
||||||
|
statistics: {
|
||||||
|
type: 'json',
|
||||||
|
description:
|
||||||
|
'Query execution statistics - elapsed time, rows read, bytes read (only with FORMAT JSON)',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
@@ -121,6 +121,7 @@ import { SupabaseBlock } from '@/blocks/blocks/supabase'
|
|||||||
import { TavilyBlock } from '@/blocks/blocks/tavily'
|
import { TavilyBlock } from '@/blocks/blocks/tavily'
|
||||||
import { TelegramBlock } from '@/blocks/blocks/telegram'
|
import { TelegramBlock } from '@/blocks/blocks/telegram'
|
||||||
import { ThinkingBlock } from '@/blocks/blocks/thinking'
|
import { ThinkingBlock } from '@/blocks/blocks/thinking'
|
||||||
|
import { TinybirdBlock } from '@/blocks/blocks/tinybird'
|
||||||
import { TranslateBlock } from '@/blocks/blocks/translate'
|
import { TranslateBlock } from '@/blocks/blocks/translate'
|
||||||
import { TrelloBlock } from '@/blocks/blocks/trello'
|
import { TrelloBlock } from '@/blocks/blocks/trello'
|
||||||
import { TtsBlock } from '@/blocks/blocks/tts'
|
import { TtsBlock } from '@/blocks/blocks/tts'
|
||||||
@@ -281,6 +282,7 @@ export const registry: Record<string, BlockConfig> = {
|
|||||||
tavily: TavilyBlock,
|
tavily: TavilyBlock,
|
||||||
telegram: TelegramBlock,
|
telegram: TelegramBlock,
|
||||||
thinking: ThinkingBlock,
|
thinking: ThinkingBlock,
|
||||||
|
tinybird: TinybirdBlock,
|
||||||
translate: TranslateBlock,
|
translate: TranslateBlock,
|
||||||
trello: TrelloBlock,
|
trello: TrelloBlock,
|
||||||
twilio_sms: TwilioSMSBlock,
|
twilio_sms: TwilioSMSBlock,
|
||||||
@@ -313,6 +315,26 @@ export const getBlock = (type: string): BlockConfig | undefined => {
|
|||||||
return registry[normalized]
|
return registry[normalized]
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export const getLatestBlock = (baseType: string): BlockConfig | undefined => {
|
||||||
|
const normalized = baseType.replace(/-/g, '_')
|
||||||
|
|
||||||
|
const versionedKeys = Object.keys(registry).filter((key) => {
|
||||||
|
const match = key.match(new RegExp(`^${normalized}_v(\\d+)$`))
|
||||||
|
return match !== null
|
||||||
|
})
|
||||||
|
|
||||||
|
if (versionedKeys.length > 0) {
|
||||||
|
const sorted = versionedKeys.sort((a, b) => {
|
||||||
|
const versionA = Number.parseInt(a.match(/_v(\d+)$/)?.[1] || '0', 10)
|
||||||
|
const versionB = Number.parseInt(b.match(/_v(\d+)$/)?.[1] || '0', 10)
|
||||||
|
return versionB - versionA
|
||||||
|
})
|
||||||
|
return registry[sorted[0]]
|
||||||
|
}
|
||||||
|
|
||||||
|
return registry[normalized]
|
||||||
|
}
|
||||||
|
|
||||||
export const getBlockByToolName = (toolName: string): BlockConfig | undefined => {
|
export const getBlockByToolName = (toolName: string): BlockConfig | undefined => {
|
||||||
return Object.values(registry).find((block) => block.tools?.access?.includes(toolName))
|
return Object.values(registry).find((block) => block.tools?.access?.includes(toolName))
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1897,6 +1897,19 @@ export function TelegramIcon(props: SVGProps<SVGSVGElement>) {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function TinybirdIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
|
return (
|
||||||
|
<svg {...props} xmlns='http://www.w3.org/2000/svg' viewBox='0 0 24 24' fill='none'>
|
||||||
|
<rect x='0' y='0' width='24' height='24' fill='#2EF598' rx='6' />
|
||||||
|
<g transform='translate(2, 2) scale(0.833)'>
|
||||||
|
<path d='M25 2.64 17.195.5 14.45 6.635z' fill='#1E7F63' />
|
||||||
|
<path d='M17.535 17.77 10.39 15.215 6.195 25.5z' fill='#1E7F63' />
|
||||||
|
<path d='M0 11.495 17.535 17.77 20.41 4.36z' fill='#1F2437' />
|
||||||
|
</g>
|
||||||
|
</svg>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
export function ClayIcon(props: SVGProps<SVGSVGElement>) {
|
export function ClayIcon(props: SVGProps<SVGSVGElement>) {
|
||||||
return (
|
return (
|
||||||
<svg {...props} xmlns='http://www.w3.org/2000/svg' width='40' height='40' viewBox='0 0 400 400'>
|
<svg {...props} xmlns='http://www.w3.org/2000/svg' width='40' height='40' viewBox='0 0 400 400'>
|
||||||
|
|||||||
@@ -378,21 +378,10 @@ function buildManualTriggerOutput(
|
|||||||
}
|
}
|
||||||
|
|
||||||
function buildIntegrationTriggerOutput(
|
function buildIntegrationTriggerOutput(
|
||||||
finalInput: unknown,
|
_finalInput: unknown,
|
||||||
workflowInput: unknown
|
workflowInput: unknown
|
||||||
): NormalizedBlockOutput {
|
): NormalizedBlockOutput {
|
||||||
const base: NormalizedBlockOutput = isPlainObject(workflowInput)
|
return isPlainObject(workflowInput) ? (workflowInput as NormalizedBlockOutput) : {}
|
||||||
? ({ ...(workflowInput as Record<string, unknown>) } as NormalizedBlockOutput)
|
|
||||||
: {}
|
|
||||||
|
|
||||||
if (isPlainObject(finalInput)) {
|
|
||||||
Object.assign(base, finalInput as Record<string, unknown>)
|
|
||||||
base.input = { ...(finalInput as Record<string, unknown>) }
|
|
||||||
} else {
|
|
||||||
base.input = finalInput
|
|
||||||
}
|
|
||||||
|
|
||||||
return mergeFilesIntoOutput(base, workflowInput)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
function extractSubBlocks(block: SerializedBlock): Record<string, unknown> | undefined {
|
function extractSubBlocks(block: SerializedBlock): Record<string, unknown> | undefined {
|
||||||
|
|||||||
@@ -22,7 +22,7 @@ import { useUndoRedoStore } from '@/stores/undo-redo'
|
|||||||
import { useWorkflowDiffStore } from '@/stores/workflow-diff/store'
|
import { useWorkflowDiffStore } from '@/stores/workflow-diff/store'
|
||||||
import { useWorkflowRegistry } from '@/stores/workflows/registry/store'
|
import { useWorkflowRegistry } from '@/stores/workflows/registry/store'
|
||||||
import { useSubBlockStore } from '@/stores/workflows/subblock/store'
|
import { useSubBlockStore } from '@/stores/workflows/subblock/store'
|
||||||
import { mergeSubblockState, normalizeName } from '@/stores/workflows/utils'
|
import { filterNewEdges, mergeSubblockState, normalizeName } from '@/stores/workflows/utils'
|
||||||
import { useWorkflowStore } from '@/stores/workflows/workflow/store'
|
import { useWorkflowStore } from '@/stores/workflows/workflow/store'
|
||||||
import type { BlockState, Loop, Parallel, Position } from '@/stores/workflows/workflow/types'
|
import type { BlockState, Loop, Parallel, Position } from '@/stores/workflows/workflow/types'
|
||||||
|
|
||||||
@@ -242,7 +242,10 @@ export function useCollaborativeWorkflow() {
|
|||||||
case EDGES_OPERATIONS.BATCH_ADD_EDGES: {
|
case EDGES_OPERATIONS.BATCH_ADD_EDGES: {
|
||||||
const { edges } = payload
|
const { edges } = payload
|
||||||
if (Array.isArray(edges) && edges.length > 0) {
|
if (Array.isArray(edges) && edges.length > 0) {
|
||||||
workflowStore.batchAddEdges(edges)
|
const newEdges = filterNewEdges(edges, workflowStore.edges)
|
||||||
|
if (newEdges.length > 0) {
|
||||||
|
workflowStore.batchAddEdges(newEdges)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
@@ -424,35 +427,6 @@ export function useCollaborativeWorkflow() {
|
|||||||
logger.info('Successfully applied batch-update-parent from remote user')
|
logger.info('Successfully applied batch-update-parent from remote user')
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
case BLOCKS_OPERATIONS.GROUP_BLOCKS: {
|
|
||||||
const { blockIds, groupId } = payload
|
|
||||||
logger.info('Received group-blocks from remote user', {
|
|
||||||
userId,
|
|
||||||
groupId,
|
|
||||||
blockCount: (blockIds || []).length,
|
|
||||||
})
|
|
||||||
|
|
||||||
if (blockIds && blockIds.length > 0 && groupId) {
|
|
||||||
workflowStore.groupBlocks(blockIds, groupId)
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.info('Successfully applied group-blocks from remote user')
|
|
||||||
break
|
|
||||||
}
|
|
||||||
case BLOCKS_OPERATIONS.UNGROUP_BLOCKS: {
|
|
||||||
const { groupId } = payload
|
|
||||||
logger.info('Received ungroup-blocks from remote user', {
|
|
||||||
userId,
|
|
||||||
groupId,
|
|
||||||
})
|
|
||||||
|
|
||||||
if (groupId) {
|
|
||||||
workflowStore.ungroupBlocks(groupId)
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.info('Successfully applied ungroup-blocks from remote user')
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
@@ -1005,6 +979,9 @@ export function useCollaborativeWorkflow() {
|
|||||||
|
|
||||||
if (edges.length === 0) return false
|
if (edges.length === 0) return false
|
||||||
|
|
||||||
|
const newEdges = filterNewEdges(edges, workflowStore.edges)
|
||||||
|
if (newEdges.length === 0) return false
|
||||||
|
|
||||||
const operationId = crypto.randomUUID()
|
const operationId = crypto.randomUUID()
|
||||||
|
|
||||||
addToQueue({
|
addToQueue({
|
||||||
@@ -1012,16 +989,16 @@ export function useCollaborativeWorkflow() {
|
|||||||
operation: {
|
operation: {
|
||||||
operation: EDGES_OPERATIONS.BATCH_ADD_EDGES,
|
operation: EDGES_OPERATIONS.BATCH_ADD_EDGES,
|
||||||
target: OPERATION_TARGETS.EDGES,
|
target: OPERATION_TARGETS.EDGES,
|
||||||
payload: { edges },
|
payload: { edges: newEdges },
|
||||||
},
|
},
|
||||||
workflowId: activeWorkflowId || '',
|
workflowId: activeWorkflowId || '',
|
||||||
userId: session?.user?.id || 'unknown',
|
userId: session?.user?.id || 'unknown',
|
||||||
})
|
})
|
||||||
|
|
||||||
workflowStore.batchAddEdges(edges)
|
workflowStore.batchAddEdges(newEdges)
|
||||||
|
|
||||||
if (!options?.skipUndoRedo) {
|
if (!options?.skipUndoRedo) {
|
||||||
edges.forEach((edge) => undoRedo.recordAddEdge(edge.id))
|
newEdges.forEach((edge) => undoRedo.recordAddEdge(edge.id))
|
||||||
}
|
}
|
||||||
|
|
||||||
return true
|
return true
|
||||||
@@ -1613,83 +1590,6 @@ export function useCollaborativeWorkflow() {
|
|||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
const collaborativeGroupBlocks = useCallback(
|
|
||||||
(blockIds: string[]) => {
|
|
||||||
if (!isInActiveRoom()) {
|
|
||||||
logger.debug('Skipping group blocks - not in active workflow')
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
|
|
||||||
if (blockIds.length < 2) {
|
|
||||||
logger.debug('Cannot group fewer than 2 blocks')
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
|
|
||||||
const groupId = crypto.randomUUID()
|
|
||||||
|
|
||||||
const operationId = crypto.randomUUID()
|
|
||||||
|
|
||||||
addToQueue({
|
|
||||||
id: operationId,
|
|
||||||
operation: {
|
|
||||||
operation: BLOCKS_OPERATIONS.GROUP_BLOCKS,
|
|
||||||
target: OPERATION_TARGETS.BLOCKS,
|
|
||||||
payload: { blockIds, groupId },
|
|
||||||
},
|
|
||||||
workflowId: activeWorkflowId || '',
|
|
||||||
userId: session?.user?.id || 'unknown',
|
|
||||||
})
|
|
||||||
|
|
||||||
workflowStore.groupBlocks(blockIds, groupId)
|
|
||||||
|
|
||||||
undoRedo.recordGroupBlocks(blockIds, groupId)
|
|
||||||
|
|
||||||
logger.info('Grouped blocks collaboratively', { groupId, blockCount: blockIds.length })
|
|
||||||
return groupId
|
|
||||||
},
|
|
||||||
[addToQueue, activeWorkflowId, session?.user?.id, isInActiveRoom, workflowStore, undoRedo]
|
|
||||||
)
|
|
||||||
|
|
||||||
const collaborativeUngroupBlocks = useCallback(
|
|
||||||
(groupId: string) => {
|
|
||||||
if (!isInActiveRoom()) {
|
|
||||||
logger.debug('Skipping ungroup blocks - not in active workflow')
|
|
||||||
return []
|
|
||||||
}
|
|
||||||
|
|
||||||
const groups = workflowStore.getGroups()
|
|
||||||
const group = groups[groupId]
|
|
||||||
|
|
||||||
if (!group) {
|
|
||||||
logger.warn('Cannot ungroup - group not found', { groupId })
|
|
||||||
return []
|
|
||||||
}
|
|
||||||
|
|
||||||
const blockIds = [...group.blockIds]
|
|
||||||
|
|
||||||
const operationId = crypto.randomUUID()
|
|
||||||
|
|
||||||
addToQueue({
|
|
||||||
id: operationId,
|
|
||||||
operation: {
|
|
||||||
operation: BLOCKS_OPERATIONS.UNGROUP_BLOCKS,
|
|
||||||
target: OPERATION_TARGETS.BLOCKS,
|
|
||||||
payload: { groupId, blockIds },
|
|
||||||
},
|
|
||||||
workflowId: activeWorkflowId || '',
|
|
||||||
userId: session?.user?.id || 'unknown',
|
|
||||||
})
|
|
||||||
|
|
||||||
workflowStore.ungroupBlocks(groupId)
|
|
||||||
|
|
||||||
undoRedo.recordUngroupBlocks(groupId, blockIds)
|
|
||||||
|
|
||||||
logger.info('Ungrouped blocks collaboratively', { groupId, blockCount: blockIds.length })
|
|
||||||
return blockIds
|
|
||||||
},
|
|
||||||
[addToQueue, activeWorkflowId, session?.user?.id, isInActiveRoom, workflowStore, undoRedo]
|
|
||||||
)
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
// Connection status
|
// Connection status
|
||||||
isConnected,
|
isConnected,
|
||||||
@@ -1728,10 +1628,6 @@ export function useCollaborativeWorkflow() {
|
|||||||
collaborativeUpdateIterationCount,
|
collaborativeUpdateIterationCount,
|
||||||
collaborativeUpdateIterationCollection,
|
collaborativeUpdateIterationCollection,
|
||||||
|
|
||||||
// Collaborative block group operations
|
|
||||||
collaborativeGroupBlocks,
|
|
||||||
collaborativeUngroupBlocks,
|
|
||||||
|
|
||||||
// Direct access to stores for non-collaborative operations
|
// Direct access to stores for non-collaborative operations
|
||||||
workflowStore,
|
workflowStore,
|
||||||
subBlockStore,
|
subBlockStore,
|
||||||
|
|||||||
@@ -22,9 +22,7 @@ import {
|
|||||||
type BatchToggleHandlesOperation,
|
type BatchToggleHandlesOperation,
|
||||||
type BatchUpdateParentOperation,
|
type BatchUpdateParentOperation,
|
||||||
createOperationEntry,
|
createOperationEntry,
|
||||||
type GroupBlocksOperation,
|
|
||||||
runWithUndoRedoRecordingSuspended,
|
runWithUndoRedoRecordingSuspended,
|
||||||
type UngroupBlocksOperation,
|
|
||||||
type UpdateParentOperation,
|
type UpdateParentOperation,
|
||||||
useUndoRedoStore,
|
useUndoRedoStore,
|
||||||
} from '@/stores/undo-redo'
|
} from '@/stores/undo-redo'
|
||||||
@@ -876,46 +874,6 @@ export function useUndoRedo() {
|
|||||||
})
|
})
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
case UNDO_REDO_OPERATIONS.GROUP_BLOCKS: {
|
|
||||||
// Undoing group = ungroup (inverse is ungroup operation)
|
|
||||||
const inverseOp = entry.inverse as unknown as UngroupBlocksOperation
|
|
||||||
const { groupId } = inverseOp.data
|
|
||||||
|
|
||||||
addToQueue({
|
|
||||||
id: opId,
|
|
||||||
operation: {
|
|
||||||
operation: BLOCKS_OPERATIONS.UNGROUP_BLOCKS,
|
|
||||||
target: OPERATION_TARGETS.BLOCKS,
|
|
||||||
payload: { groupId, blockIds: inverseOp.data.blockIds },
|
|
||||||
},
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
userId,
|
|
||||||
})
|
|
||||||
|
|
||||||
workflowStore.ungroupBlocks(groupId)
|
|
||||||
logger.debug('Undid group blocks', { groupId })
|
|
||||||
break
|
|
||||||
}
|
|
||||||
case UNDO_REDO_OPERATIONS.UNGROUP_BLOCKS: {
|
|
||||||
// Undoing ungroup = re-group (inverse is group operation)
|
|
||||||
const inverseOp = entry.inverse as unknown as GroupBlocksOperation
|
|
||||||
const { groupId, blockIds } = inverseOp.data
|
|
||||||
|
|
||||||
addToQueue({
|
|
||||||
id: opId,
|
|
||||||
operation: {
|
|
||||||
operation: BLOCKS_OPERATIONS.GROUP_BLOCKS,
|
|
||||||
target: OPERATION_TARGETS.BLOCKS,
|
|
||||||
payload: { groupId, blockIds },
|
|
||||||
},
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
userId,
|
|
||||||
})
|
|
||||||
|
|
||||||
workflowStore.groupBlocks(blockIds, groupId)
|
|
||||||
logger.debug('Undid ungroup blocks', { groupId })
|
|
||||||
break
|
|
||||||
}
|
|
||||||
case UNDO_REDO_OPERATIONS.APPLY_DIFF: {
|
case UNDO_REDO_OPERATIONS.APPLY_DIFF: {
|
||||||
const applyDiffInverse = entry.inverse as any
|
const applyDiffInverse = entry.inverse as any
|
||||||
const { baselineSnapshot } = applyDiffInverse.data
|
const { baselineSnapshot } = applyDiffInverse.data
|
||||||
@@ -1524,46 +1482,6 @@ export function useUndoRedo() {
|
|||||||
})
|
})
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
case UNDO_REDO_OPERATIONS.GROUP_BLOCKS: {
|
|
||||||
// Redo group = group again
|
|
||||||
const groupOp = entry.operation as GroupBlocksOperation
|
|
||||||
const { groupId, blockIds } = groupOp.data
|
|
||||||
|
|
||||||
addToQueue({
|
|
||||||
id: opId,
|
|
||||||
operation: {
|
|
||||||
operation: BLOCKS_OPERATIONS.GROUP_BLOCKS,
|
|
||||||
target: OPERATION_TARGETS.BLOCKS,
|
|
||||||
payload: { groupId, blockIds },
|
|
||||||
},
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
userId,
|
|
||||||
})
|
|
||||||
|
|
||||||
workflowStore.groupBlocks(blockIds, groupId)
|
|
||||||
logger.debug('Redid group blocks', { groupId })
|
|
||||||
break
|
|
||||||
}
|
|
||||||
case UNDO_REDO_OPERATIONS.UNGROUP_BLOCKS: {
|
|
||||||
// Redo ungroup = ungroup again
|
|
||||||
const ungroupOp = entry.operation as UngroupBlocksOperation
|
|
||||||
const { groupId } = ungroupOp.data
|
|
||||||
|
|
||||||
addToQueue({
|
|
||||||
id: opId,
|
|
||||||
operation: {
|
|
||||||
operation: BLOCKS_OPERATIONS.UNGROUP_BLOCKS,
|
|
||||||
target: OPERATION_TARGETS.BLOCKS,
|
|
||||||
payload: { groupId, blockIds: ungroupOp.data.blockIds },
|
|
||||||
},
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
userId,
|
|
||||||
})
|
|
||||||
|
|
||||||
workflowStore.ungroupBlocks(groupId)
|
|
||||||
logger.debug('Redid ungroup blocks', { groupId })
|
|
||||||
break
|
|
||||||
}
|
|
||||||
case UNDO_REDO_OPERATIONS.APPLY_DIFF: {
|
case UNDO_REDO_OPERATIONS.APPLY_DIFF: {
|
||||||
// Redo apply-diff means re-applying the proposed state with diff markers
|
// Redo apply-diff means re-applying the proposed state with diff markers
|
||||||
const applyDiffOp = entry.operation as any
|
const applyDiffOp = entry.operation as any
|
||||||
@@ -1875,66 +1793,6 @@ export function useUndoRedo() {
|
|||||||
[activeWorkflowId, userId, undoRedoStore]
|
[activeWorkflowId, userId, undoRedoStore]
|
||||||
)
|
)
|
||||||
|
|
||||||
const recordGroupBlocks = useCallback(
|
|
||||||
(blockIds: string[], groupId: string) => {
|
|
||||||
if (!activeWorkflowId || blockIds.length === 0) return
|
|
||||||
|
|
||||||
const operation: GroupBlocksOperation = {
|
|
||||||
id: crypto.randomUUID(),
|
|
||||||
type: UNDO_REDO_OPERATIONS.GROUP_BLOCKS,
|
|
||||||
timestamp: Date.now(),
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
userId,
|
|
||||||
data: { groupId, blockIds },
|
|
||||||
}
|
|
||||||
|
|
||||||
const inverse: UngroupBlocksOperation = {
|
|
||||||
id: crypto.randomUUID(),
|
|
||||||
type: UNDO_REDO_OPERATIONS.UNGROUP_BLOCKS,
|
|
||||||
timestamp: Date.now(),
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
userId,
|
|
||||||
data: { groupId, blockIds },
|
|
||||||
}
|
|
||||||
|
|
||||||
const entry = createOperationEntry(operation, inverse)
|
|
||||||
undoRedoStore.push(activeWorkflowId, userId, entry)
|
|
||||||
|
|
||||||
logger.debug('Recorded group blocks', { groupId, blockCount: blockIds.length })
|
|
||||||
},
|
|
||||||
[activeWorkflowId, userId, undoRedoStore]
|
|
||||||
)
|
|
||||||
|
|
||||||
const recordUngroupBlocks = useCallback(
|
|
||||||
(groupId: string, blockIds: string[], parentGroupId?: string) => {
|
|
||||||
if (!activeWorkflowId || blockIds.length === 0) return
|
|
||||||
|
|
||||||
const operation: UngroupBlocksOperation = {
|
|
||||||
id: crypto.randomUUID(),
|
|
||||||
type: UNDO_REDO_OPERATIONS.UNGROUP_BLOCKS,
|
|
||||||
timestamp: Date.now(),
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
userId,
|
|
||||||
data: { groupId, blockIds, parentGroupId },
|
|
||||||
}
|
|
||||||
|
|
||||||
const inverse: GroupBlocksOperation = {
|
|
||||||
id: crypto.randomUUID(),
|
|
||||||
type: UNDO_REDO_OPERATIONS.GROUP_BLOCKS,
|
|
||||||
timestamp: Date.now(),
|
|
||||||
workflowId: activeWorkflowId,
|
|
||||||
userId,
|
|
||||||
data: { groupId, blockIds },
|
|
||||||
}
|
|
||||||
|
|
||||||
const entry = createOperationEntry(operation, inverse)
|
|
||||||
undoRedoStore.push(activeWorkflowId, userId, entry)
|
|
||||||
|
|
||||||
logger.debug('Recorded ungroup blocks', { groupId, blockCount: blockIds.length })
|
|
||||||
},
|
|
||||||
[activeWorkflowId, userId, undoRedoStore]
|
|
||||||
)
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
recordBatchAddBlocks,
|
recordBatchAddBlocks,
|
||||||
recordBatchRemoveBlocks,
|
recordBatchRemoveBlocks,
|
||||||
@@ -1948,8 +1806,6 @@ export function useUndoRedo() {
|
|||||||
recordApplyDiff,
|
recordApplyDiff,
|
||||||
recordAcceptDiff,
|
recordAcceptDiff,
|
||||||
recordRejectDiff,
|
recordRejectDiff,
|
||||||
recordGroupBlocks,
|
|
||||||
recordUngroupBlocks,
|
|
||||||
undo,
|
undo,
|
||||||
redo,
|
redo,
|
||||||
getStackSizes,
|
getStackSizes,
|
||||||
|
|||||||
@@ -36,9 +36,10 @@ class ApiKeyInterceptor implements CallInterceptor {
|
|||||||
/**
|
/**
|
||||||
* Create an A2A client from an agent URL with optional API key authentication
|
* Create an A2A client from an agent URL with optional API key authentication
|
||||||
*
|
*
|
||||||
* The agent URL should be the full endpoint URL (e.g., /api/a2a/serve/{agentId}).
|
* Supports both standard A2A agents (agent card at /.well-known/agent.json)
|
||||||
* We pass an empty path to createFromUrl so it uses the URL directly for agent card
|
* and Sim Studio agents (agent card at root URL via GET).
|
||||||
* discovery (GET on the URL) instead of appending .well-known/agent-card.json.
|
*
|
||||||
|
* Tries standard path first, falls back to root URL for compatibility.
|
||||||
*/
|
*/
|
||||||
export async function createA2AClient(agentUrl: string, apiKey?: string): Promise<Client> {
|
export async function createA2AClient(agentUrl: string, apiKey?: string): Promise<Client> {
|
||||||
const factoryOptions = apiKey
|
const factoryOptions = apiKey
|
||||||
@@ -49,6 +50,18 @@ export async function createA2AClient(agentUrl: string, apiKey?: string): Promis
|
|||||||
})
|
})
|
||||||
: ClientFactoryOptions.default
|
: ClientFactoryOptions.default
|
||||||
const factory = new ClientFactory(factoryOptions)
|
const factory = new ClientFactory(factoryOptions)
|
||||||
|
|
||||||
|
// Try standard A2A path first (/.well-known/agent.json)
|
||||||
|
try {
|
||||||
|
return await factory.createFromUrl(agentUrl, '/.well-known/agent.json')
|
||||||
|
} catch (standardError) {
|
||||||
|
logger.debug('Standard agent card path failed, trying root URL', {
|
||||||
|
agentUrl,
|
||||||
|
error: standardError instanceof Error ? standardError.message : String(standardError),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// Fall back to root URL (Sim Studio compatibility)
|
||||||
return factory.createFromUrl(agentUrl, '')
|
return factory.createFromUrl(agentUrl, '')
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -656,7 +656,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.id.toString(),
|
id: `${profile.id.toString()}-${crypto.randomUUID()}`,
|
||||||
name: profile.name || profile.login,
|
name: profile.name || profile.login,
|
||||||
email: profile.email,
|
email: profile.email,
|
||||||
image: profile.avatar_url,
|
image: profile.avatar_url,
|
||||||
@@ -962,7 +962,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: uniqueId,
|
id: `${uniqueId}-${crypto.randomUUID()}`,
|
||||||
name: 'Wealthbox User',
|
name: 'Wealthbox User',
|
||||||
email: `${uniqueId}@wealthbox.user`,
|
email: `${uniqueId}@wealthbox.user`,
|
||||||
emailVerified: false,
|
emailVerified: false,
|
||||||
@@ -1016,7 +1016,7 @@ export const auth = betterAuth({
|
|||||||
const user = data.data
|
const user = data.data
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: user.id.toString(),
|
id: `${user.id.toString()}-${crypto.randomUUID()}`,
|
||||||
name: user.name,
|
name: user.name,
|
||||||
email: user.email,
|
email: user.email,
|
||||||
emailVerified: user.activated,
|
emailVerified: user.activated,
|
||||||
@@ -1108,7 +1108,7 @@ export const auth = betterAuth({
|
|||||||
})
|
})
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: data.user_id || data.hub_id.toString(),
|
id: `${data.user_id || data.hub_id.toString()}-${crypto.randomUUID()}`,
|
||||||
name: data.user || 'HubSpot User',
|
name: data.user || 'HubSpot User',
|
||||||
email: data.user || `hubspot-${data.hub_id}@hubspot.com`,
|
email: data.user || `hubspot-${data.hub_id}@hubspot.com`,
|
||||||
emailVerified: true,
|
emailVerified: true,
|
||||||
@@ -1162,7 +1162,7 @@ export const auth = betterAuth({
|
|||||||
const data = await response.json()
|
const data = await response.json()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: data.user_id || data.sub,
|
id: `${data.user_id || data.sub}-${crypto.randomUUID()}`,
|
||||||
name: data.name || 'Salesforce User',
|
name: data.name || 'Salesforce User',
|
||||||
email: data.email || `salesforce-${data.user_id}@salesforce.com`,
|
email: data.email || `salesforce-${data.user_id}@salesforce.com`,
|
||||||
emailVerified: data.email_verified || true,
|
emailVerified: data.email_verified || true,
|
||||||
@@ -1221,7 +1221,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.data.id,
|
id: `${profile.data.id}-${crypto.randomUUID()}`,
|
||||||
name: profile.data.name || 'X User',
|
name: profile.data.name || 'X User',
|
||||||
email: `${profile.data.username}@x.com`,
|
email: `${profile.data.username}@x.com`,
|
||||||
image: profile.data.profile_image_url,
|
image: profile.data.profile_image_url,
|
||||||
@@ -1295,7 +1295,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.account_id,
|
id: `${profile.account_id}-${crypto.randomUUID()}`,
|
||||||
name: profile.name || profile.display_name || 'Confluence User',
|
name: profile.name || profile.display_name || 'Confluence User',
|
||||||
email: profile.email || `${profile.account_id}@atlassian.com`,
|
email: profile.email || `${profile.account_id}@atlassian.com`,
|
||||||
image: profile.picture || undefined,
|
image: profile.picture || undefined,
|
||||||
@@ -1406,7 +1406,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.account_id,
|
id: `${profile.account_id}-${crypto.randomUUID()}`,
|
||||||
name: profile.name || profile.display_name || 'Jira User',
|
name: profile.name || profile.display_name || 'Jira User',
|
||||||
email: profile.email || `${profile.account_id}@atlassian.com`,
|
email: profile.email || `${profile.account_id}@atlassian.com`,
|
||||||
image: profile.picture || undefined,
|
image: profile.picture || undefined,
|
||||||
@@ -1456,7 +1456,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: data.id,
|
id: `${data.id}-${crypto.randomUUID()}`,
|
||||||
name: data.email ? data.email.split('@')[0] : 'Airtable User',
|
name: data.email ? data.email.split('@')[0] : 'Airtable User',
|
||||||
email: data.email || `${data.id}@airtable.user`,
|
email: data.email || `${data.id}@airtable.user`,
|
||||||
emailVerified: !!data.email,
|
emailVerified: !!data.email,
|
||||||
@@ -1505,7 +1505,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.bot?.owner?.user?.id || profile.id,
|
id: `${profile.bot?.owner?.user?.id || profile.id}-${crypto.randomUUID()}`,
|
||||||
name: profile.name || profile.bot?.owner?.user?.name || 'Notion User',
|
name: profile.name || profile.bot?.owner?.user?.name || 'Notion User',
|
||||||
email: profile.person?.email || `${profile.id}@notion.user`,
|
email: profile.person?.email || `${profile.id}@notion.user`,
|
||||||
emailVerified: !!profile.person?.email,
|
emailVerified: !!profile.person?.email,
|
||||||
@@ -1572,7 +1572,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: data.id,
|
id: `${data.id}-${crypto.randomUUID()}`,
|
||||||
name: data.name || 'Reddit User',
|
name: data.name || 'Reddit User',
|
||||||
email: `${data.name}@reddit.user`,
|
email: `${data.name}@reddit.user`,
|
||||||
image: data.icon_img || undefined,
|
image: data.icon_img || undefined,
|
||||||
@@ -1644,7 +1644,7 @@ export const auth = betterAuth({
|
|||||||
const viewer = data.viewer
|
const viewer = data.viewer
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: viewer.id,
|
id: `${viewer.id}-${crypto.randomUUID()}`,
|
||||||
email: viewer.email,
|
email: viewer.email,
|
||||||
name: viewer.name,
|
name: viewer.name,
|
||||||
emailVerified: true,
|
emailVerified: true,
|
||||||
@@ -1707,7 +1707,7 @@ export const auth = betterAuth({
|
|||||||
const data = await response.json()
|
const data = await response.json()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: data.account_id,
|
id: `${data.account_id}-${crypto.randomUUID()}`,
|
||||||
email: data.email,
|
email: data.email,
|
||||||
name: data.name?.display_name || data.email,
|
name: data.name?.display_name || data.email,
|
||||||
emailVerified: data.email_verified || false,
|
emailVerified: data.email_verified || false,
|
||||||
@@ -1758,7 +1758,7 @@ export const auth = betterAuth({
|
|||||||
const now = new Date()
|
const now = new Date()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.gid,
|
id: `${profile.gid}-${crypto.randomUUID()}`,
|
||||||
name: profile.name || 'Asana User',
|
name: profile.name || 'Asana User',
|
||||||
email: profile.email || `${profile.gid}@asana.user`,
|
email: profile.email || `${profile.gid}@asana.user`,
|
||||||
image: profile.photo?.image_128x128 || undefined,
|
image: profile.photo?.image_128x128 || undefined,
|
||||||
@@ -1834,7 +1834,7 @@ export const auth = betterAuth({
|
|||||||
logger.info('Slack credential identifier', { teamId, userId, uniqueId, teamName })
|
logger.info('Slack credential identifier', { teamId, userId, uniqueId, teamName })
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: uniqueId,
|
id: `${uniqueId}-${crypto.randomUUID()}`,
|
||||||
name: teamName,
|
name: teamName,
|
||||||
email: `${teamId}-${userId}@slack.bot`,
|
email: `${teamId}-${userId}@slack.bot`,
|
||||||
emailVerified: false,
|
emailVerified: false,
|
||||||
@@ -1884,7 +1884,7 @@ export const auth = betterAuth({
|
|||||||
const uniqueId = `webflow-${userId}`
|
const uniqueId = `webflow-${userId}`
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: uniqueId,
|
id: `${uniqueId}-${crypto.randomUUID()}`,
|
||||||
name: data.user_name || 'Webflow User',
|
name: data.user_name || 'Webflow User',
|
||||||
email: `${uniqueId.replace(/[^a-zA-Z0-9]/g, '')}@webflow.user`,
|
email: `${uniqueId.replace(/[^a-zA-Z0-9]/g, '')}@webflow.user`,
|
||||||
emailVerified: false,
|
emailVerified: false,
|
||||||
@@ -1931,7 +1931,7 @@ export const auth = betterAuth({
|
|||||||
const profile = await response.json()
|
const profile = await response.json()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.sub,
|
id: `${profile.sub}-${crypto.randomUUID()}`,
|
||||||
name: profile.name || 'LinkedIn User',
|
name: profile.name || 'LinkedIn User',
|
||||||
email: profile.email || `${profile.sub}@linkedin.user`,
|
email: profile.email || `${profile.sub}@linkedin.user`,
|
||||||
emailVerified: profile.email_verified || true,
|
emailVerified: profile.email_verified || true,
|
||||||
@@ -1993,7 +1993,7 @@ export const auth = betterAuth({
|
|||||||
const profile = await response.json()
|
const profile = await response.json()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.id,
|
id: `${profile.id}-${crypto.randomUUID()}`,
|
||||||
name:
|
name:
|
||||||
`${profile.first_name || ''} ${profile.last_name || ''}`.trim() || 'Zoom User',
|
`${profile.first_name || ''} ${profile.last_name || ''}`.trim() || 'Zoom User',
|
||||||
email: profile.email || `${profile.id}@zoom.user`,
|
email: profile.email || `${profile.id}@zoom.user`,
|
||||||
@@ -2060,7 +2060,7 @@ export const auth = betterAuth({
|
|||||||
const profile = await response.json()
|
const profile = await response.json()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.id,
|
id: `${profile.id}-${crypto.randomUUID()}`,
|
||||||
name: profile.display_name || 'Spotify User',
|
name: profile.display_name || 'Spotify User',
|
||||||
email: profile.email || `${profile.id}@spotify.user`,
|
email: profile.email || `${profile.id}@spotify.user`,
|
||||||
emailVerified: true,
|
emailVerified: true,
|
||||||
@@ -2108,7 +2108,7 @@ export const auth = betterAuth({
|
|||||||
const profile = await response.json()
|
const profile = await response.json()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
id: profile.ID?.toString() || profile.id?.toString(),
|
id: `${profile.ID?.toString() || profile.id?.toString()}-${crypto.randomUUID()}`,
|
||||||
name: profile.display_name || profile.username || 'WordPress User',
|
name: profile.display_name || profile.username || 'WordPress User',
|
||||||
email: profile.email || `${profile.username}@wordpress.com`,
|
email: profile.email || `${profile.username}@wordpress.com`,
|
||||||
emailVerified: profile.email_verified || false,
|
emailVerified: profile.email_verified || false,
|
||||||
|
|||||||
@@ -18,7 +18,7 @@ export const DEFAULT_ENTERPRISE_TIER_COST_LIMIT = 200
|
|||||||
* Base charge applied to every workflow execution
|
* Base charge applied to every workflow execution
|
||||||
* This charge is applied regardless of whether the workflow uses AI models
|
* This charge is applied regardless of whether the workflow uses AI models
|
||||||
*/
|
*/
|
||||||
export const BASE_EXECUTION_CHARGE = 0.001
|
export const BASE_EXECUTION_CHARGE = 0.005
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Fixed cost for search tool invocation (in dollars)
|
* Fixed cost for search tool invocation (in dollars)
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import {
|
|||||||
|
|
||||||
// Mock the billing constants
|
// Mock the billing constants
|
||||||
vi.mock('@/lib/billing/constants', () => ({
|
vi.mock('@/lib/billing/constants', () => ({
|
||||||
BASE_EXECUTION_CHARGE: 0.001,
|
BASE_EXECUTION_CHARGE: 0.005,
|
||||||
}))
|
}))
|
||||||
|
|
||||||
vi.mock('@sim/logger', () => loggerMock)
|
vi.mock('@sim/logger', () => loggerMock)
|
||||||
@@ -148,7 +148,7 @@ describe('createEnvironmentObject', () => {
|
|||||||
})
|
})
|
||||||
|
|
||||||
describe('calculateCostSummary', () => {
|
describe('calculateCostSummary', () => {
|
||||||
const BASE_EXECUTION_CHARGE = 0.001
|
const BASE_EXECUTION_CHARGE = 0.005
|
||||||
|
|
||||||
test('should return base execution charge for empty trace spans', () => {
|
test('should return base execution charge for empty trace spans', () => {
|
||||||
const result = calculateCostSummary([])
|
const result = calculateCostSummary([])
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { getBlock } from '@/blocks/registry'
|
import { getLatestBlock } from '@/blocks/registry'
|
||||||
import { getAllTriggers } from '@/triggers'
|
import { getAllTriggers } from '@/triggers'
|
||||||
|
|
||||||
export interface TriggerOption {
|
export interface TriggerOption {
|
||||||
@@ -49,22 +49,13 @@ export function getTriggerOptions(): TriggerOption[] {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
const block = getBlock(provider)
|
const block = getLatestBlock(provider)
|
||||||
|
|
||||||
if (block) {
|
providerMap.set(provider, {
|
||||||
providerMap.set(provider, {
|
value: provider,
|
||||||
value: provider,
|
label: block?.name || formatProviderName(provider),
|
||||||
label: block.name, // Use block's display name (e.g., "Slack", "GitHub")
|
color: block?.bgColor || '#6b7280',
|
||||||
color: block.bgColor || '#6b7280', // Use block's hex color, fallback to gray
|
})
|
||||||
})
|
|
||||||
} else {
|
|
||||||
const label = formatProviderName(provider)
|
|
||||||
providerMap.set(provider, {
|
|
||||||
value: provider,
|
|
||||||
label,
|
|
||||||
color: '#6b7280', // gray fallback
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const integrationOptions = Array.from(providerMap.values()).sort((a, b) =>
|
const integrationOptions = Array.from(providerMap.values()).sort((a, b) =>
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -16,61 +16,9 @@ import type { BlockState } from '@/stores/workflows/workflow/types'
|
|||||||
|
|
||||||
const logger = createLogger('AutoLayout')
|
const logger = createLogger('AutoLayout')
|
||||||
|
|
||||||
/** Default block dimensions for layout calculations */
|
|
||||||
const DEFAULT_BLOCK_WIDTH = 250
|
|
||||||
const DEFAULT_BLOCK_HEIGHT = 100
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Identifies groups from blocks and calculates their bounding boxes.
|
|
||||||
* Returns a map of groupId to group info including bounding box and member block IDs.
|
|
||||||
*/
|
|
||||||
function identifyGroups(blocks: Record<string, BlockState>): Map<
|
|
||||||
string,
|
|
||||||
{
|
|
||||||
blockIds: string[]
|
|
||||||
bounds: { minX: number; minY: number; maxX: number; maxY: number }
|
|
||||||
}
|
|
||||||
> {
|
|
||||||
const groups = new Map<
|
|
||||||
string,
|
|
||||||
{
|
|
||||||
blockIds: string[]
|
|
||||||
bounds: { minX: number; minY: number; maxX: number; maxY: number }
|
|
||||||
}
|
|
||||||
>()
|
|
||||||
|
|
||||||
// Group blocks by their groupId
|
|
||||||
for (const [blockId, block] of Object.entries(blocks)) {
|
|
||||||
const groupId = block.data?.groupId
|
|
||||||
if (!groupId) continue
|
|
||||||
|
|
||||||
if (!groups.has(groupId)) {
|
|
||||||
groups.set(groupId, {
|
|
||||||
blockIds: [],
|
|
||||||
bounds: { minX: Infinity, minY: Infinity, maxX: -Infinity, maxY: -Infinity },
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
const group = groups.get(groupId)!
|
|
||||||
group.blockIds.push(blockId)
|
|
||||||
|
|
||||||
// Update bounding box
|
|
||||||
const blockWidth = block.data?.width ?? DEFAULT_BLOCK_WIDTH
|
|
||||||
const blockHeight = block.data?.height ?? block.height ?? DEFAULT_BLOCK_HEIGHT
|
|
||||||
|
|
||||||
group.bounds.minX = Math.min(group.bounds.minX, block.position.x)
|
|
||||||
group.bounds.minY = Math.min(group.bounds.minY, block.position.y)
|
|
||||||
group.bounds.maxX = Math.max(group.bounds.maxX, block.position.x + blockWidth)
|
|
||||||
group.bounds.maxY = Math.max(group.bounds.maxY, block.position.y + blockHeight)
|
|
||||||
}
|
|
||||||
|
|
||||||
return groups
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Applies automatic layout to all blocks in a workflow.
|
* Applies automatic layout to all blocks in a workflow.
|
||||||
* Positions blocks in layers based on their connections (edges).
|
* Positions blocks in layers based on their connections (edges).
|
||||||
* Groups are treated as single units and laid out together.
|
|
||||||
*/
|
*/
|
||||||
export function applyAutoLayout(
|
export function applyAutoLayout(
|
||||||
blocks: Record<string, BlockState>,
|
blocks: Record<string, BlockState>,
|
||||||
@@ -88,11 +36,6 @@ export function applyAutoLayout(
|
|||||||
const horizontalSpacing = options.horizontalSpacing ?? DEFAULT_HORIZONTAL_SPACING
|
const horizontalSpacing = options.horizontalSpacing ?? DEFAULT_HORIZONTAL_SPACING
|
||||||
const verticalSpacing = options.verticalSpacing ?? DEFAULT_VERTICAL_SPACING
|
const verticalSpacing = options.verticalSpacing ?? DEFAULT_VERTICAL_SPACING
|
||||||
|
|
||||||
// Identify groups and their bounding boxes
|
|
||||||
const groups = identifyGroups(blocksCopy)
|
|
||||||
|
|
||||||
logger.info('Identified block groups for layout', { groupCount: groups.size })
|
|
||||||
|
|
||||||
// Pre-calculate container dimensions by laying out their children (bottom-up)
|
// Pre-calculate container dimensions by laying out their children (bottom-up)
|
||||||
// This ensures accurate widths/heights before root-level layout
|
// This ensures accurate widths/heights before root-level layout
|
||||||
prepareContainerDimensions(
|
prepareContainerDimensions(
|
||||||
@@ -106,112 +49,19 @@ export function applyAutoLayout(
|
|||||||
const { root: rootBlockIds } = getBlocksByParent(blocksCopy)
|
const { root: rootBlockIds } = getBlocksByParent(blocksCopy)
|
||||||
const layoutRootIds = filterLayoutEligibleBlockIds(rootBlockIds, blocksCopy)
|
const layoutRootIds = filterLayoutEligibleBlockIds(rootBlockIds, blocksCopy)
|
||||||
|
|
||||||
// For groups, we need to:
|
|
||||||
// 1. Create virtual blocks representing each group
|
|
||||||
// 2. Replace grouped blocks with their group's virtual block
|
|
||||||
// 3. Layout the virtual blocks + ungrouped blocks
|
|
||||||
// 4. Apply position deltas to grouped blocks
|
|
||||||
|
|
||||||
// Track which blocks are in groups at root level
|
|
||||||
const groupedRootBlockIds = new Set<string>()
|
|
||||||
const groupRepresentatives = new Map<string, string>() // groupId -> representative blockId
|
|
||||||
|
|
||||||
// Store ORIGINAL positions of all grouped blocks before any modifications
|
|
||||||
const originalBlockPositions = new Map<string, { x: number; y: number }>()
|
|
||||||
for (const [_groupId, group] of groups) {
|
|
||||||
for (const blockId of group.blockIds) {
|
|
||||||
if (blocksCopy[blockId]) {
|
|
||||||
originalBlockPositions.set(blockId, { ...blocksCopy[blockId].position })
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
for (const [groupId, group] of groups) {
|
|
||||||
// Find if any blocks in this group are at root level
|
|
||||||
const rootGroupBlocks = group.blockIds.filter((id) => layoutRootIds.includes(id))
|
|
||||||
if (rootGroupBlocks.length > 0) {
|
|
||||||
// Mark all blocks in this group as grouped
|
|
||||||
for (const blockId of rootGroupBlocks) {
|
|
||||||
groupedRootBlockIds.add(blockId)
|
|
||||||
}
|
|
||||||
// Use the first block as the group's representative for layout
|
|
||||||
const representativeId = rootGroupBlocks[0]
|
|
||||||
groupRepresentatives.set(groupId, representativeId)
|
|
||||||
|
|
||||||
// Update the representative block's dimensions to match the group's bounding box
|
|
||||||
const bounds = group.bounds
|
|
||||||
const groupWidth = bounds.maxX - bounds.minX
|
|
||||||
const groupHeight = bounds.maxY - bounds.minY
|
|
||||||
|
|
||||||
blocksCopy[representativeId] = {
|
|
||||||
...blocksCopy[representativeId],
|
|
||||||
data: {
|
|
||||||
...blocksCopy[representativeId].data,
|
|
||||||
width: groupWidth,
|
|
||||||
height: groupHeight,
|
|
||||||
},
|
|
||||||
// Position at the group's top-left corner
|
|
||||||
position: { x: bounds.minX, y: bounds.minY },
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Build the blocks to layout: ungrouped blocks + group representatives
|
|
||||||
const rootBlocks: Record<string, BlockState> = {}
|
const rootBlocks: Record<string, BlockState> = {}
|
||||||
for (const id of layoutRootIds) {
|
for (const id of layoutRootIds) {
|
||||||
// Skip grouped blocks that aren't representatives
|
rootBlocks[id] = blocksCopy[id]
|
||||||
if (groupedRootBlockIds.has(id)) {
|
|
||||||
// Only include if this is a group representative
|
|
||||||
for (const [groupId, repId] of groupRepresentatives) {
|
|
||||||
if (repId === id) {
|
|
||||||
rootBlocks[id] = blocksCopy[id]
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
rootBlocks[id] = blocksCopy[id]
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Remap edges: edges involving grouped blocks should connect to the representative
|
const rootEdges = edges.filter(
|
||||||
const blockToGroup = new Map<string, string>() // blockId -> groupId
|
(edge) => layoutRootIds.includes(edge.source) && layoutRootIds.includes(edge.target)
|
||||||
for (const [groupId, group] of groups) {
|
)
|
||||||
for (const blockId of group.blockIds) {
|
|
||||||
blockToGroup.set(blockId, groupId)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const layoutBlockIds = new Set(Object.keys(rootBlocks))
|
|
||||||
const rootEdges = edges
|
|
||||||
.map((edge) => {
|
|
||||||
let source = edge.source
|
|
||||||
let target = edge.target
|
|
||||||
|
|
||||||
// Remap source if it's in a group
|
|
||||||
const sourceGroupId = blockToGroup.get(source)
|
|
||||||
if (sourceGroupId && groupRepresentatives.has(sourceGroupId)) {
|
|
||||||
source = groupRepresentatives.get(sourceGroupId)!
|
|
||||||
}
|
|
||||||
|
|
||||||
// Remap target if it's in a group
|
|
||||||
const targetGroupId = blockToGroup.get(target)
|
|
||||||
if (targetGroupId && groupRepresentatives.has(targetGroupId)) {
|
|
||||||
target = groupRepresentatives.get(targetGroupId)!
|
|
||||||
}
|
|
||||||
|
|
||||||
return { ...edge, source, target }
|
|
||||||
})
|
|
||||||
.filter((edge) => layoutBlockIds.has(edge.source) && layoutBlockIds.has(edge.target))
|
|
||||||
|
|
||||||
// Calculate subflow depths before laying out root blocks
|
// Calculate subflow depths before laying out root blocks
|
||||||
|
// This ensures blocks connected to subflow ends are positioned correctly
|
||||||
const subflowDepths = calculateSubflowDepths(blocksCopy, edges, assignLayers)
|
const subflowDepths = calculateSubflowDepths(blocksCopy, edges, assignLayers)
|
||||||
|
|
||||||
// Store old positions for groups to calculate deltas
|
|
||||||
const oldGroupPositions = new Map<string, { x: number; y: number }>()
|
|
||||||
for (const [groupId, repId] of groupRepresentatives) {
|
|
||||||
oldGroupPositions.set(groupId, { ...blocksCopy[repId].position })
|
|
||||||
}
|
|
||||||
|
|
||||||
if (Object.keys(rootBlocks).length > 0) {
|
if (Object.keys(rootBlocks).length > 0) {
|
||||||
const { nodes } = layoutBlocksCore(rootBlocks, rootEdges, {
|
const { nodes } = layoutBlocksCore(rootBlocks, rootEdges, {
|
||||||
isContainer: false,
|
isContainer: false,
|
||||||
@@ -219,49 +69,15 @@ export function applyAutoLayout(
|
|||||||
subflowDepths,
|
subflowDepths,
|
||||||
})
|
})
|
||||||
|
|
||||||
// Apply positions to ungrouped blocks and group representatives
|
|
||||||
for (const node of nodes.values()) {
|
for (const node of nodes.values()) {
|
||||||
blocksCopy[node.id].position = node.position
|
blocksCopy[node.id].position = node.position
|
||||||
}
|
}
|
||||||
|
|
||||||
// For each group, calculate the delta and apply to ALL blocks in the group
|
|
||||||
for (const [groupId, repId] of groupRepresentatives) {
|
|
||||||
const oldGroupTopLeft = oldGroupPositions.get(groupId)!
|
|
||||||
const newGroupTopLeft = blocksCopy[repId].position
|
|
||||||
const deltaX = newGroupTopLeft.x - oldGroupTopLeft.x
|
|
||||||
const deltaY = newGroupTopLeft.y - oldGroupTopLeft.y
|
|
||||||
|
|
||||||
const group = groups.get(groupId)!
|
|
||||||
// Apply delta to ALL blocks in the group using their ORIGINAL positions
|
|
||||||
for (const blockId of group.blockIds) {
|
|
||||||
if (layoutRootIds.includes(blockId)) {
|
|
||||||
const originalPos = originalBlockPositions.get(blockId)
|
|
||||||
if (originalPos) {
|
|
||||||
blocksCopy[blockId].position = {
|
|
||||||
x: originalPos.x + deltaX,
|
|
||||||
y: originalPos.y + deltaY,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Restore the representative's original dimensions
|
|
||||||
const originalBlock = blocks[repId]
|
|
||||||
if (originalBlock) {
|
|
||||||
blocksCopy[repId].data = {
|
|
||||||
...blocksCopy[repId].data,
|
|
||||||
width: originalBlock.data?.width,
|
|
||||||
height: originalBlock.data?.height,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
layoutContainers(blocksCopy, edges, options)
|
layoutContainers(blocksCopy, edges, options)
|
||||||
|
|
||||||
logger.info('Auto layout completed successfully', {
|
logger.info('Auto layout completed successfully', {
|
||||||
blockCount: Object.keys(blocksCopy).length,
|
blockCount: Object.keys(blocksCopy).length,
|
||||||
groupCount: groups.size,
|
|
||||||
})
|
})
|
||||||
|
|
||||||
return {
|
return {
|
||||||
|
|||||||
@@ -26,53 +26,9 @@ export interface TargetedLayoutOptions extends LayoutOptions {
|
|||||||
horizontalSpacing?: number
|
horizontalSpacing?: number
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Identifies block groups from the blocks' groupId data.
|
|
||||||
* Returns a map of groupId to array of block IDs in that group.
|
|
||||||
*/
|
|
||||||
function identifyBlockGroups(blocks: Record<string, BlockState>): Map<string, string[]> {
|
|
||||||
const groups = new Map<string, string[]>()
|
|
||||||
|
|
||||||
for (const [blockId, block] of Object.entries(blocks)) {
|
|
||||||
const groupId = block.data?.groupId
|
|
||||||
if (!groupId) continue
|
|
||||||
|
|
||||||
if (!groups.has(groupId)) {
|
|
||||||
groups.set(groupId, [])
|
|
||||||
}
|
|
||||||
groups.get(groupId)!.push(blockId)
|
|
||||||
}
|
|
||||||
|
|
||||||
return groups
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Expands changed block IDs to include all blocks in the same group.
|
|
||||||
* If any block in a group changed, all blocks in that group should be treated as changed.
|
|
||||||
*/
|
|
||||||
function expandChangedToGroups(
|
|
||||||
changedBlockIds: string[],
|
|
||||||
blockGroups: Map<string, string[]>,
|
|
||||||
blocks: Record<string, BlockState>
|
|
||||||
): Set<string> {
|
|
||||||
const expandedSet = new Set(changedBlockIds)
|
|
||||||
|
|
||||||
for (const blockId of changedBlockIds) {
|
|
||||||
const groupId = blocks[blockId]?.data?.groupId
|
|
||||||
if (groupId && blockGroups.has(groupId)) {
|
|
||||||
for (const groupBlockId of blockGroups.get(groupId)!) {
|
|
||||||
expandedSet.add(groupBlockId)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return expandedSet
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Applies targeted layout to only reposition changed blocks.
|
* Applies targeted layout to only reposition changed blocks.
|
||||||
* Unchanged blocks act as anchors to preserve existing layout.
|
* Unchanged blocks act as anchors to preserve existing layout.
|
||||||
* Blocks in groups are moved together as a unit.
|
|
||||||
*/
|
*/
|
||||||
export function applyTargetedLayout(
|
export function applyTargetedLayout(
|
||||||
blocks: Record<string, BlockState>,
|
blocks: Record<string, BlockState>,
|
||||||
@@ -89,14 +45,9 @@ export function applyTargetedLayout(
|
|||||||
return blocks
|
return blocks
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const changedSet = new Set(changedBlockIds)
|
||||||
const blocksCopy: Record<string, BlockState> = JSON.parse(JSON.stringify(blocks))
|
const blocksCopy: Record<string, BlockState> = JSON.parse(JSON.stringify(blocks))
|
||||||
|
|
||||||
// Identify block groups
|
|
||||||
const blockGroups = identifyBlockGroups(blocksCopy)
|
|
||||||
|
|
||||||
// Expand changed set to include all blocks in affected groups
|
|
||||||
const changedSet = expandChangedToGroups(changedBlockIds, blockGroups, blocksCopy)
|
|
||||||
|
|
||||||
// Pre-calculate container dimensions by laying out their children (bottom-up)
|
// Pre-calculate container dimensions by laying out their children (bottom-up)
|
||||||
// This ensures accurate widths/heights before root-level layout
|
// This ensures accurate widths/heights before root-level layout
|
||||||
prepareContainerDimensions(
|
prepareContainerDimensions(
|
||||||
@@ -120,8 +71,7 @@ export function applyTargetedLayout(
|
|||||||
changedSet,
|
changedSet,
|
||||||
verticalSpacing,
|
verticalSpacing,
|
||||||
horizontalSpacing,
|
horizontalSpacing,
|
||||||
subflowDepths,
|
subflowDepths
|
||||||
blockGroups
|
|
||||||
)
|
)
|
||||||
|
|
||||||
for (const [parentId, childIds] of groups.children.entries()) {
|
for (const [parentId, childIds] of groups.children.entries()) {
|
||||||
@@ -133,8 +83,7 @@ export function applyTargetedLayout(
|
|||||||
changedSet,
|
changedSet,
|
||||||
verticalSpacing,
|
verticalSpacing,
|
||||||
horizontalSpacing,
|
horizontalSpacing,
|
||||||
subflowDepths,
|
subflowDepths
|
||||||
blockGroups
|
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -143,7 +92,6 @@ export function applyTargetedLayout(
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Layouts a group of blocks (either root level or within a container)
|
* Layouts a group of blocks (either root level or within a container)
|
||||||
* Blocks in block groups are moved together as a unit.
|
|
||||||
*/
|
*/
|
||||||
function layoutGroup(
|
function layoutGroup(
|
||||||
parentId: string | null,
|
parentId: string | null,
|
||||||
@@ -153,8 +101,7 @@ function layoutGroup(
|
|||||||
changedSet: Set<string>,
|
changedSet: Set<string>,
|
||||||
verticalSpacing: number,
|
verticalSpacing: number,
|
||||||
horizontalSpacing: number,
|
horizontalSpacing: number,
|
||||||
subflowDepths: Map<string, number>,
|
subflowDepths: Map<string, number>
|
||||||
blockGroups: Map<string, string[]>
|
|
||||||
): void {
|
): void {
|
||||||
if (childIds.length === 0) return
|
if (childIds.length === 0) return
|
||||||
|
|
||||||
@@ -194,7 +141,7 @@ function layoutGroup(
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Store old positions for anchor calculation and group delta tracking
|
// Store old positions for anchor calculation
|
||||||
const oldPositions = new Map<string, { x: number; y: number }>()
|
const oldPositions = new Map<string, { x: number; y: number }>()
|
||||||
for (const id of layoutEligibleChildIds) {
|
for (const id of layoutEligibleChildIds) {
|
||||||
const block = blocks[id]
|
const block = blocks[id]
|
||||||
@@ -238,47 +185,14 @@ function layoutGroup(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Track which groups have already had their deltas applied
|
|
||||||
const processedGroups = new Set<string>()
|
|
||||||
|
|
||||||
// Apply new positions only to blocks that need layout
|
// Apply new positions only to blocks that need layout
|
||||||
for (const id of needsLayout) {
|
for (const id of needsLayout) {
|
||||||
const block = blocks[id]
|
const block = blocks[id]
|
||||||
const newPos = layoutPositions.get(id)
|
const newPos = layoutPositions.get(id)
|
||||||
if (!block || !newPos) continue
|
if (!block || !newPos) continue
|
||||||
|
block.position = {
|
||||||
const groupId = block.data?.groupId
|
x: newPos.x + offsetX,
|
||||||
|
y: newPos.y + offsetY,
|
||||||
// If this block is in a group, move all blocks in the group together
|
|
||||||
if (groupId && blockGroups.has(groupId) && !processedGroups.has(groupId)) {
|
|
||||||
processedGroups.add(groupId)
|
|
||||||
|
|
||||||
// Calculate the delta for this block (the one that needs layout)
|
|
||||||
const oldPos = oldPositions.get(id)
|
|
||||||
if (oldPos) {
|
|
||||||
const deltaX = newPos.x + offsetX - oldPos.x
|
|
||||||
const deltaY = newPos.y + offsetY - oldPos.y
|
|
||||||
|
|
||||||
// Apply delta to ALL blocks in the group using their original positions
|
|
||||||
for (const groupBlockId of blockGroups.get(groupId)!) {
|
|
||||||
const groupBlock = blocks[groupBlockId]
|
|
||||||
if (groupBlock && layoutEligibleChildIds.includes(groupBlockId)) {
|
|
||||||
const groupOriginalPos = oldPositions.get(groupBlockId)
|
|
||||||
if (groupOriginalPos) {
|
|
||||||
groupBlock.position = {
|
|
||||||
x: groupOriginalPos.x + deltaX,
|
|
||||||
y: groupOriginalPos.y + deltaY,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else if (!groupId) {
|
|
||||||
// Non-grouped block - apply position normally
|
|
||||||
block.position = {
|
|
||||||
x: newPos.x + offsetX,
|
|
||||||
y: newPos.y + offsetY,
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -41,18 +41,11 @@ export function isContainerType(blockType: string): boolean {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Checks if a block should be excluded from autolayout.
|
* Checks if a block should be excluded from autolayout
|
||||||
* Note blocks are excluded unless they are part of a group.
|
|
||||||
*/
|
*/
|
||||||
export function shouldSkipAutoLayout(block?: BlockState, isInGroup?: boolean): boolean {
|
export function shouldSkipAutoLayout(block?: BlockState): boolean {
|
||||||
if (!block) return true
|
if (!block) return true
|
||||||
// If the block type is normally excluded (e.g., note), but it's in a group, include it
|
return AUTO_LAYOUT_EXCLUDED_TYPES.has(block.type)
|
||||||
if (AUTO_LAYOUT_EXCLUDED_TYPES.has(block.type)) {
|
|
||||||
// Check if block is in a group - if so, include it in layout
|
|
||||||
const blockIsInGroup = isInGroup ?? !!block.data?.groupId
|
|
||||||
return !blockIsInGroup
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|||||||
@@ -2290,7 +2290,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
webhookId: { value: null },
|
webhookId: { value: null },
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
@@ -2302,7 +2302,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
webhookId: { value: 'wh_123456' },
|
webhookId: { value: 'wh_123456' },
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
@@ -2318,7 +2318,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
triggerPath: { value: '' },
|
triggerPath: { value: '' },
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
@@ -2330,7 +2330,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
triggerPath: { value: '/api/webhooks/abc123' },
|
triggerPath: { value: '/api/webhooks/abc123' },
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
@@ -2346,7 +2346,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
webhookId: { value: null },
|
webhookId: { value: null },
|
||||||
triggerPath: { value: '' },
|
triggerPath: { value: '' },
|
||||||
},
|
},
|
||||||
@@ -2359,7 +2359,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
webhookId: { value: 'wh_123456' },
|
webhookId: { value: 'wh_123456' },
|
||||||
triggerPath: { value: '/api/webhooks/abc123' },
|
triggerPath: { value: '/api/webhooks/abc123' },
|
||||||
},
|
},
|
||||||
@@ -2371,14 +2371,18 @@ describe('hasWorkflowChanged', () => {
|
|||||||
})
|
})
|
||||||
|
|
||||||
it.concurrent(
|
it.concurrent(
|
||||||
'should detect change when triggerConfig differs but runtime metadata also differs',
|
'should detect change when actual config differs but runtime metadata also differs',
|
||||||
() => {
|
() => {
|
||||||
|
// Test that when a real config field changes along with runtime metadata,
|
||||||
|
// the change is still detected. Using 'model' as the config field since
|
||||||
|
// triggerConfig is now excluded from comparison (individual trigger fields
|
||||||
|
// are compared separately).
|
||||||
const deployedState = createWorkflowState({
|
const deployedState = createWorkflowState({
|
||||||
blocks: {
|
blocks: {
|
||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
webhookId: { value: null },
|
webhookId: { value: null },
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
@@ -2390,7 +2394,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'pull_request' } },
|
model: { value: 'gpt-4o' },
|
||||||
webhookId: { value: 'wh_123456' },
|
webhookId: { value: 'wh_123456' },
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
@@ -2402,8 +2406,12 @@ describe('hasWorkflowChanged', () => {
|
|||||||
)
|
)
|
||||||
|
|
||||||
it.concurrent(
|
it.concurrent(
|
||||||
'should not detect change when runtime metadata is added to current state',
|
'should not detect change when triggerConfig differs (individual fields compared separately)',
|
||||||
() => {
|
() => {
|
||||||
|
// triggerConfig is excluded from comparison because:
|
||||||
|
// 1. Individual trigger fields are stored as separate subblocks and compared individually
|
||||||
|
// 2. The client populates triggerConfig with default values from trigger definitions,
|
||||||
|
// which aren't present in the deployed state, causing false positive change detection
|
||||||
const deployedState = createWorkflowState({
|
const deployedState = createWorkflowState({
|
||||||
blocks: {
|
blocks: {
|
||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
@@ -2420,7 +2428,36 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
triggerConfig: { value: { event: 'pull_request', extraField: true } },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(hasWorkflowChanged(currentState, deployedState)).toBe(false)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
it.concurrent(
|
||||||
|
'should not detect change when runtime metadata is added to current state',
|
||||||
|
() => {
|
||||||
|
const deployedState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
model: { value: 'gpt-4' },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const currentState = createWorkflowState({
|
||||||
|
blocks: {
|
||||||
|
block1: createBlock('block1', {
|
||||||
|
type: 'starter',
|
||||||
|
subBlocks: {
|
||||||
|
model: { value: 'gpt-4' },
|
||||||
webhookId: { value: 'wh_123456' },
|
webhookId: { value: 'wh_123456' },
|
||||||
triggerPath: { value: '/api/webhooks/abc123' },
|
triggerPath: { value: '/api/webhooks/abc123' },
|
||||||
},
|
},
|
||||||
@@ -2440,7 +2477,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
webhookId: { value: 'wh_old123' },
|
webhookId: { value: 'wh_old123' },
|
||||||
triggerPath: { value: '/api/webhooks/old' },
|
triggerPath: { value: '/api/webhooks/old' },
|
||||||
},
|
},
|
||||||
@@ -2453,7 +2490,7 @@ describe('hasWorkflowChanged', () => {
|
|||||||
block1: createBlock('block1', {
|
block1: createBlock('block1', {
|
||||||
type: 'starter',
|
type: 'starter',
|
||||||
subBlocks: {
|
subBlocks: {
|
||||||
triggerConfig: { value: { event: 'push' } },
|
model: { value: 'gpt-4' },
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -1174,6 +1174,5 @@ export function stripWorkflowDiffMarkers(state: WorkflowState): WorkflowState {
|
|||||||
edges: structuredClone(state.edges || []),
|
edges: structuredClone(state.edges || []),
|
||||||
loops: structuredClone(state.loops || {}),
|
loops: structuredClone(state.loops || {}),
|
||||||
parallels: structuredClone(state.parallels || {}),
|
parallels: structuredClone(state.parallels || {}),
|
||||||
groups: structuredClone(state.groups || {}),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -104,6 +104,7 @@
|
|||||||
"groq-sdk": "^0.15.0",
|
"groq-sdk": "^0.15.0",
|
||||||
"html-to-image": "1.11.13",
|
"html-to-image": "1.11.13",
|
||||||
"html-to-text": "^9.0.5",
|
"html-to-text": "^9.0.5",
|
||||||
|
"idb-keyval": "6.2.2",
|
||||||
"imapflow": "1.2.4",
|
"imapflow": "1.2.4",
|
||||||
"input-otp": "^1.4.2",
|
"input-otp": "^1.4.2",
|
||||||
"ioredis": "^5.6.0",
|
"ioredis": "^5.6.0",
|
||||||
|
|||||||
@@ -16,8 +16,6 @@ export const BLOCKS_OPERATIONS = {
|
|||||||
BATCH_TOGGLE_ENABLED: 'batch-toggle-enabled',
|
BATCH_TOGGLE_ENABLED: 'batch-toggle-enabled',
|
||||||
BATCH_TOGGLE_HANDLES: 'batch-toggle-handles',
|
BATCH_TOGGLE_HANDLES: 'batch-toggle-handles',
|
||||||
BATCH_UPDATE_PARENT: 'batch-update-parent',
|
BATCH_UPDATE_PARENT: 'batch-update-parent',
|
||||||
GROUP_BLOCKS: 'group-blocks',
|
|
||||||
UNGROUP_BLOCKS: 'ungroup-blocks',
|
|
||||||
} as const
|
} as const
|
||||||
|
|
||||||
export type BlocksOperation = (typeof BLOCKS_OPERATIONS)[keyof typeof BLOCKS_OPERATIONS]
|
export type BlocksOperation = (typeof BLOCKS_OPERATIONS)[keyof typeof BLOCKS_OPERATIONS]
|
||||||
@@ -89,8 +87,6 @@ export const UNDO_REDO_OPERATIONS = {
|
|||||||
APPLY_DIFF: 'apply-diff',
|
APPLY_DIFF: 'apply-diff',
|
||||||
ACCEPT_DIFF: 'accept-diff',
|
ACCEPT_DIFF: 'accept-diff',
|
||||||
REJECT_DIFF: 'reject-diff',
|
REJECT_DIFF: 'reject-diff',
|
||||||
GROUP_BLOCKS: 'group-blocks',
|
|
||||||
UNGROUP_BLOCKS: 'ungroup-blocks',
|
|
||||||
} as const
|
} as const
|
||||||
|
|
||||||
export type UndoRedoOperation = (typeof UNDO_REDO_OPERATIONS)[keyof typeof UNDO_REDO_OPERATIONS]
|
export type UndoRedoOperation = (typeof UNDO_REDO_OPERATIONS)[keyof typeof UNDO_REDO_OPERATIONS]
|
||||||
|
|||||||
@@ -810,104 +810,6 @@ async function handleBlocksOperationTx(
|
|||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
case BLOCKS_OPERATIONS.GROUP_BLOCKS: {
|
|
||||||
const { blockIds, groupId } = payload
|
|
||||||
if (!Array.isArray(blockIds) || blockIds.length === 0 || !groupId) {
|
|
||||||
logger.debug('Invalid payload for group blocks operation')
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.info(`Grouping ${blockIds.length} blocks into group ${groupId} in workflow ${workflowId}`)
|
|
||||||
|
|
||||||
// Update blocks: set groupId and push to groupStack
|
|
||||||
for (const blockId of blockIds) {
|
|
||||||
const [currentBlock] = await tx
|
|
||||||
.select({ data: workflowBlocks.data })
|
|
||||||
.from(workflowBlocks)
|
|
||||||
.where(and(eq(workflowBlocks.id, blockId), eq(workflowBlocks.workflowId, workflowId)))
|
|
||||||
.limit(1)
|
|
||||||
|
|
||||||
if (!currentBlock) {
|
|
||||||
logger.warn(`Block ${blockId} not found for grouping`)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
const currentData = (currentBlock?.data || {}) as Record<string, any>
|
|
||||||
const currentStack = Array.isArray(currentData.groupStack) ? currentData.groupStack : []
|
|
||||||
const updatedData = {
|
|
||||||
...currentData,
|
|
||||||
groupId,
|
|
||||||
groupStack: [...currentStack, groupId],
|
|
||||||
}
|
|
||||||
|
|
||||||
await tx
|
|
||||||
.update(workflowBlocks)
|
|
||||||
.set({
|
|
||||||
data: updatedData,
|
|
||||||
updatedAt: new Date(),
|
|
||||||
})
|
|
||||||
.where(and(eq(workflowBlocks.id, blockId), eq(workflowBlocks.workflowId, workflowId)))
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.debug(`Grouped ${blockIds.length} blocks into group ${groupId}`)
|
|
||||||
break
|
|
||||||
}
|
|
||||||
|
|
||||||
case BLOCKS_OPERATIONS.UNGROUP_BLOCKS: {
|
|
||||||
const { groupId, blockIds } = payload
|
|
||||||
if (!groupId || !Array.isArray(blockIds)) {
|
|
||||||
logger.debug('Invalid payload for ungroup blocks operation')
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.info(`Ungrouping ${blockIds.length} blocks from group ${groupId} in workflow ${workflowId}`)
|
|
||||||
|
|
||||||
// Update blocks: pop from groupStack and set groupId to the previous level
|
|
||||||
for (const blockId of blockIds) {
|
|
||||||
const [currentBlock] = await tx
|
|
||||||
.select({ data: workflowBlocks.data })
|
|
||||||
.from(workflowBlocks)
|
|
||||||
.where(and(eq(workflowBlocks.id, blockId), eq(workflowBlocks.workflowId, workflowId)))
|
|
||||||
.limit(1)
|
|
||||||
|
|
||||||
if (!currentBlock) {
|
|
||||||
logger.warn(`Block ${blockId} not found for ungrouping`)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
const currentData = (currentBlock?.data || {}) as Record<string, any>
|
|
||||||
const currentStack = Array.isArray(currentData.groupStack) ? [...currentData.groupStack] : []
|
|
||||||
|
|
||||||
// Pop the current groupId from the stack
|
|
||||||
if (currentStack.length > 0 && currentStack[currentStack.length - 1] === groupId) {
|
|
||||||
currentStack.pop()
|
|
||||||
}
|
|
||||||
|
|
||||||
// The new groupId is the top of the remaining stack, or undefined if empty
|
|
||||||
const newGroupId = currentStack.length > 0 ? currentStack[currentStack.length - 1] : undefined
|
|
||||||
|
|
||||||
let updatedData: Record<string, any>
|
|
||||||
if (newGroupId) {
|
|
||||||
updatedData = { ...currentData, groupId: newGroupId, groupStack: currentStack }
|
|
||||||
} else {
|
|
||||||
// Remove groupId and groupStack if stack is empty
|
|
||||||
const { groupId: _removed, groupStack: _removedStack, ...restData } = currentData
|
|
||||||
updatedData = restData
|
|
||||||
}
|
|
||||||
|
|
||||||
await tx
|
|
||||||
.update(workflowBlocks)
|
|
||||||
.set({
|
|
||||||
data: updatedData,
|
|
||||||
updatedAt: new Date(),
|
|
||||||
})
|
|
||||||
.where(and(eq(workflowBlocks.id, blockId), eq(workflowBlocks.workflowId, workflowId)))
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.debug(`Ungrouped ${blockIds.length} blocks from group ${groupId}`)
|
|
||||||
break
|
|
||||||
}
|
|
||||||
|
|
||||||
default:
|
default:
|
||||||
throw new Error(`Unsupported blocks operation: ${operation}`)
|
throw new Error(`Unsupported blocks operation: ${operation}`)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -465,70 +465,6 @@ export function setupOperationsHandlers(
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
if (
|
|
||||||
target === OPERATION_TARGETS.BLOCKS &&
|
|
||||||
operation === BLOCKS_OPERATIONS.GROUP_BLOCKS
|
|
||||||
) {
|
|
||||||
await persistWorkflowOperation(workflowId, {
|
|
||||||
operation,
|
|
||||||
target,
|
|
||||||
payload,
|
|
||||||
timestamp: operationTimestamp,
|
|
||||||
userId: session.userId,
|
|
||||||
})
|
|
||||||
|
|
||||||
room.lastModified = Date.now()
|
|
||||||
|
|
||||||
socket.to(workflowId).emit('workflow-operation', {
|
|
||||||
operation,
|
|
||||||
target,
|
|
||||||
payload,
|
|
||||||
timestamp: operationTimestamp,
|
|
||||||
senderId: socket.id,
|
|
||||||
userId: session.userId,
|
|
||||||
userName: session.userName,
|
|
||||||
metadata: { workflowId, operationId: crypto.randomUUID() },
|
|
||||||
})
|
|
||||||
|
|
||||||
if (operationId) {
|
|
||||||
socket.emit('operation-confirmed', { operationId, serverTimestamp: Date.now() })
|
|
||||||
}
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if (
|
|
||||||
target === OPERATION_TARGETS.BLOCKS &&
|
|
||||||
operation === BLOCKS_OPERATIONS.UNGROUP_BLOCKS
|
|
||||||
) {
|
|
||||||
await persistWorkflowOperation(workflowId, {
|
|
||||||
operation,
|
|
||||||
target,
|
|
||||||
payload,
|
|
||||||
timestamp: operationTimestamp,
|
|
||||||
userId: session.userId,
|
|
||||||
})
|
|
||||||
|
|
||||||
room.lastModified = Date.now()
|
|
||||||
|
|
||||||
socket.to(workflowId).emit('workflow-operation', {
|
|
||||||
operation,
|
|
||||||
target,
|
|
||||||
payload,
|
|
||||||
timestamp: operationTimestamp,
|
|
||||||
senderId: socket.id,
|
|
||||||
userId: session.userId,
|
|
||||||
userName: session.userName,
|
|
||||||
metadata: { workflowId, operationId: crypto.randomUUID() },
|
|
||||||
})
|
|
||||||
|
|
||||||
if (operationId) {
|
|
||||||
socket.emit('operation-confirmed', { operationId, serverTimestamp: Date.now() })
|
|
||||||
}
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if (target === OPERATION_TARGETS.EDGES && operation === EDGES_OPERATIONS.BATCH_ADD_EDGES) {
|
if (target === OPERATION_TARGETS.EDGES && operation === EDGES_OPERATIONS.BATCH_ADD_EDGES) {
|
||||||
await persistWorkflowOperation(workflowId, {
|
await persistWorkflowOperation(workflowId, {
|
||||||
operation,
|
operation,
|
||||||
|
|||||||
@@ -30,8 +30,6 @@ const WRITE_OPERATIONS: string[] = [
|
|||||||
BLOCKS_OPERATIONS.BATCH_TOGGLE_ENABLED,
|
BLOCKS_OPERATIONS.BATCH_TOGGLE_ENABLED,
|
||||||
BLOCKS_OPERATIONS.BATCH_TOGGLE_HANDLES,
|
BLOCKS_OPERATIONS.BATCH_TOGGLE_HANDLES,
|
||||||
BLOCKS_OPERATIONS.BATCH_UPDATE_PARENT,
|
BLOCKS_OPERATIONS.BATCH_UPDATE_PARENT,
|
||||||
BLOCKS_OPERATIONS.GROUP_BLOCKS,
|
|
||||||
BLOCKS_OPERATIONS.UNGROUP_BLOCKS,
|
|
||||||
// Edge operations
|
// Edge operations
|
||||||
EDGE_OPERATIONS.ADD,
|
EDGE_OPERATIONS.ADD,
|
||||||
EDGE_OPERATIONS.REMOVE,
|
EDGE_OPERATIONS.REMOVE,
|
||||||
|
|||||||
@@ -221,30 +221,6 @@ export const BatchUpdateParentSchema = z.object({
|
|||||||
operationId: z.string().optional(),
|
operationId: z.string().optional(),
|
||||||
})
|
})
|
||||||
|
|
||||||
export const GroupBlocksSchema = z.object({
|
|
||||||
operation: z.literal(BLOCKS_OPERATIONS.GROUP_BLOCKS),
|
|
||||||
target: z.literal(OPERATION_TARGETS.BLOCKS),
|
|
||||||
payload: z.object({
|
|
||||||
blockIds: z.array(z.string()),
|
|
||||||
groupId: z.string(),
|
|
||||||
name: z.string().optional(),
|
|
||||||
}),
|
|
||||||
timestamp: z.number(),
|
|
||||||
operationId: z.string().optional(),
|
|
||||||
})
|
|
||||||
|
|
||||||
export const UngroupBlocksSchema = z.object({
|
|
||||||
operation: z.literal(BLOCKS_OPERATIONS.UNGROUP_BLOCKS),
|
|
||||||
target: z.literal(OPERATION_TARGETS.BLOCKS),
|
|
||||||
payload: z.object({
|
|
||||||
groupId: z.string(),
|
|
||||||
blockIds: z.array(z.string()),
|
|
||||||
parentGroupId: z.string().optional(),
|
|
||||||
}),
|
|
||||||
timestamp: z.number(),
|
|
||||||
operationId: z.string().optional(),
|
|
||||||
})
|
|
||||||
|
|
||||||
export const WorkflowOperationSchema = z.union([
|
export const WorkflowOperationSchema = z.union([
|
||||||
BlockOperationSchema,
|
BlockOperationSchema,
|
||||||
BatchPositionUpdateSchema,
|
BatchPositionUpdateSchema,
|
||||||
@@ -253,8 +229,6 @@ export const WorkflowOperationSchema = z.union([
|
|||||||
BatchToggleEnabledSchema,
|
BatchToggleEnabledSchema,
|
||||||
BatchToggleHandlesSchema,
|
BatchToggleHandlesSchema,
|
||||||
BatchUpdateParentSchema,
|
BatchUpdateParentSchema,
|
||||||
GroupBlocksSchema,
|
|
||||||
UngroupBlocksSchema,
|
|
||||||
EdgeOperationSchema,
|
EdgeOperationSchema,
|
||||||
BatchAddEdgesSchema,
|
BatchAddEdgesSchema,
|
||||||
BatchRemoveEdgesSchema,
|
BatchRemoveEdgesSchema,
|
||||||
|
|||||||
@@ -25,7 +25,6 @@ function captureWorkflowSnapshot(): WorkflowState {
|
|||||||
edges: rawState.edges || [],
|
edges: rawState.edges || [],
|
||||||
loops: rawState.loops || {},
|
loops: rawState.loops || {},
|
||||||
parallels: rawState.parallels || {},
|
parallels: rawState.parallels || {},
|
||||||
groups: rawState.groups || {},
|
|
||||||
lastSaved: Date.now(),
|
lastSaved: Date.now(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -422,7 +422,8 @@ function abortAllInProgressTools(set: any, get: () => CopilotStore) {
|
|||||||
* Loads messages from DB for UI rendering.
|
* Loads messages from DB for UI rendering.
|
||||||
* Messages are stored exactly as they render, so we just need to:
|
* Messages are stored exactly as they render, so we just need to:
|
||||||
* 1. Register client tool instances for any tool calls
|
* 1. Register client tool instances for any tool calls
|
||||||
* 2. Return the messages as-is
|
* 2. Clear any streaming flags (messages loaded from DB are never actively streaming)
|
||||||
|
* 3. Return the messages
|
||||||
*/
|
*/
|
||||||
function normalizeMessagesForUI(messages: CopilotMessage[]): CopilotMessage[] {
|
function normalizeMessagesForUI(messages: CopilotMessage[]): CopilotMessage[] {
|
||||||
try {
|
try {
|
||||||
@@ -438,23 +439,57 @@ function normalizeMessagesForUI(messages: CopilotMessage[]): CopilotMessage[] {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Register client tool instances for all tool calls so they can be looked up
|
// Register client tool instances and clear streaming flags for all tool calls
|
||||||
for (const message of messages) {
|
for (const message of messages) {
|
||||||
|
// Clear from contentBlocks (current format)
|
||||||
if (message.contentBlocks) {
|
if (message.contentBlocks) {
|
||||||
for (const block of message.contentBlocks as any[]) {
|
for (const block of message.contentBlocks as any[]) {
|
||||||
if (block?.type === 'tool_call' && block.toolCall) {
|
if (block?.type === 'tool_call' && block.toolCall) {
|
||||||
registerToolCallInstances(block.toolCall)
|
registerToolCallInstances(block.toolCall)
|
||||||
|
clearStreamingFlags(block.toolCall)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
// Clear from toolCalls array (legacy format)
|
||||||
|
if (message.toolCalls) {
|
||||||
|
for (const toolCall of message.toolCalls) {
|
||||||
|
registerToolCallInstances(toolCall)
|
||||||
|
clearStreamingFlags(toolCall)
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
// Return messages as-is - they're already in the correct format for rendering
|
// Return messages - they're already in the correct format for rendering
|
||||||
return messages
|
return messages
|
||||||
} catch {
|
} catch {
|
||||||
return messages
|
return messages
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Recursively clears streaming flags from a tool call and its nested subagent tool calls.
|
||||||
|
* This ensures messages loaded from DB don't appear to be streaming.
|
||||||
|
*/
|
||||||
|
function clearStreamingFlags(toolCall: any): void {
|
||||||
|
if (!toolCall) return
|
||||||
|
|
||||||
|
// Always set subAgentStreaming to false - messages loaded from DB are never streaming
|
||||||
|
toolCall.subAgentStreaming = false
|
||||||
|
|
||||||
|
// Clear nested subagent tool calls
|
||||||
|
if (Array.isArray(toolCall.subAgentBlocks)) {
|
||||||
|
for (const block of toolCall.subAgentBlocks) {
|
||||||
|
if (block?.type === 'subagent_tool_call' && block.toolCall) {
|
||||||
|
clearStreamingFlags(block.toolCall)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (Array.isArray(toolCall.subAgentToolCalls)) {
|
||||||
|
for (const subTc of toolCall.subAgentToolCalls) {
|
||||||
|
clearStreamingFlags(subTc)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Recursively registers client tool instances for a tool call and its nested subagent tool calls.
|
* Recursively registers client tool instances for a tool call and its nested subagent tool calls.
|
||||||
*/
|
*/
|
||||||
|
|||||||
@@ -1,2 +1,3 @@
|
|||||||
|
export { indexedDBStorage } from './storage'
|
||||||
export { useTerminalConsoleStore } from './store'
|
export { useTerminalConsoleStore } from './store'
|
||||||
export type { ConsoleEntry, ConsoleStore, ConsoleUpdate } from './types'
|
export type { ConsoleEntry, ConsoleStore, ConsoleUpdate } from './types'
|
||||||
|
|||||||
81
apps/sim/stores/terminal/console/storage.ts
Normal file
81
apps/sim/stores/terminal/console/storage.ts
Normal file
@@ -0,0 +1,81 @@
|
|||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import { del, get, set } from 'idb-keyval'
|
||||||
|
import type { StateStorage } from 'zustand/middleware'
|
||||||
|
|
||||||
|
const logger = createLogger('ConsoleStorage')
|
||||||
|
|
||||||
|
const STORE_KEY = 'terminal-console-store'
|
||||||
|
const MIGRATION_KEY = 'terminal-console-store-migrated'
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Promise that resolves when migration is complete.
|
||||||
|
* Used to ensure getItem waits for migration before reading.
|
||||||
|
*/
|
||||||
|
let migrationPromise: Promise<void> | null = null
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Migrates existing console data from localStorage to IndexedDB.
|
||||||
|
* Runs once on first load, then marks migration as complete.
|
||||||
|
*/
|
||||||
|
async function migrateFromLocalStorage(): Promise<void> {
|
||||||
|
if (typeof window === 'undefined') return
|
||||||
|
|
||||||
|
try {
|
||||||
|
const migrated = await get<boolean>(MIGRATION_KEY)
|
||||||
|
if (migrated) return
|
||||||
|
|
||||||
|
const localData = localStorage.getItem(STORE_KEY)
|
||||||
|
if (localData) {
|
||||||
|
await set(STORE_KEY, localData)
|
||||||
|
localStorage.removeItem(STORE_KEY)
|
||||||
|
logger.info('Migrated console store to IndexedDB')
|
||||||
|
}
|
||||||
|
|
||||||
|
await set(MIGRATION_KEY, true)
|
||||||
|
} catch (error) {
|
||||||
|
logger.warn('Migration from localStorage failed', { error })
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (typeof window !== 'undefined') {
|
||||||
|
migrationPromise = migrateFromLocalStorage().finally(() => {
|
||||||
|
migrationPromise = null
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
export const indexedDBStorage: StateStorage = {
|
||||||
|
getItem: async (name: string): Promise<string | null> => {
|
||||||
|
if (typeof window === 'undefined') return null
|
||||||
|
|
||||||
|
// Ensure migration completes before reading
|
||||||
|
if (migrationPromise) {
|
||||||
|
await migrationPromise
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
const value = await get<string>(name)
|
||||||
|
return value ?? null
|
||||||
|
} catch (error) {
|
||||||
|
logger.warn('IndexedDB read failed', { name, error })
|
||||||
|
return null
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
setItem: async (name: string, value: string): Promise<void> => {
|
||||||
|
if (typeof window === 'undefined') return
|
||||||
|
try {
|
||||||
|
await set(name, value)
|
||||||
|
} catch (error) {
|
||||||
|
logger.warn('IndexedDB write failed', { name, error })
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
removeItem: async (name: string): Promise<void> => {
|
||||||
|
if (typeof window === 'undefined') return
|
||||||
|
try {
|
||||||
|
await del(name)
|
||||||
|
} catch (error) {
|
||||||
|
logger.warn('IndexedDB delete failed', { name, error })
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
@@ -1,18 +1,22 @@
|
|||||||
import { createLogger } from '@sim/logger'
|
import { createLogger } from '@sim/logger'
|
||||||
import { create } from 'zustand'
|
import { create } from 'zustand'
|
||||||
import { devtools, persist } from 'zustand/middleware'
|
import { createJSONStorage, devtools, persist } from 'zustand/middleware'
|
||||||
import { redactApiKeys } from '@/lib/core/security/redaction'
|
import { redactApiKeys } from '@/lib/core/security/redaction'
|
||||||
import type { NormalizedBlockOutput } from '@/executor/types'
|
import type { NormalizedBlockOutput } from '@/executor/types'
|
||||||
import { useExecutionStore } from '@/stores/execution'
|
import { useExecutionStore } from '@/stores/execution'
|
||||||
import { useNotificationStore } from '@/stores/notifications'
|
import { useNotificationStore } from '@/stores/notifications'
|
||||||
import { useGeneralStore } from '@/stores/settings/general'
|
import { useGeneralStore } from '@/stores/settings/general'
|
||||||
|
import { indexedDBStorage } from '@/stores/terminal/console/storage'
|
||||||
import type { ConsoleEntry, ConsoleStore, ConsoleUpdate } from '@/stores/terminal/console/types'
|
import type { ConsoleEntry, ConsoleStore, ConsoleUpdate } from '@/stores/terminal/console/types'
|
||||||
|
|
||||||
const logger = createLogger('TerminalConsoleStore')
|
const logger = createLogger('TerminalConsoleStore')
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Updates a NormalizedBlockOutput with new content
|
* Maximum number of console entries to keep per workflow.
|
||||||
|
* Keeps the stored data size reasonable and improves performance.
|
||||||
*/
|
*/
|
||||||
|
const MAX_ENTRIES_PER_WORKFLOW = 500
|
||||||
|
|
||||||
const updateBlockOutput = (
|
const updateBlockOutput = (
|
||||||
existingOutput: NormalizedBlockOutput | undefined,
|
existingOutput: NormalizedBlockOutput | undefined,
|
||||||
contentUpdate: string
|
contentUpdate: string
|
||||||
@@ -23,9 +27,6 @@ const updateBlockOutput = (
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Checks if output represents a streaming object that should be skipped
|
|
||||||
*/
|
|
||||||
const isStreamingOutput = (output: any): boolean => {
|
const isStreamingOutput = (output: any): boolean => {
|
||||||
if (typeof ReadableStream !== 'undefined' && output instanceof ReadableStream) {
|
if (typeof ReadableStream !== 'undefined' && output instanceof ReadableStream) {
|
||||||
return true
|
return true
|
||||||
@@ -44,9 +45,6 @@ const isStreamingOutput = (output: any): boolean => {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Checks if entry should be skipped to prevent duplicates
|
|
||||||
*/
|
|
||||||
const shouldSkipEntry = (output: any): boolean => {
|
const shouldSkipEntry = (output: any): boolean => {
|
||||||
if (typeof output !== 'object' || !output) {
|
if (typeof output !== 'object' || !output) {
|
||||||
return false
|
return false
|
||||||
@@ -69,6 +67,9 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
|||||||
(set, get) => ({
|
(set, get) => ({
|
||||||
entries: [],
|
entries: [],
|
||||||
isOpen: false,
|
isOpen: false,
|
||||||
|
_hasHydrated: false,
|
||||||
|
|
||||||
|
setHasHydrated: (hasHydrated) => set({ _hasHydrated: hasHydrated }),
|
||||||
|
|
||||||
addConsole: (entry: Omit<ConsoleEntry, 'id' | 'timestamp'>) => {
|
addConsole: (entry: Omit<ConsoleEntry, 'id' | 'timestamp'>) => {
|
||||||
set((state) => {
|
set((state) => {
|
||||||
@@ -94,7 +95,15 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
|||||||
timestamp: new Date().toISOString(),
|
timestamp: new Date().toISOString(),
|
||||||
}
|
}
|
||||||
|
|
||||||
return { entries: [newEntry, ...state.entries] }
|
const newEntries = [newEntry, ...state.entries]
|
||||||
|
const workflowCounts = new Map<string, number>()
|
||||||
|
const trimmedEntries = newEntries.filter((entry) => {
|
||||||
|
const count = workflowCounts.get(entry.workflowId) || 0
|
||||||
|
if (count >= MAX_ENTRIES_PER_WORKFLOW) return false
|
||||||
|
workflowCounts.set(entry.workflowId, count + 1)
|
||||||
|
return true
|
||||||
|
})
|
||||||
|
return { entries: trimmedEntries }
|
||||||
})
|
})
|
||||||
|
|
||||||
const newEntry = get().entries[0]
|
const newEntry = get().entries[0]
|
||||||
@@ -130,10 +139,6 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
|||||||
return newEntry
|
return newEntry
|
||||||
},
|
},
|
||||||
|
|
||||||
/**
|
|
||||||
* Clears console entries for a specific workflow and clears the run path
|
|
||||||
* @param workflowId - The workflow ID to clear entries for
|
|
||||||
*/
|
|
||||||
clearWorkflowConsole: (workflowId: string) => {
|
clearWorkflowConsole: (workflowId: string) => {
|
||||||
set((state) => ({
|
set((state) => ({
|
||||||
entries: state.entries.filter((entry) => entry.workflowId !== workflowId),
|
entries: state.entries.filter((entry) => entry.workflowId !== workflowId),
|
||||||
@@ -148,9 +153,6 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Formats a value for CSV export
|
|
||||||
*/
|
|
||||||
const formatCSVValue = (value: any): string => {
|
const formatCSVValue = (value: any): string => {
|
||||||
if (value === null || value === undefined) {
|
if (value === null || value === undefined) {
|
||||||
return ''
|
return ''
|
||||||
@@ -297,7 +299,35 @@ export const useTerminalConsoleStore = create<ConsoleStore>()(
|
|||||||
}),
|
}),
|
||||||
{
|
{
|
||||||
name: 'terminal-console-store',
|
name: 'terminal-console-store',
|
||||||
|
storage: createJSONStorage(() => indexedDBStorage),
|
||||||
|
partialize: (state) => ({
|
||||||
|
entries: state.entries,
|
||||||
|
isOpen: state.isOpen,
|
||||||
|
}),
|
||||||
|
onRehydrateStorage: () => (_state, error) => {
|
||||||
|
if (error) {
|
||||||
|
logger.error('Failed to rehydrate console store', { error })
|
||||||
|
}
|
||||||
|
},
|
||||||
|
merge: (persistedState, currentState) => {
|
||||||
|
const persisted = persistedState as Partial<ConsoleStore> | undefined
|
||||||
|
return {
|
||||||
|
...currentState,
|
||||||
|
entries: persisted?.entries ?? currentState.entries,
|
||||||
|
isOpen: persisted?.isOpen ?? currentState.isOpen,
|
||||||
|
}
|
||||||
|
},
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if (typeof window !== 'undefined') {
|
||||||
|
useTerminalConsoleStore.persist.onFinishHydration(() => {
|
||||||
|
useTerminalConsoleStore.setState({ _hasHydrated: true })
|
||||||
|
})
|
||||||
|
|
||||||
|
if (useTerminalConsoleStore.persist.hasHydrated()) {
|
||||||
|
useTerminalConsoleStore.setState({ _hasHydrated: true })
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -1,9 +1,6 @@
|
|||||||
import type { NormalizedBlockOutput } from '@/executor/types'
|
import type { NormalizedBlockOutput } from '@/executor/types'
|
||||||
import type { SubflowType } from '@/stores/workflows/workflow/types'
|
import type { SubflowType } from '@/stores/workflows/workflow/types'
|
||||||
|
|
||||||
/**
|
|
||||||
* Console entry for terminal logs
|
|
||||||
*/
|
|
||||||
export interface ConsoleEntry {
|
export interface ConsoleEntry {
|
||||||
id: string
|
id: string
|
||||||
timestamp: string
|
timestamp: string
|
||||||
@@ -25,9 +22,6 @@ export interface ConsoleEntry {
|
|||||||
iterationType?: SubflowType
|
iterationType?: SubflowType
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Console update payload for partial updates
|
|
||||||
*/
|
|
||||||
export interface ConsoleUpdate {
|
export interface ConsoleUpdate {
|
||||||
content?: string
|
content?: string
|
||||||
output?: Partial<NormalizedBlockOutput>
|
output?: Partial<NormalizedBlockOutput>
|
||||||
@@ -40,9 +34,6 @@ export interface ConsoleUpdate {
|
|||||||
input?: any
|
input?: any
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Console store state and actions
|
|
||||||
*/
|
|
||||||
export interface ConsoleStore {
|
export interface ConsoleStore {
|
||||||
entries: ConsoleEntry[]
|
entries: ConsoleEntry[]
|
||||||
isOpen: boolean
|
isOpen: boolean
|
||||||
@@ -52,4 +43,6 @@ export interface ConsoleStore {
|
|||||||
getWorkflowEntries: (workflowId: string) => ConsoleEntry[]
|
getWorkflowEntries: (workflowId: string) => ConsoleEntry[]
|
||||||
toggleConsole: () => void
|
toggleConsole: () => void
|
||||||
updateConsole: (blockId: string, update: string | ConsoleUpdate, executionId?: string) => void
|
updateConsole: (blockId: string, update: string | ConsoleUpdate, executionId?: string) => void
|
||||||
|
_hasHydrated: boolean
|
||||||
|
setHasHydrated: (hasHydrated: boolean) => void
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -126,23 +126,6 @@ export interface RejectDiffOperation extends BaseOperation {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface GroupBlocksOperation extends BaseOperation {
|
|
||||||
type: typeof UNDO_REDO_OPERATIONS.GROUP_BLOCKS
|
|
||||||
data: {
|
|
||||||
groupId: string
|
|
||||||
blockIds: string[]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
export interface UngroupBlocksOperation extends BaseOperation {
|
|
||||||
type: typeof UNDO_REDO_OPERATIONS.UNGROUP_BLOCKS
|
|
||||||
data: {
|
|
||||||
groupId: string
|
|
||||||
blockIds: string[]
|
|
||||||
parentGroupId?: string
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
export type Operation =
|
export type Operation =
|
||||||
| BatchAddBlocksOperation
|
| BatchAddBlocksOperation
|
||||||
| BatchRemoveBlocksOperation
|
| BatchRemoveBlocksOperation
|
||||||
@@ -156,8 +139,6 @@ export type Operation =
|
|||||||
| ApplyDiffOperation
|
| ApplyDiffOperation
|
||||||
| AcceptDiffOperation
|
| AcceptDiffOperation
|
||||||
| RejectDiffOperation
|
| RejectDiffOperation
|
||||||
| GroupBlocksOperation
|
|
||||||
| UngroupBlocksOperation
|
|
||||||
|
|
||||||
export interface OperationEntry {
|
export interface OperationEntry {
|
||||||
id: string
|
id: string
|
||||||
|
|||||||
@@ -6,10 +6,8 @@ import type {
|
|||||||
BatchRemoveBlocksOperation,
|
BatchRemoveBlocksOperation,
|
||||||
BatchRemoveEdgesOperation,
|
BatchRemoveEdgesOperation,
|
||||||
BatchUpdateParentOperation,
|
BatchUpdateParentOperation,
|
||||||
GroupBlocksOperation,
|
|
||||||
Operation,
|
Operation,
|
||||||
OperationEntry,
|
OperationEntry,
|
||||||
UngroupBlocksOperation,
|
|
||||||
} from '@/stores/undo-redo/types'
|
} from '@/stores/undo-redo/types'
|
||||||
|
|
||||||
export function createOperationEntry(operation: Operation, inverse: Operation): OperationEntry {
|
export function createOperationEntry(operation: Operation, inverse: Operation): OperationEntry {
|
||||||
@@ -166,30 +164,6 @@ export function createInverseOperation(operation: Operation): Operation {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
case UNDO_REDO_OPERATIONS.GROUP_BLOCKS: {
|
|
||||||
const op = operation as GroupBlocksOperation
|
|
||||||
return {
|
|
||||||
...operation,
|
|
||||||
type: UNDO_REDO_OPERATIONS.UNGROUP_BLOCKS,
|
|
||||||
data: {
|
|
||||||
groupId: op.data.groupId,
|
|
||||||
blockIds: op.data.blockIds,
|
|
||||||
},
|
|
||||||
} as UngroupBlocksOperation
|
|
||||||
}
|
|
||||||
|
|
||||||
case UNDO_REDO_OPERATIONS.UNGROUP_BLOCKS: {
|
|
||||||
const op = operation as UngroupBlocksOperation
|
|
||||||
return {
|
|
||||||
...operation,
|
|
||||||
type: UNDO_REDO_OPERATIONS.GROUP_BLOCKS,
|
|
||||||
data: {
|
|
||||||
groupId: op.data.groupId,
|
|
||||||
blockIds: op.data.blockIds,
|
|
||||||
},
|
|
||||||
} as GroupBlocksOperation
|
|
||||||
}
|
|
||||||
|
|
||||||
default: {
|
default: {
|
||||||
const exhaustiveCheck: never = operation
|
const exhaustiveCheck: never = operation
|
||||||
throw new Error(`Unhandled operation type: ${(exhaustiveCheck as Operation).type}`)
|
throw new Error(`Unhandled operation type: ${(exhaustiveCheck as Operation).type}`)
|
||||||
|
|||||||
@@ -16,7 +16,6 @@ export function cloneWorkflowState(state: WorkflowState): WorkflowState {
|
|||||||
edges: structuredClone(state.edges || []),
|
edges: structuredClone(state.edges || []),
|
||||||
loops: structuredClone(state.loops || {}),
|
loops: structuredClone(state.loops || {}),
|
||||||
parallels: structuredClone(state.parallels || {}),
|
parallels: structuredClone(state.parallels || {}),
|
||||||
groups: structuredClone(state.groups || {}),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -298,26 +298,11 @@ export const useWorkflowRegistry = create<WorkflowRegistry>()(
|
|||||||
let workflowState: any
|
let workflowState: any
|
||||||
|
|
||||||
if (workflowData?.state) {
|
if (workflowData?.state) {
|
||||||
const blocks = workflowData.state.blocks || {}
|
|
||||||
|
|
||||||
// Reconstruct groups from blocks' groupId data
|
|
||||||
const reconstructedGroups: Record<string, { id: string; blockIds: string[] }> = {}
|
|
||||||
Object.entries(blocks).forEach(([blockId, block]: [string, any]) => {
|
|
||||||
const groupId = block?.data?.groupId
|
|
||||||
if (groupId) {
|
|
||||||
if (!reconstructedGroups[groupId]) {
|
|
||||||
reconstructedGroups[groupId] = { id: groupId, blockIds: [] }
|
|
||||||
}
|
|
||||||
reconstructedGroups[groupId].blockIds.push(blockId)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
workflowState = {
|
workflowState = {
|
||||||
blocks,
|
blocks: workflowData.state.blocks || {},
|
||||||
edges: workflowData.state.edges || [],
|
edges: workflowData.state.edges || [],
|
||||||
loops: workflowData.state.loops || {},
|
loops: workflowData.state.loops || {},
|
||||||
parallels: workflowData.state.parallels || {},
|
parallels: workflowData.state.parallels || {},
|
||||||
groups: reconstructedGroups,
|
|
||||||
lastSaved: Date.now(),
|
lastSaved: Date.now(),
|
||||||
deploymentStatuses: {},
|
deploymentStatuses: {},
|
||||||
}
|
}
|
||||||
@@ -327,7 +312,6 @@ export const useWorkflowRegistry = create<WorkflowRegistry>()(
|
|||||||
edges: [],
|
edges: [],
|
||||||
loops: {},
|
loops: {},
|
||||||
parallels: {},
|
parallels: {},
|
||||||
groups: {},
|
|
||||||
deploymentStatuses: {},
|
deploymentStatuses: {},
|
||||||
lastSaved: Date.now(),
|
lastSaved: Date.now(),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,5 +1,19 @@
|
|||||||
import type { Edge } from 'reactflow'
|
import type { Edge } from 'reactflow'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
|
|
||||||
|
export function filterNewEdges(edgesToAdd: Edge[], currentEdges: Edge[]): Edge[] {
|
||||||
|
return edgesToAdd.filter((edge) => {
|
||||||
|
if (edge.source === edge.target) return false
|
||||||
|
return !currentEdges.some(
|
||||||
|
(e) =>
|
||||||
|
e.source === edge.source &&
|
||||||
|
e.sourceHandle === edge.sourceHandle &&
|
||||||
|
e.target === edge.target &&
|
||||||
|
e.targetHandle === edge.targetHandle
|
||||||
|
)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
import { getBlockOutputs } from '@/lib/workflows/blocks/block-outputs'
|
import { getBlockOutputs } from '@/lib/workflows/blocks/block-outputs'
|
||||||
import { getBlock } from '@/blocks'
|
import { getBlock } from '@/blocks'
|
||||||
import { normalizeName } from '@/executor/constants'
|
import { normalizeName } from '@/executor/constants'
|
||||||
|
|||||||
@@ -297,7 +297,7 @@ describe('workflow store', () => {
|
|||||||
expectEdgeConnects(edges, 'block-1', 'block-2')
|
expectEdgeConnects(edges, 'block-1', 'block-2')
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should not add duplicate edges', () => {
|
it('should not add duplicate connections', () => {
|
||||||
const { addBlock, batchAddEdges } = useWorkflowStore.getState()
|
const { addBlock, batchAddEdges } = useWorkflowStore.getState()
|
||||||
|
|
||||||
addBlock('block-1', 'starter', 'Start', { x: 0, y: 0 })
|
addBlock('block-1', 'starter', 'Start', { x: 0, y: 0 })
|
||||||
@@ -309,17 +309,6 @@ describe('workflow store', () => {
|
|||||||
const state = useWorkflowStore.getState()
|
const state = useWorkflowStore.getState()
|
||||||
expectEdgeCount(state, 1)
|
expectEdgeCount(state, 1)
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should prevent self-referencing edges', () => {
|
|
||||||
const { addBlock, batchAddEdges } = useWorkflowStore.getState()
|
|
||||||
|
|
||||||
addBlock('block-1', 'function', 'Self', { x: 0, y: 0 })
|
|
||||||
|
|
||||||
batchAddEdges([{ id: 'e1', source: 'block-1', target: 'block-1' }])
|
|
||||||
|
|
||||||
const state = useWorkflowStore.getState()
|
|
||||||
expectEdgeCount(state, 0)
|
|
||||||
})
|
|
||||||
})
|
})
|
||||||
|
|
||||||
describe('batchRemoveEdges', () => {
|
describe('batchRemoveEdges', () => {
|
||||||
|
|||||||
@@ -9,7 +9,12 @@ import { getBlock } from '@/blocks'
|
|||||||
import type { SubBlockConfig } from '@/blocks/types'
|
import type { SubBlockConfig } from '@/blocks/types'
|
||||||
import { useWorkflowRegistry } from '@/stores/workflows/registry/store'
|
import { useWorkflowRegistry } from '@/stores/workflows/registry/store'
|
||||||
import { useSubBlockStore } from '@/stores/workflows/subblock/store'
|
import { useSubBlockStore } from '@/stores/workflows/subblock/store'
|
||||||
import { getUniqueBlockName, mergeSubblockState, normalizeName } from '@/stores/workflows/utils'
|
import {
|
||||||
|
filterNewEdges,
|
||||||
|
getUniqueBlockName,
|
||||||
|
mergeSubblockState,
|
||||||
|
normalizeName,
|
||||||
|
} from '@/stores/workflows/utils'
|
||||||
import type {
|
import type {
|
||||||
Position,
|
Position,
|
||||||
SubBlockState,
|
SubBlockState,
|
||||||
@@ -95,7 +100,6 @@ const initialState = {
|
|||||||
edges: [],
|
edges: [],
|
||||||
loops: {},
|
loops: {},
|
||||||
parallels: {},
|
parallels: {},
|
||||||
groups: {},
|
|
||||||
lastSaved: undefined,
|
lastSaved: undefined,
|
||||||
deploymentStatuses: {},
|
deploymentStatuses: {},
|
||||||
needsRedeployment: false,
|
needsRedeployment: false,
|
||||||
@@ -497,25 +501,11 @@ export const useWorkflowStore = create<WorkflowStore>()(
|
|||||||
|
|
||||||
batchAddEdges: (edges: Edge[]) => {
|
batchAddEdges: (edges: Edge[]) => {
|
||||||
const currentEdges = get().edges
|
const currentEdges = get().edges
|
||||||
|
const filtered = filterNewEdges(edges, currentEdges)
|
||||||
const newEdges = [...currentEdges]
|
const newEdges = [...currentEdges]
|
||||||
const existingEdgeIds = new Set(currentEdges.map((e) => e.id))
|
|
||||||
// Track existing connections to prevent duplicates (same source->target)
|
|
||||||
const existingConnections = new Set(currentEdges.map((e) => `${e.source}->${e.target}`))
|
|
||||||
|
|
||||||
for (const edge of edges) {
|
for (const edge of filtered) {
|
||||||
// Skip if edge ID already exists
|
|
||||||
if (existingEdgeIds.has(edge.id)) continue
|
|
||||||
|
|
||||||
// Skip self-referencing edges
|
|
||||||
if (edge.source === edge.target) continue
|
|
||||||
|
|
||||||
// Skip if connection already exists (same source and target)
|
|
||||||
const connectionKey = `${edge.source}->${edge.target}`
|
|
||||||
if (existingConnections.has(connectionKey)) continue
|
|
||||||
|
|
||||||
// Skip if would create a cycle
|
|
||||||
if (wouldCreateCycle([...newEdges], edge.source, edge.target)) continue
|
if (wouldCreateCycle([...newEdges], edge.source, edge.target)) continue
|
||||||
|
|
||||||
newEdges.push({
|
newEdges.push({
|
||||||
id: edge.id || crypto.randomUUID(),
|
id: edge.id || crypto.randomUUID(),
|
||||||
source: edge.source,
|
source: edge.source,
|
||||||
@@ -525,8 +515,6 @@ export const useWorkflowStore = create<WorkflowStore>()(
|
|||||||
type: edge.type || 'default',
|
type: edge.type || 'default',
|
||||||
data: edge.data || {},
|
data: edge.data || {},
|
||||||
})
|
})
|
||||||
existingEdgeIds.add(edge.id)
|
|
||||||
existingConnections.add(connectionKey)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const blocks = get().blocks
|
const blocks = get().blocks
|
||||||
@@ -578,7 +566,6 @@ export const useWorkflowStore = create<WorkflowStore>()(
|
|||||||
edges: state.edges,
|
edges: state.edges,
|
||||||
loops: state.loops,
|
loops: state.loops,
|
||||||
parallels: state.parallels,
|
parallels: state.parallels,
|
||||||
groups: state.groups,
|
|
||||||
lastSaved: state.lastSaved,
|
lastSaved: state.lastSaved,
|
||||||
deploymentStatuses: state.deploymentStatuses,
|
deploymentStatuses: state.deploymentStatuses,
|
||||||
needsRedeployment: state.needsRedeployment,
|
needsRedeployment: state.needsRedeployment,
|
||||||
@@ -599,7 +586,6 @@ export const useWorkflowStore = create<WorkflowStore>()(
|
|||||||
Object.keys(workflowState.parallels || {}).length > 0
|
Object.keys(workflowState.parallels || {}).length > 0
|
||||||
? workflowState.parallels
|
? workflowState.parallels
|
||||||
: generateParallelBlocks(nextBlocks)
|
: generateParallelBlocks(nextBlocks)
|
||||||
const nextGroups = workflowState.groups || state.groups
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
...state,
|
...state,
|
||||||
@@ -607,7 +593,6 @@ export const useWorkflowStore = create<WorkflowStore>()(
|
|||||||
edges: nextEdges,
|
edges: nextEdges,
|
||||||
loops: nextLoops,
|
loops: nextLoops,
|
||||||
parallels: nextParallels,
|
parallels: nextParallels,
|
||||||
groups: nextGroups,
|
|
||||||
deploymentStatuses: workflowState.deploymentStatuses || state.deploymentStatuses,
|
deploymentStatuses: workflowState.deploymentStatuses || state.deploymentStatuses,
|
||||||
needsRedeployment:
|
needsRedeployment:
|
||||||
workflowState.needsRedeployment !== undefined
|
workflowState.needsRedeployment !== undefined
|
||||||
@@ -1337,126 +1322,6 @@ export const useWorkflowStore = create<WorkflowStore>()(
|
|||||||
getDragStartPosition: () => {
|
getDragStartPosition: () => {
|
||||||
return get().dragStartPosition || null
|
return get().dragStartPosition || null
|
||||||
},
|
},
|
||||||
|
|
||||||
groupBlocks: (blockIds: string[], groupId?: string) => {
|
|
||||||
if (blockIds.length === 0) return ''
|
|
||||||
|
|
||||||
const newGroupId = groupId || crypto.randomUUID()
|
|
||||||
const currentGroups = get().groups || {}
|
|
||||||
const currentBlocks = get().blocks
|
|
||||||
|
|
||||||
// Create the new group with all selected block IDs
|
|
||||||
const updatedGroups = { ...currentGroups }
|
|
||||||
updatedGroups[newGroupId] = {
|
|
||||||
id: newGroupId,
|
|
||||||
blockIds: [...blockIds],
|
|
||||||
}
|
|
||||||
|
|
||||||
// Update blocks: set groupId and push to groupStack
|
|
||||||
const newBlocks = { ...currentBlocks }
|
|
||||||
for (const blockId of blockIds) {
|
|
||||||
if (newBlocks[blockId]) {
|
|
||||||
const currentData = newBlocks[blockId].data || {}
|
|
||||||
const currentStack = Array.isArray(currentData.groupStack) ? currentData.groupStack : []
|
|
||||||
newBlocks[blockId] = {
|
|
||||||
...newBlocks[blockId],
|
|
||||||
data: {
|
|
||||||
...currentData,
|
|
||||||
groupId: newGroupId,
|
|
||||||
groupStack: [...currentStack, newGroupId],
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
set({
|
|
||||||
blocks: newBlocks,
|
|
||||||
groups: updatedGroups,
|
|
||||||
})
|
|
||||||
|
|
||||||
get().updateLastSaved()
|
|
||||||
logger.info('Created block group', {
|
|
||||||
groupId: newGroupId,
|
|
||||||
blockCount: blockIds.length,
|
|
||||||
})
|
|
||||||
return newGroupId
|
|
||||||
},
|
|
||||||
|
|
||||||
ungroupBlocks: (groupId: string) => {
|
|
||||||
const currentGroups = get().groups || {}
|
|
||||||
const currentBlocks = get().blocks
|
|
||||||
const group = currentGroups[groupId]
|
|
||||||
|
|
||||||
if (!group) {
|
|
||||||
logger.warn('Attempted to ungroup non-existent group', { groupId })
|
|
||||||
return []
|
|
||||||
}
|
|
||||||
|
|
||||||
const blockIds = [...group.blockIds]
|
|
||||||
|
|
||||||
// Remove the group from the groups record
|
|
||||||
const updatedGroups = { ...currentGroups }
|
|
||||||
delete updatedGroups[groupId]
|
|
||||||
|
|
||||||
// Update blocks: pop from groupStack and set groupId to the previous level
|
|
||||||
const newBlocks = { ...currentBlocks }
|
|
||||||
for (const blockId of blockIds) {
|
|
||||||
if (newBlocks[blockId]) {
|
|
||||||
const currentData = { ...newBlocks[blockId].data }
|
|
||||||
const currentStack = Array.isArray(currentData.groupStack)
|
|
||||||
? [...currentData.groupStack]
|
|
||||||
: []
|
|
||||||
|
|
||||||
// Pop the current groupId from the stack
|
|
||||||
if (currentStack.length > 0 && currentStack[currentStack.length - 1] === groupId) {
|
|
||||||
currentStack.pop()
|
|
||||||
}
|
|
||||||
|
|
||||||
// The new groupId is the top of the remaining stack, or undefined if empty
|
|
||||||
const newGroupId =
|
|
||||||
currentStack.length > 0 ? currentStack[currentStack.length - 1] : undefined
|
|
||||||
|
|
||||||
if (newGroupId) {
|
|
||||||
currentData.groupId = newGroupId
|
|
||||||
currentData.groupStack = currentStack
|
|
||||||
} else {
|
|
||||||
// Remove groupId and groupStack if stack is empty
|
|
||||||
delete currentData.groupId
|
|
||||||
delete currentData.groupStack
|
|
||||||
}
|
|
||||||
|
|
||||||
newBlocks[blockId] = {
|
|
||||||
...newBlocks[blockId],
|
|
||||||
data: currentData,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
set({
|
|
||||||
blocks: newBlocks,
|
|
||||||
groups: updatedGroups,
|
|
||||||
})
|
|
||||||
|
|
||||||
get().updateLastSaved()
|
|
||||||
logger.info('Ungrouped blocks', {
|
|
||||||
groupId,
|
|
||||||
blockCount: blockIds.length,
|
|
||||||
})
|
|
||||||
return blockIds
|
|
||||||
},
|
|
||||||
|
|
||||||
getGroupBlockIds: (groupId: string) => {
|
|
||||||
const groups = get().groups || {}
|
|
||||||
const group = groups[groupId]
|
|
||||||
|
|
||||||
if (!group) return []
|
|
||||||
|
|
||||||
return [...group.blockIds]
|
|
||||||
},
|
|
||||||
|
|
||||||
getGroups: () => {
|
|
||||||
return get().groups || {}
|
|
||||||
},
|
|
||||||
}),
|
}),
|
||||||
{ name: 'workflow-store' }
|
{ name: 'workflow-store' }
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -63,11 +63,6 @@ export interface BlockData {
|
|||||||
|
|
||||||
// Container node type (for ReactFlow node type determination)
|
// Container node type (for ReactFlow node type determination)
|
||||||
type?: string
|
type?: string
|
||||||
|
|
||||||
// Block group membership
|
|
||||||
groupId?: string
|
|
||||||
/** Stack of group IDs for hierarchical grouping (oldest to newest) */
|
|
||||||
groupStack?: string[]
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface BlockLayoutState {
|
export interface BlockLayoutState {
|
||||||
@@ -149,20 +144,6 @@ export interface Variable {
|
|||||||
value: unknown
|
value: unknown
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Represents a group of blocks on the canvas.
|
|
||||||
* Groups can be nested (a group can contain other groups via block membership).
|
|
||||||
* When a block is in a group, it stores the groupId in its data.
|
|
||||||
*/
|
|
||||||
export interface BlockGroup {
|
|
||||||
/** Unique identifier for the group */
|
|
||||||
id: string
|
|
||||||
/** Optional display name for the group */
|
|
||||||
name?: string
|
|
||||||
/** Block IDs that are direct members of this group */
|
|
||||||
blockIds: string[]
|
|
||||||
}
|
|
||||||
|
|
||||||
export interface DragStartPosition {
|
export interface DragStartPosition {
|
||||||
id: string
|
id: string
|
||||||
x: number
|
x: number
|
||||||
@@ -176,8 +157,6 @@ export interface WorkflowState {
|
|||||||
lastSaved?: number
|
lastSaved?: number
|
||||||
loops: Record<string, Loop>
|
loops: Record<string, Loop>
|
||||||
parallels: Record<string, Parallel>
|
parallels: Record<string, Parallel>
|
||||||
/** Block groups for organizing blocks on the canvas */
|
|
||||||
groups?: Record<string, BlockGroup>
|
|
||||||
lastUpdate?: number
|
lastUpdate?: number
|
||||||
metadata?: {
|
metadata?: {
|
||||||
name?: string
|
name?: string
|
||||||
@@ -264,28 +243,6 @@ export interface WorkflowActions {
|
|||||||
workflowState: WorkflowState,
|
workflowState: WorkflowState,
|
||||||
options?: { updateLastSaved?: boolean }
|
options?: { updateLastSaved?: boolean }
|
||||||
) => void
|
) => void
|
||||||
|
|
||||||
// Block group operations
|
|
||||||
/**
|
|
||||||
* Groups the specified blocks together.
|
|
||||||
* If any blocks are already in a group, they are removed from their current group first.
|
|
||||||
* @returns The new group ID
|
|
||||||
*/
|
|
||||||
groupBlocks: (blockIds: string[], groupId?: string) => string
|
|
||||||
/**
|
|
||||||
* Ungroups a group, removing it and releasing its blocks.
|
|
||||||
* If the group has a parent group, blocks are moved to the parent group.
|
|
||||||
* @returns The block IDs that were in the group
|
|
||||||
*/
|
|
||||||
ungroupBlocks: (groupId: string) => string[]
|
|
||||||
/**
|
|
||||||
* Gets all block IDs in a group, including blocks in nested groups (recursive).
|
|
||||||
*/
|
|
||||||
getGroupBlockIds: (groupId: string, recursive?: boolean) => string[]
|
|
||||||
/**
|
|
||||||
* Gets all groups in the workflow.
|
|
||||||
*/
|
|
||||||
getGroups: () => Record<string, BlockGroup>
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export type WorkflowStore = WorkflowState & WorkflowActions
|
export type WorkflowStore = WorkflowState & WorkflowActions
|
||||||
|
|||||||
@@ -30,11 +30,14 @@ export const a2aCancelTaskTool: ToolConfig<A2ACancelTaskParams, A2ACancelTaskRes
|
|||||||
headers: () => ({
|
headers: () => ({
|
||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
}),
|
}),
|
||||||
body: (params: A2ACancelTaskParams) => ({
|
body: (params: A2ACancelTaskParams) => {
|
||||||
agentUrl: params.agentUrl,
|
const body: Record<string, string> = {
|
||||||
taskId: params.taskId,
|
agentUrl: params.agentUrl,
|
||||||
apiKey: params.apiKey,
|
taskId: params.taskId,
|
||||||
}),
|
}
|
||||||
|
if (params.apiKey) body.apiKey = params.apiKey
|
||||||
|
return body
|
||||||
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
transformResponse: async (response: Response) => {
|
transformResponse: async (response: Response) => {
|
||||||
|
|||||||
@@ -38,12 +38,16 @@ export const a2aDeletePushNotificationTool: ToolConfig<
|
|||||||
headers: () => ({
|
headers: () => ({
|
||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
}),
|
}),
|
||||||
body: (params) => ({
|
body: (params) => {
|
||||||
agentUrl: params.agentUrl,
|
const body: Record<string, string> = {
|
||||||
taskId: params.taskId,
|
agentUrl: params.agentUrl,
|
||||||
pushNotificationConfigId: params.pushNotificationConfigId,
|
taskId: params.taskId,
|
||||||
apiKey: params.apiKey,
|
}
|
||||||
}),
|
if (params.pushNotificationConfigId)
|
||||||
|
body.pushNotificationConfigId = params.pushNotificationConfigId
|
||||||
|
if (params.apiKey) body.apiKey = params.apiKey
|
||||||
|
return body
|
||||||
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
transformResponse: async (response: Response) => {
|
transformResponse: async (response: Response) => {
|
||||||
|
|||||||
@@ -25,10 +25,13 @@ export const a2aGetAgentCardTool: ToolConfig<A2AGetAgentCardParams, A2AGetAgentC
|
|||||||
headers: () => ({
|
headers: () => ({
|
||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
}),
|
}),
|
||||||
body: (params) => ({
|
body: (params) => {
|
||||||
agentUrl: params.agentUrl,
|
const body: Record<string, string> = {
|
||||||
apiKey: params.apiKey,
|
agentUrl: params.agentUrl,
|
||||||
}),
|
}
|
||||||
|
if (params.apiKey) body.apiKey = params.apiKey
|
||||||
|
return body
|
||||||
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
transformResponse: async (response: Response) => {
|
transformResponse: async (response: Response) => {
|
||||||
|
|||||||
@@ -33,11 +33,14 @@ export const a2aGetPushNotificationTool: ToolConfig<
|
|||||||
headers: () => ({
|
headers: () => ({
|
||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
}),
|
}),
|
||||||
body: (params) => ({
|
body: (params) => {
|
||||||
agentUrl: params.agentUrl,
|
const body: Record<string, string> = {
|
||||||
taskId: params.taskId,
|
agentUrl: params.agentUrl,
|
||||||
apiKey: params.apiKey,
|
taskId: params.taskId,
|
||||||
}),
|
}
|
||||||
|
if (params.apiKey) body.apiKey = params.apiKey
|
||||||
|
return body
|
||||||
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
transformResponse: async (response: Response) => {
|
transformResponse: async (response: Response) => {
|
||||||
|
|||||||
@@ -34,12 +34,15 @@ export const a2aGetTaskTool: ToolConfig<A2AGetTaskParams, A2AGetTaskResponse> =
|
|||||||
headers: () => ({
|
headers: () => ({
|
||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
}),
|
}),
|
||||||
body: (params: A2AGetTaskParams) => ({
|
body: (params: A2AGetTaskParams) => {
|
||||||
agentUrl: params.agentUrl,
|
const body: Record<string, string | number> = {
|
||||||
taskId: params.taskId,
|
agentUrl: params.agentUrl,
|
||||||
apiKey: params.apiKey,
|
taskId: params.taskId,
|
||||||
historyLength: params.historyLength,
|
}
|
||||||
}),
|
if (params.apiKey) body.apiKey = params.apiKey
|
||||||
|
if (params.historyLength) body.historyLength = params.historyLength
|
||||||
|
return body
|
||||||
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
transformResponse: async (response: Response) => {
|
transformResponse: async (response: Response) => {
|
||||||
|
|||||||
@@ -5,7 +5,6 @@ import { a2aGetPushNotificationTool } from './get_push_notification'
|
|||||||
import { a2aGetTaskTool } from './get_task'
|
import { a2aGetTaskTool } from './get_task'
|
||||||
import { a2aResubscribeTool } from './resubscribe'
|
import { a2aResubscribeTool } from './resubscribe'
|
||||||
import { a2aSendMessageTool } from './send_message'
|
import { a2aSendMessageTool } from './send_message'
|
||||||
import { a2aSendMessageStreamTool } from './send_message_stream'
|
|
||||||
import { a2aSetPushNotificationTool } from './set_push_notification'
|
import { a2aSetPushNotificationTool } from './set_push_notification'
|
||||||
|
|
||||||
export {
|
export {
|
||||||
@@ -16,6 +15,5 @@ export {
|
|||||||
a2aGetTaskTool,
|
a2aGetTaskTool,
|
||||||
a2aResubscribeTool,
|
a2aResubscribeTool,
|
||||||
a2aSendMessageTool,
|
a2aSendMessageTool,
|
||||||
a2aSendMessageStreamTool,
|
|
||||||
a2aSetPushNotificationTool,
|
a2aSetPushNotificationTool,
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -30,11 +30,14 @@ export const a2aResubscribeTool: ToolConfig<A2AResubscribeParams, A2AResubscribe
|
|||||||
headers: () => ({
|
headers: () => ({
|
||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
}),
|
}),
|
||||||
body: (params: A2AResubscribeParams) => ({
|
body: (params: A2AResubscribeParams) => {
|
||||||
agentUrl: params.agentUrl,
|
const body: Record<string, string> = {
|
||||||
taskId: params.taskId,
|
agentUrl: params.agentUrl,
|
||||||
apiKey: params.apiKey,
|
taskId: params.taskId,
|
||||||
}),
|
}
|
||||||
|
if (params.apiKey) body.apiKey = params.apiKey
|
||||||
|
return body
|
||||||
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
transformResponse: async (response) => {
|
transformResponse: async (response) => {
|
||||||
|
|||||||
@@ -26,6 +26,14 @@ export const a2aSendMessageTool: ToolConfig<A2ASendMessageParams, A2ASendMessage
|
|||||||
type: 'string',
|
type: 'string',
|
||||||
description: 'Context ID for conversation continuity',
|
description: 'Context ID for conversation continuity',
|
||||||
},
|
},
|
||||||
|
data: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'Structured data to include with the message (JSON string)',
|
||||||
|
},
|
||||||
|
files: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Files to include with the message',
|
||||||
|
},
|
||||||
apiKey: {
|
apiKey: {
|
||||||
type: 'string',
|
type: 'string',
|
||||||
description: 'API key for authentication',
|
description: 'API key for authentication',
|
||||||
@@ -35,7 +43,21 @@ export const a2aSendMessageTool: ToolConfig<A2ASendMessageParams, A2ASendMessage
|
|||||||
request: {
|
request: {
|
||||||
url: '/api/tools/a2a/send-message',
|
url: '/api/tools/a2a/send-message',
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
headers: () => ({}),
|
headers: () => ({
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
}),
|
||||||
|
body: (params) => {
|
||||||
|
const body: Record<string, unknown> = {
|
||||||
|
agentUrl: params.agentUrl,
|
||||||
|
message: params.message,
|
||||||
|
}
|
||||||
|
if (params.taskId) body.taskId = params.taskId
|
||||||
|
if (params.contextId) body.contextId = params.contextId
|
||||||
|
if (params.data) body.data = params.data
|
||||||
|
if (params.files && params.files.length > 0) body.files = params.files
|
||||||
|
if (params.apiKey) body.apiKey = params.apiKey
|
||||||
|
return body
|
||||||
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
transformResponse: async (response: Response) => {
|
transformResponse: async (response: Response) => {
|
||||||
|
|||||||
@@ -1,81 +0,0 @@
|
|||||||
import type { ToolConfig } from '@/tools/types'
|
|
||||||
import type { A2ASendMessageParams, A2ASendMessageResponse } from './types'
|
|
||||||
|
|
||||||
export const a2aSendMessageStreamTool: ToolConfig<A2ASendMessageParams, A2ASendMessageResponse> = {
|
|
||||||
id: 'a2a_send_message_stream',
|
|
||||||
name: 'A2A Send Message (Streaming)',
|
|
||||||
description: 'Send a message to an external A2A-compatible agent with real-time streaming.',
|
|
||||||
version: '1.0.0',
|
|
||||||
|
|
||||||
params: {
|
|
||||||
agentUrl: {
|
|
||||||
type: 'string',
|
|
||||||
required: true,
|
|
||||||
description: 'The A2A agent endpoint URL',
|
|
||||||
},
|
|
||||||
message: {
|
|
||||||
type: 'string',
|
|
||||||
required: true,
|
|
||||||
description: 'Message to send to the agent',
|
|
||||||
},
|
|
||||||
taskId: {
|
|
||||||
type: 'string',
|
|
||||||
description: 'Task ID for continuing an existing task',
|
|
||||||
},
|
|
||||||
contextId: {
|
|
||||||
type: 'string',
|
|
||||||
description: 'Context ID for conversation continuity',
|
|
||||||
},
|
|
||||||
apiKey: {
|
|
||||||
type: 'string',
|
|
||||||
description: 'API key for authentication',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
|
|
||||||
request: {
|
|
||||||
url: '/api/tools/a2a/send-message-stream',
|
|
||||||
method: 'POST',
|
|
||||||
headers: () => ({
|
|
||||||
'Content-Type': 'application/json',
|
|
||||||
}),
|
|
||||||
body: (params) => ({
|
|
||||||
agentUrl: params.agentUrl,
|
|
||||||
message: params.message,
|
|
||||||
taskId: params.taskId,
|
|
||||||
contextId: params.contextId,
|
|
||||||
apiKey: params.apiKey,
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
|
|
||||||
transformResponse: async (response: Response) => {
|
|
||||||
const data = await response.json()
|
|
||||||
return data
|
|
||||||
},
|
|
||||||
|
|
||||||
outputs: {
|
|
||||||
content: {
|
|
||||||
type: 'string',
|
|
||||||
description: 'The text response from the agent',
|
|
||||||
},
|
|
||||||
taskId: {
|
|
||||||
type: 'string',
|
|
||||||
description: 'Task ID for follow-up interactions',
|
|
||||||
},
|
|
||||||
contextId: {
|
|
||||||
type: 'string',
|
|
||||||
description: 'Context ID for conversation continuity',
|
|
||||||
},
|
|
||||||
state: {
|
|
||||||
type: 'string',
|
|
||||||
description: 'Task state',
|
|
||||||
},
|
|
||||||
artifacts: {
|
|
||||||
type: 'array',
|
|
||||||
description: 'Structured output artifacts',
|
|
||||||
},
|
|
||||||
history: {
|
|
||||||
type: 'array',
|
|
||||||
description: 'Full message history',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
@@ -42,13 +42,16 @@ export const a2aSetPushNotificationTool: ToolConfig<
|
|||||||
headers: () => ({
|
headers: () => ({
|
||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
}),
|
}),
|
||||||
body: (params: A2ASetPushNotificationParams) => ({
|
body: (params: A2ASetPushNotificationParams) => {
|
||||||
agentUrl: params.agentUrl,
|
const body: Record<string, string> = {
|
||||||
taskId: params.taskId,
|
agentUrl: params.agentUrl,
|
||||||
webhookUrl: params.webhookUrl,
|
taskId: params.taskId,
|
||||||
token: params.token,
|
webhookUrl: params.webhookUrl,
|
||||||
apiKey: params.apiKey,
|
}
|
||||||
}),
|
if (params.token) body.token = params.token
|
||||||
|
if (params.apiKey) body.apiKey = params.apiKey
|
||||||
|
return body
|
||||||
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
transformResponse: async (response: Response) => {
|
transformResponse: async (response: Response) => {
|
||||||
|
|||||||
@@ -25,11 +25,20 @@ export interface A2AGetAgentCardResponse extends ToolResponse {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export interface A2ASendMessageFileInput {
|
||||||
|
type: 'file' | 'url'
|
||||||
|
data: string
|
||||||
|
name: string
|
||||||
|
mime?: string
|
||||||
|
}
|
||||||
|
|
||||||
export interface A2ASendMessageParams {
|
export interface A2ASendMessageParams {
|
||||||
agentUrl: string
|
agentUrl: string
|
||||||
message: string
|
message: string
|
||||||
taskId?: string
|
taskId?: string
|
||||||
contextId?: string
|
contextId?: string
|
||||||
|
data?: string
|
||||||
|
files?: A2ASendMessageFileInput[]
|
||||||
apiKey?: string
|
apiKey?: string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -5,7 +5,6 @@ import {
|
|||||||
a2aGetPushNotificationTool,
|
a2aGetPushNotificationTool,
|
||||||
a2aGetTaskTool,
|
a2aGetTaskTool,
|
||||||
a2aResubscribeTool,
|
a2aResubscribeTool,
|
||||||
a2aSendMessageStreamTool,
|
|
||||||
a2aSendMessageTool,
|
a2aSendMessageTool,
|
||||||
a2aSetPushNotificationTool,
|
a2aSetPushNotificationTool,
|
||||||
} from '@/tools/a2a'
|
} from '@/tools/a2a'
|
||||||
@@ -1180,6 +1179,8 @@ import {
|
|||||||
slackCanvasTool,
|
slackCanvasTool,
|
||||||
slackDeleteMessageTool,
|
slackDeleteMessageTool,
|
||||||
slackDownloadTool,
|
slackDownloadTool,
|
||||||
|
slackGetMessageTool,
|
||||||
|
slackGetThreadTool,
|
||||||
slackGetUserTool,
|
slackGetUserTool,
|
||||||
slackListChannelsTool,
|
slackListChannelsTool,
|
||||||
slackListMembersTool,
|
slackListMembersTool,
|
||||||
@@ -1380,6 +1381,7 @@ import {
|
|||||||
telegramSendVideoTool,
|
telegramSendVideoTool,
|
||||||
} from '@/tools/telegram'
|
} from '@/tools/telegram'
|
||||||
import { thinkingTool } from '@/tools/thinking'
|
import { thinkingTool } from '@/tools/thinking'
|
||||||
|
import { tinybirdEventsTool, tinybirdQueryTool } from '@/tools/tinybird'
|
||||||
import {
|
import {
|
||||||
trelloAddCommentTool,
|
trelloAddCommentTool,
|
||||||
trelloCreateCardTool,
|
trelloCreateCardTool,
|
||||||
@@ -1541,7 +1543,6 @@ export const tools: Record<string, ToolConfig> = {
|
|||||||
a2a_get_task: a2aGetTaskTool,
|
a2a_get_task: a2aGetTaskTool,
|
||||||
a2a_resubscribe: a2aResubscribeTool,
|
a2a_resubscribe: a2aResubscribeTool,
|
||||||
a2a_send_message: a2aSendMessageTool,
|
a2a_send_message: a2aSendMessageTool,
|
||||||
a2a_send_message_stream: a2aSendMessageStreamTool,
|
|
||||||
a2a_set_push_notification: a2aSetPushNotificationTool,
|
a2a_set_push_notification: a2aSetPushNotificationTool,
|
||||||
arxiv_search: arxivSearchTool,
|
arxiv_search: arxivSearchTool,
|
||||||
arxiv_get_paper: arxivGetPaperTool,
|
arxiv_get_paper: arxivGetPaperTool,
|
||||||
@@ -1731,6 +1732,8 @@ export const tools: Record<string, ToolConfig> = {
|
|||||||
slack_list_members: slackListMembersTool,
|
slack_list_members: slackListMembersTool,
|
||||||
slack_list_users: slackListUsersTool,
|
slack_list_users: slackListUsersTool,
|
||||||
slack_get_user: slackGetUserTool,
|
slack_get_user: slackGetUserTool,
|
||||||
|
slack_get_message: slackGetMessageTool,
|
||||||
|
slack_get_thread: slackGetThreadTool,
|
||||||
slack_canvas: slackCanvasTool,
|
slack_canvas: slackCanvasTool,
|
||||||
slack_download: slackDownloadTool,
|
slack_download: slackDownloadTool,
|
||||||
slack_update_message: slackUpdateMessageTool,
|
slack_update_message: slackUpdateMessageTool,
|
||||||
@@ -2235,6 +2238,8 @@ export const tools: Record<string, ToolConfig> = {
|
|||||||
apollo_email_accounts: apolloEmailAccountsTool,
|
apollo_email_accounts: apolloEmailAccountsTool,
|
||||||
mistral_parser: mistralParserTool,
|
mistral_parser: mistralParserTool,
|
||||||
thinking_tool: thinkingTool,
|
thinking_tool: thinkingTool,
|
||||||
|
tinybird_events: tinybirdEventsTool,
|
||||||
|
tinybird_query: tinybirdQueryTool,
|
||||||
stagehand_extract: stagehandExtractTool,
|
stagehand_extract: stagehandExtractTool,
|
||||||
stagehand_agent: stagehandAgentTool,
|
stagehand_agent: stagehandAgentTool,
|
||||||
mem0_add_memories: mem0AddMemoriesTool,
|
mem0_add_memories: mem0AddMemoriesTool,
|
||||||
|
|||||||
213
apps/sim/tools/slack/get_message.ts
Normal file
213
apps/sim/tools/slack/get_message.ts
Normal file
@@ -0,0 +1,213 @@
|
|||||||
|
import type { SlackGetMessageParams, SlackGetMessageResponse } from '@/tools/slack/types'
|
||||||
|
import type { ToolConfig } from '@/tools/types'
|
||||||
|
|
||||||
|
export const slackGetMessageTool: ToolConfig<SlackGetMessageParams, SlackGetMessageResponse> = {
|
||||||
|
id: 'slack_get_message',
|
||||||
|
name: 'Slack Get Message',
|
||||||
|
description:
|
||||||
|
'Retrieve a specific message by its timestamp. Useful for getting a thread parent message.',
|
||||||
|
version: '1.0.0',
|
||||||
|
|
||||||
|
oauth: {
|
||||||
|
required: true,
|
||||||
|
provider: 'slack',
|
||||||
|
},
|
||||||
|
|
||||||
|
params: {
|
||||||
|
authMethod: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Authentication method: oauth or bot_token',
|
||||||
|
},
|
||||||
|
botToken: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Bot token for Custom Bot',
|
||||||
|
},
|
||||||
|
accessToken: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'hidden',
|
||||||
|
description: 'OAuth access token or bot token for Slack API',
|
||||||
|
},
|
||||||
|
channel: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Slack channel ID (e.g., C1234567890)',
|
||||||
|
},
|
||||||
|
timestamp: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-or-llm',
|
||||||
|
description: 'Message timestamp to retrieve (e.g., 1405894322.002768)',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
request: {
|
||||||
|
url: (params: SlackGetMessageParams) => {
|
||||||
|
const url = new URL('https://slack.com/api/conversations.history')
|
||||||
|
url.searchParams.append('channel', params.channel?.trim() ?? '')
|
||||||
|
url.searchParams.append('oldest', params.timestamp?.trim() ?? '')
|
||||||
|
url.searchParams.append('limit', '1')
|
||||||
|
url.searchParams.append('inclusive', 'true')
|
||||||
|
return url.toString()
|
||||||
|
},
|
||||||
|
method: 'GET',
|
||||||
|
headers: (params: SlackGetMessageParams) => ({
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
Authorization: `Bearer ${params.accessToken || params.botToken}`,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
|
||||||
|
transformResponse: async (response: Response) => {
|
||||||
|
const data = await response.json()
|
||||||
|
|
||||||
|
if (!data.ok) {
|
||||||
|
if (data.error === 'missing_scope') {
|
||||||
|
throw new Error(
|
||||||
|
'Missing required permissions. Please reconnect your Slack account with the necessary scopes (channels:history, groups:history).'
|
||||||
|
)
|
||||||
|
}
|
||||||
|
if (data.error === 'invalid_auth') {
|
||||||
|
throw new Error('Invalid authentication. Please check your Slack credentials.')
|
||||||
|
}
|
||||||
|
if (data.error === 'channel_not_found') {
|
||||||
|
throw new Error('Channel not found. Please check the channel ID.')
|
||||||
|
}
|
||||||
|
throw new Error(data.error || 'Failed to get message from Slack')
|
||||||
|
}
|
||||||
|
|
||||||
|
const messages = data.messages || []
|
||||||
|
if (messages.length === 0) {
|
||||||
|
throw new Error('Message not found')
|
||||||
|
}
|
||||||
|
|
||||||
|
const msg = messages[0]
|
||||||
|
const message = {
|
||||||
|
type: msg.type ?? 'message',
|
||||||
|
ts: msg.ts,
|
||||||
|
text: msg.text ?? '',
|
||||||
|
user: msg.user ?? null,
|
||||||
|
bot_id: msg.bot_id ?? null,
|
||||||
|
username: msg.username ?? null,
|
||||||
|
channel: msg.channel ?? null,
|
||||||
|
team: msg.team ?? null,
|
||||||
|
thread_ts: msg.thread_ts ?? null,
|
||||||
|
parent_user_id: msg.parent_user_id ?? null,
|
||||||
|
reply_count: msg.reply_count ?? null,
|
||||||
|
reply_users_count: msg.reply_users_count ?? null,
|
||||||
|
latest_reply: msg.latest_reply ?? null,
|
||||||
|
subscribed: msg.subscribed ?? null,
|
||||||
|
last_read: msg.last_read ?? null,
|
||||||
|
unread_count: msg.unread_count ?? null,
|
||||||
|
subtype: msg.subtype ?? null,
|
||||||
|
reactions: msg.reactions ?? [],
|
||||||
|
is_starred: msg.is_starred ?? false,
|
||||||
|
pinned_to: msg.pinned_to ?? [],
|
||||||
|
files: (msg.files ?? []).map((f: any) => ({
|
||||||
|
id: f.id,
|
||||||
|
name: f.name,
|
||||||
|
mimetype: f.mimetype,
|
||||||
|
size: f.size,
|
||||||
|
url_private: f.url_private ?? null,
|
||||||
|
permalink: f.permalink ?? null,
|
||||||
|
mode: f.mode ?? null,
|
||||||
|
})),
|
||||||
|
attachments: msg.attachments ?? [],
|
||||||
|
blocks: msg.blocks ?? [],
|
||||||
|
edited: msg.edited ?? null,
|
||||||
|
permalink: msg.permalink ?? null,
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
message,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
outputs: {
|
||||||
|
message: {
|
||||||
|
type: 'object',
|
||||||
|
description: 'The retrieved message object',
|
||||||
|
properties: {
|
||||||
|
type: { type: 'string', description: 'Message type' },
|
||||||
|
ts: { type: 'string', description: 'Message timestamp' },
|
||||||
|
text: { type: 'string', description: 'Message text content' },
|
||||||
|
user: { type: 'string', description: 'User ID who sent the message' },
|
||||||
|
bot_id: { type: 'string', description: 'Bot ID if sent by a bot', optional: true },
|
||||||
|
username: { type: 'string', description: 'Display username', optional: true },
|
||||||
|
channel: { type: 'string', description: 'Channel ID', optional: true },
|
||||||
|
team: { type: 'string', description: 'Team ID', optional: true },
|
||||||
|
thread_ts: { type: 'string', description: 'Thread parent timestamp', optional: true },
|
||||||
|
parent_user_id: { type: 'string', description: 'User ID of thread parent', optional: true },
|
||||||
|
reply_count: { type: 'number', description: 'Number of thread replies', optional: true },
|
||||||
|
reply_users_count: {
|
||||||
|
type: 'number',
|
||||||
|
description: 'Number of users who replied',
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
latest_reply: { type: 'string', description: 'Timestamp of latest reply', optional: true },
|
||||||
|
subtype: { type: 'string', description: 'Message subtype', optional: true },
|
||||||
|
reactions: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Array of reactions on this message',
|
||||||
|
items: {
|
||||||
|
type: 'object',
|
||||||
|
properties: {
|
||||||
|
name: { type: 'string', description: 'Emoji name' },
|
||||||
|
count: { type: 'number', description: 'Number of reactions' },
|
||||||
|
users: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'User IDs who reacted',
|
||||||
|
items: { type: 'string' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
is_starred: { type: 'boolean', description: 'Whether message is starred', optional: true },
|
||||||
|
pinned_to: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Channel IDs where message is pinned',
|
||||||
|
items: { type: 'string' },
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
files: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Files attached to message',
|
||||||
|
items: {
|
||||||
|
type: 'object',
|
||||||
|
properties: {
|
||||||
|
id: { type: 'string', description: 'File ID' },
|
||||||
|
name: { type: 'string', description: 'File name' },
|
||||||
|
mimetype: { type: 'string', description: 'MIME type' },
|
||||||
|
size: { type: 'number', description: 'File size in bytes' },
|
||||||
|
url_private: { type: 'string', description: 'Private download URL' },
|
||||||
|
permalink: { type: 'string', description: 'Permanent link to file' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
attachments: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Legacy attachments',
|
||||||
|
items: { type: 'object' },
|
||||||
|
},
|
||||||
|
blocks: { type: 'array', description: 'Block Kit blocks', items: { type: 'object' } },
|
||||||
|
edited: {
|
||||||
|
type: 'object',
|
||||||
|
description: 'Edit information if message was edited',
|
||||||
|
properties: {
|
||||||
|
user: { type: 'string', description: 'User ID who edited' },
|
||||||
|
ts: { type: 'string', description: 'Edit timestamp' },
|
||||||
|
},
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
permalink: { type: 'string', description: 'Permanent link to message', optional: true },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
224
apps/sim/tools/slack/get_thread.ts
Normal file
224
apps/sim/tools/slack/get_thread.ts
Normal file
@@ -0,0 +1,224 @@
|
|||||||
|
import type { SlackGetThreadParams, SlackGetThreadResponse } from '@/tools/slack/types'
|
||||||
|
import type { ToolConfig } from '@/tools/types'
|
||||||
|
|
||||||
|
export const slackGetThreadTool: ToolConfig<SlackGetThreadParams, SlackGetThreadResponse> = {
|
||||||
|
id: 'slack_get_thread',
|
||||||
|
name: 'Slack Get Thread',
|
||||||
|
description:
|
||||||
|
'Retrieve an entire thread including the parent message and all replies. Useful for getting full conversation context.',
|
||||||
|
version: '1.0.0',
|
||||||
|
|
||||||
|
oauth: {
|
||||||
|
required: true,
|
||||||
|
provider: 'slack',
|
||||||
|
},
|
||||||
|
|
||||||
|
params: {
|
||||||
|
authMethod: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Authentication method: oauth or bot_token',
|
||||||
|
},
|
||||||
|
botToken: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Bot token for Custom Bot',
|
||||||
|
},
|
||||||
|
accessToken: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'hidden',
|
||||||
|
description: 'OAuth access token or bot token for Slack API',
|
||||||
|
},
|
||||||
|
channel: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Slack channel ID (e.g., C1234567890)',
|
||||||
|
},
|
||||||
|
threadTs: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-or-llm',
|
||||||
|
description: 'Thread timestamp (thread_ts) to retrieve (e.g., 1405894322.002768)',
|
||||||
|
},
|
||||||
|
limit: {
|
||||||
|
type: 'number',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-or-llm',
|
||||||
|
description: 'Maximum number of messages to return (default: 100, max: 200)',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
request: {
|
||||||
|
url: (params: SlackGetThreadParams) => {
|
||||||
|
const url = new URL('https://slack.com/api/conversations.replies')
|
||||||
|
url.searchParams.append('channel', params.channel?.trim() ?? '')
|
||||||
|
url.searchParams.append('ts', params.threadTs?.trim() ?? '')
|
||||||
|
url.searchParams.append('inclusive', 'true')
|
||||||
|
const limit = params.limit ? Math.min(Number(params.limit), 200) : 100
|
||||||
|
url.searchParams.append('limit', String(limit))
|
||||||
|
return url.toString()
|
||||||
|
},
|
||||||
|
method: 'GET',
|
||||||
|
headers: (params: SlackGetThreadParams) => ({
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
Authorization: `Bearer ${params.accessToken || params.botToken}`,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
|
||||||
|
transformResponse: async (response: Response) => {
|
||||||
|
const data = await response.json()
|
||||||
|
|
||||||
|
if (!data.ok) {
|
||||||
|
if (data.error === 'missing_scope') {
|
||||||
|
throw new Error(
|
||||||
|
'Missing required permissions. Please reconnect your Slack account with the necessary scopes (channels:history, groups:history).'
|
||||||
|
)
|
||||||
|
}
|
||||||
|
if (data.error === 'invalid_auth') {
|
||||||
|
throw new Error('Invalid authentication. Please check your Slack credentials.')
|
||||||
|
}
|
||||||
|
if (data.error === 'channel_not_found') {
|
||||||
|
throw new Error('Channel not found. Please check the channel ID.')
|
||||||
|
}
|
||||||
|
if (data.error === 'thread_not_found') {
|
||||||
|
throw new Error('Thread not found. Please check the thread timestamp.')
|
||||||
|
}
|
||||||
|
throw new Error(data.error || 'Failed to get thread from Slack')
|
||||||
|
}
|
||||||
|
|
||||||
|
const rawMessages = data.messages || []
|
||||||
|
if (rawMessages.length === 0) {
|
||||||
|
throw new Error('Thread not found')
|
||||||
|
}
|
||||||
|
|
||||||
|
const messages = rawMessages.map((msg: any) => ({
|
||||||
|
type: msg.type ?? 'message',
|
||||||
|
ts: msg.ts,
|
||||||
|
text: msg.text ?? '',
|
||||||
|
user: msg.user ?? null,
|
||||||
|
bot_id: msg.bot_id ?? null,
|
||||||
|
username: msg.username ?? null,
|
||||||
|
channel: msg.channel ?? null,
|
||||||
|
team: msg.team ?? null,
|
||||||
|
thread_ts: msg.thread_ts ?? null,
|
||||||
|
parent_user_id: msg.parent_user_id ?? null,
|
||||||
|
reply_count: msg.reply_count ?? null,
|
||||||
|
reply_users_count: msg.reply_users_count ?? null,
|
||||||
|
latest_reply: msg.latest_reply ?? null,
|
||||||
|
subscribed: msg.subscribed ?? null,
|
||||||
|
last_read: msg.last_read ?? null,
|
||||||
|
unread_count: msg.unread_count ?? null,
|
||||||
|
subtype: msg.subtype ?? null,
|
||||||
|
reactions: msg.reactions ?? [],
|
||||||
|
is_starred: msg.is_starred ?? false,
|
||||||
|
pinned_to: msg.pinned_to ?? [],
|
||||||
|
files: (msg.files ?? []).map((f: any) => ({
|
||||||
|
id: f.id,
|
||||||
|
name: f.name,
|
||||||
|
mimetype: f.mimetype,
|
||||||
|
size: f.size,
|
||||||
|
url_private: f.url_private ?? null,
|
||||||
|
permalink: f.permalink ?? null,
|
||||||
|
mode: f.mode ?? null,
|
||||||
|
})),
|
||||||
|
attachments: msg.attachments ?? [],
|
||||||
|
blocks: msg.blocks ?? [],
|
||||||
|
edited: msg.edited ?? null,
|
||||||
|
permalink: msg.permalink ?? null,
|
||||||
|
}))
|
||||||
|
|
||||||
|
// First message is always the parent
|
||||||
|
const parentMessage = messages[0]
|
||||||
|
// Remaining messages are replies
|
||||||
|
const replies = messages.slice(1)
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
parentMessage,
|
||||||
|
replies,
|
||||||
|
messages,
|
||||||
|
replyCount: replies.length,
|
||||||
|
hasMore: data.has_more ?? false,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
outputs: {
|
||||||
|
parentMessage: {
|
||||||
|
type: 'object',
|
||||||
|
description: 'The thread parent message',
|
||||||
|
properties: {
|
||||||
|
type: { type: 'string', description: 'Message type' },
|
||||||
|
ts: { type: 'string', description: 'Message timestamp' },
|
||||||
|
text: { type: 'string', description: 'Message text content' },
|
||||||
|
user: { type: 'string', description: 'User ID who sent the message' },
|
||||||
|
bot_id: { type: 'string', description: 'Bot ID if sent by a bot', optional: true },
|
||||||
|
username: { type: 'string', description: 'Display username', optional: true },
|
||||||
|
reply_count: { type: 'number', description: 'Total number of thread replies' },
|
||||||
|
reply_users_count: { type: 'number', description: 'Number of users who replied' },
|
||||||
|
latest_reply: { type: 'string', description: 'Timestamp of latest reply' },
|
||||||
|
reactions: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Array of reactions on the parent message',
|
||||||
|
items: {
|
||||||
|
type: 'object',
|
||||||
|
properties: {
|
||||||
|
name: { type: 'string', description: 'Emoji name' },
|
||||||
|
count: { type: 'number', description: 'Number of reactions' },
|
||||||
|
users: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'User IDs who reacted',
|
||||||
|
items: { type: 'string' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
files: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Files attached to the parent message',
|
||||||
|
items: {
|
||||||
|
type: 'object',
|
||||||
|
properties: {
|
||||||
|
id: { type: 'string', description: 'File ID' },
|
||||||
|
name: { type: 'string', description: 'File name' },
|
||||||
|
mimetype: { type: 'string', description: 'MIME type' },
|
||||||
|
size: { type: 'number', description: 'File size in bytes' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
replies: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'Array of reply messages in the thread (excluding the parent)',
|
||||||
|
items: {
|
||||||
|
type: 'object',
|
||||||
|
properties: {
|
||||||
|
ts: { type: 'string', description: 'Message timestamp' },
|
||||||
|
text: { type: 'string', description: 'Message text content' },
|
||||||
|
user: { type: 'string', description: 'User ID who sent the reply' },
|
||||||
|
reactions: { type: 'array', description: 'Reactions on the reply' },
|
||||||
|
files: { type: 'array', description: 'Files attached to the reply' },
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
messages: {
|
||||||
|
type: 'array',
|
||||||
|
description: 'All messages in the thread (parent + replies) in chronological order',
|
||||||
|
items: { type: 'object' },
|
||||||
|
},
|
||||||
|
replyCount: {
|
||||||
|
type: 'number',
|
||||||
|
description: 'Number of replies returned in this response',
|
||||||
|
},
|
||||||
|
hasMore: {
|
||||||
|
type: 'boolean',
|
||||||
|
description: 'Whether there are more messages in the thread (pagination needed)',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
@@ -2,6 +2,8 @@ import { slackAddReactionTool } from '@/tools/slack/add_reaction'
|
|||||||
import { slackCanvasTool } from '@/tools/slack/canvas'
|
import { slackCanvasTool } from '@/tools/slack/canvas'
|
||||||
import { slackDeleteMessageTool } from '@/tools/slack/delete_message'
|
import { slackDeleteMessageTool } from '@/tools/slack/delete_message'
|
||||||
import { slackDownloadTool } from '@/tools/slack/download'
|
import { slackDownloadTool } from '@/tools/slack/download'
|
||||||
|
import { slackGetMessageTool } from '@/tools/slack/get_message'
|
||||||
|
import { slackGetThreadTool } from '@/tools/slack/get_thread'
|
||||||
import { slackGetUserTool } from '@/tools/slack/get_user'
|
import { slackGetUserTool } from '@/tools/slack/get_user'
|
||||||
import { slackListChannelsTool } from '@/tools/slack/list_channels'
|
import { slackListChannelsTool } from '@/tools/slack/list_channels'
|
||||||
import { slackListMembersTool } from '@/tools/slack/list_members'
|
import { slackListMembersTool } from '@/tools/slack/list_members'
|
||||||
@@ -22,4 +24,6 @@ export {
|
|||||||
slackListMembersTool,
|
slackListMembersTool,
|
||||||
slackListUsersTool,
|
slackListUsersTool,
|
||||||
slackGetUserTool,
|
slackGetUserTool,
|
||||||
|
slackGetMessageTool,
|
||||||
|
slackGetThreadTool,
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -71,6 +71,17 @@ export interface SlackGetUserParams extends SlackBaseParams {
|
|||||||
userId: string
|
userId: string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export interface SlackGetMessageParams extends SlackBaseParams {
|
||||||
|
channel: string
|
||||||
|
timestamp: string
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface SlackGetThreadParams extends SlackBaseParams {
|
||||||
|
channel: string
|
||||||
|
threadTs: string
|
||||||
|
limit?: number
|
||||||
|
}
|
||||||
|
|
||||||
export interface SlackMessageResponse extends ToolResponse {
|
export interface SlackMessageResponse extends ToolResponse {
|
||||||
output: {
|
output: {
|
||||||
// Legacy properties for backward compatibility
|
// Legacy properties for backward compatibility
|
||||||
@@ -305,6 +316,22 @@ export interface SlackGetUserResponse extends ToolResponse {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export interface SlackGetMessageResponse extends ToolResponse {
|
||||||
|
output: {
|
||||||
|
message: SlackMessage
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface SlackGetThreadResponse extends ToolResponse {
|
||||||
|
output: {
|
||||||
|
parentMessage: SlackMessage
|
||||||
|
replies: SlackMessage[]
|
||||||
|
messages: SlackMessage[]
|
||||||
|
replyCount: number
|
||||||
|
hasMore: boolean
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
export type SlackResponse =
|
export type SlackResponse =
|
||||||
| SlackCanvasResponse
|
| SlackCanvasResponse
|
||||||
| SlackMessageReaderResponse
|
| SlackMessageReaderResponse
|
||||||
@@ -317,3 +344,5 @@ export type SlackResponse =
|
|||||||
| SlackListMembersResponse
|
| SlackListMembersResponse
|
||||||
| SlackListUsersResponse
|
| SlackListUsersResponse
|
||||||
| SlackGetUserResponse
|
| SlackGetUserResponse
|
||||||
|
| SlackGetMessageResponse
|
||||||
|
| SlackGetThreadResponse
|
||||||
|
|||||||
128
apps/sim/tools/tinybird/events.ts
Normal file
128
apps/sim/tools/tinybird/events.ts
Normal file
@@ -0,0 +1,128 @@
|
|||||||
|
import { gzipSync } from 'zlib'
|
||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import type { TinybirdEventsParams, TinybirdEventsResponse } from '@/tools/tinybird/types'
|
||||||
|
import type { ToolConfig } from '@/tools/types'
|
||||||
|
|
||||||
|
const logger = createLogger('tinybird-events')
|
||||||
|
|
||||||
|
export const eventsTool: ToolConfig<TinybirdEventsParams, TinybirdEventsResponse> = {
|
||||||
|
id: 'tinybird_events',
|
||||||
|
name: 'Tinybird Events',
|
||||||
|
description:
|
||||||
|
'Send events to a Tinybird Data Source using the Events API. Supports JSON and NDJSON formats with optional gzip compression.',
|
||||||
|
version: '1.0.0',
|
||||||
|
errorExtractor: 'nested-error-object',
|
||||||
|
|
||||||
|
params: {
|
||||||
|
base_url: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description:
|
||||||
|
'Tinybird API base URL (e.g., https://api.tinybird.co or https://api.us-east.tinybird.co)',
|
||||||
|
},
|
||||||
|
datasource: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-or-llm',
|
||||||
|
description: 'Name of the Tinybird Data Source to send events to',
|
||||||
|
},
|
||||||
|
data: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-or-llm',
|
||||||
|
description:
|
||||||
|
'Data to send as NDJSON (newline-delimited JSON) or JSON string. Each event should be a valid JSON object.',
|
||||||
|
},
|
||||||
|
wait: {
|
||||||
|
type: 'boolean',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description:
|
||||||
|
'Wait for database acknowledgment before responding. Enables safer retries but introduces latency. Defaults to false.',
|
||||||
|
},
|
||||||
|
format: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Format of the events data: "ndjson" (default) or "json"',
|
||||||
|
},
|
||||||
|
compression: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Compression format: "none" (default) or "gzip"',
|
||||||
|
},
|
||||||
|
token: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Tinybird API Token with DATASOURCE:APPEND or DATASOURCE:CREATE scope',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
request: {
|
||||||
|
url: (params) => {
|
||||||
|
const baseUrl = params.base_url.endsWith('/') ? params.base_url.slice(0, -1) : params.base_url
|
||||||
|
const url = new URL(`${baseUrl}/v0/events`)
|
||||||
|
url.searchParams.set('name', params.datasource)
|
||||||
|
if (params.wait) {
|
||||||
|
url.searchParams.set('wait', 'true')
|
||||||
|
}
|
||||||
|
return url.toString()
|
||||||
|
},
|
||||||
|
method: 'POST',
|
||||||
|
headers: (params) => {
|
||||||
|
const headers: Record<string, string> = {
|
||||||
|
Authorization: `Bearer ${params.token}`,
|
||||||
|
}
|
||||||
|
|
||||||
|
if (params.compression === 'gzip') {
|
||||||
|
headers['Content-Encoding'] = 'gzip'
|
||||||
|
}
|
||||||
|
|
||||||
|
if (params.format === 'json') {
|
||||||
|
headers['Content-Type'] = 'application/json'
|
||||||
|
} else {
|
||||||
|
headers['Content-Type'] = 'application/x-ndjson'
|
||||||
|
}
|
||||||
|
|
||||||
|
return headers
|
||||||
|
},
|
||||||
|
body: (params) => {
|
||||||
|
const data = params.data
|
||||||
|
if (params.compression === 'gzip') {
|
||||||
|
return gzipSync(Buffer.from(data, 'utf-8'))
|
||||||
|
}
|
||||||
|
return data
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
transformResponse: async (response: Response) => {
|
||||||
|
const data = await response.json()
|
||||||
|
|
||||||
|
logger.info('Successfully sent events to Tinybird', {
|
||||||
|
successful: data.successful_rows,
|
||||||
|
quarantined: data.quarantined_rows,
|
||||||
|
})
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
successful_rows: data.successful_rows ?? 0,
|
||||||
|
quarantined_rows: data.quarantined_rows ?? 0,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
outputs: {
|
||||||
|
successful_rows: {
|
||||||
|
type: 'number',
|
||||||
|
description: 'Number of rows successfully ingested',
|
||||||
|
},
|
||||||
|
quarantined_rows: {
|
||||||
|
type: 'number',
|
||||||
|
description: 'Number of rows quarantined (failed validation)',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
5
apps/sim/tools/tinybird/index.ts
Normal file
5
apps/sim/tools/tinybird/index.ts
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
import { eventsTool } from '@/tools/tinybird/events'
|
||||||
|
import { queryTool } from '@/tools/tinybird/query'
|
||||||
|
|
||||||
|
export const tinybirdEventsTool = eventsTool
|
||||||
|
export const tinybirdQueryTool = queryTool
|
||||||
139
apps/sim/tools/tinybird/query.ts
Normal file
139
apps/sim/tools/tinybird/query.ts
Normal file
@@ -0,0 +1,139 @@
|
|||||||
|
import { createLogger } from '@sim/logger'
|
||||||
|
import type { TinybirdQueryParams, TinybirdQueryResponse } from '@/tools/tinybird/types'
|
||||||
|
import type { ToolConfig } from '@/tools/types'
|
||||||
|
|
||||||
|
const logger = createLogger('tinybird-query')
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Tinybird Query Tool
|
||||||
|
*
|
||||||
|
* Executes SQL queries against Tinybird and returns results in the format specified in the query.
|
||||||
|
* - FORMAT JSON: Returns structured data with rows/statistics metadata
|
||||||
|
* - FORMAT CSV/TSV/etc: Returns raw text string
|
||||||
|
*
|
||||||
|
* The tool automatically detects the response format based on Content-Type headers.
|
||||||
|
*/
|
||||||
|
export const queryTool: ToolConfig<TinybirdQueryParams, TinybirdQueryResponse> = {
|
||||||
|
id: 'tinybird_query',
|
||||||
|
name: 'Tinybird Query',
|
||||||
|
description: 'Execute SQL queries against Tinybird Pipes and Data Sources using the Query API.',
|
||||||
|
version: '1.0.0',
|
||||||
|
errorExtractor: 'nested-error-object',
|
||||||
|
|
||||||
|
params: {
|
||||||
|
base_url: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Tinybird API base URL (e.g., https://api.tinybird.co)',
|
||||||
|
},
|
||||||
|
query: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-or-llm',
|
||||||
|
description:
|
||||||
|
'SQL query to execute. Specify your desired output format (e.g., FORMAT JSON, FORMAT CSV, FORMAT TSV). JSON format provides structured data, while other formats return raw text.',
|
||||||
|
},
|
||||||
|
pipeline: {
|
||||||
|
type: 'string',
|
||||||
|
required: false,
|
||||||
|
visibility: 'user-or-llm',
|
||||||
|
description: 'Optional pipe name. When provided, enables SELECT * FROM _ syntax',
|
||||||
|
},
|
||||||
|
token: {
|
||||||
|
type: 'string',
|
||||||
|
required: true,
|
||||||
|
visibility: 'user-only',
|
||||||
|
description: 'Tinybird API Token with PIPE:READ scope',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
request: {
|
||||||
|
url: (params) => {
|
||||||
|
const baseUrl = params.base_url.endsWith('/') ? params.base_url.slice(0, -1) : params.base_url
|
||||||
|
return `${baseUrl}/v0/sql`
|
||||||
|
},
|
||||||
|
method: 'POST',
|
||||||
|
headers: (params) => ({
|
||||||
|
'Content-Type': 'application/x-www-form-urlencoded',
|
||||||
|
Authorization: `Bearer ${params.token}`,
|
||||||
|
}),
|
||||||
|
body: (params) => {
|
||||||
|
const searchParams = new URLSearchParams()
|
||||||
|
searchParams.set('q', params.query)
|
||||||
|
if (params.pipeline) {
|
||||||
|
searchParams.set('pipeline', params.pipeline)
|
||||||
|
}
|
||||||
|
return searchParams.toString()
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
transformResponse: async (response: Response) => {
|
||||||
|
const responseText = await response.text()
|
||||||
|
const contentType = response.headers.get('content-type') || ''
|
||||||
|
|
||||||
|
// Check if response is JSON based on content-type or try parsing
|
||||||
|
const isJson = contentType.includes('application/json') || contentType.includes('text/json')
|
||||||
|
|
||||||
|
if (isJson) {
|
||||||
|
try {
|
||||||
|
const data = JSON.parse(responseText)
|
||||||
|
logger.info('Successfully executed Tinybird query (JSON)', {
|
||||||
|
rows: data.rows,
|
||||||
|
elapsed: data.statistics?.elapsed,
|
||||||
|
})
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
data: data.data || [],
|
||||||
|
rows: data.rows || 0,
|
||||||
|
statistics: data.statistics
|
||||||
|
? {
|
||||||
|
elapsed: data.statistics.elapsed,
|
||||||
|
rows_read: data.statistics.rows_read,
|
||||||
|
bytes_read: data.statistics.bytes_read,
|
||||||
|
}
|
||||||
|
: undefined,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
} catch (parseError) {
|
||||||
|
logger.error('Failed to parse JSON response', {
|
||||||
|
contentType,
|
||||||
|
parseError: parseError instanceof Error ? parseError.message : String(parseError),
|
||||||
|
})
|
||||||
|
throw new Error(
|
||||||
|
`Invalid JSON response: ${parseError instanceof Error ? parseError.message : 'Parse error'}`
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// For non-JSON formats (CSV, TSV, etc.), return as raw text
|
||||||
|
logger.info('Successfully executed Tinybird query (non-JSON)', { contentType })
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
output: {
|
||||||
|
data: responseText,
|
||||||
|
rows: undefined,
|
||||||
|
statistics: undefined,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
outputs: {
|
||||||
|
data: {
|
||||||
|
type: 'json',
|
||||||
|
description:
|
||||||
|
'Query result data. For FORMAT JSON: array of objects. For other formats (CSV, TSV, etc.): raw text string.',
|
||||||
|
},
|
||||||
|
rows: {
|
||||||
|
type: 'number',
|
||||||
|
description: 'Number of rows returned (only available with FORMAT JSON)',
|
||||||
|
},
|
||||||
|
statistics: {
|
||||||
|
type: 'json',
|
||||||
|
description:
|
||||||
|
'Query execution statistics - elapsed time, rows read, bytes read (only available with FORMAT JSON)',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
59
apps/sim/tools/tinybird/types.ts
Normal file
59
apps/sim/tools/tinybird/types.ts
Normal file
@@ -0,0 +1,59 @@
|
|||||||
|
import type { ToolResponse } from '@/tools/types'
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Base parameters for Tinybird API tools
|
||||||
|
*/
|
||||||
|
export interface TinybirdBaseParams {
|
||||||
|
token: string
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Parameters for sending events to Tinybird
|
||||||
|
*/
|
||||||
|
export interface TinybirdEventsParams extends TinybirdBaseParams {
|
||||||
|
base_url: string
|
||||||
|
datasource: string
|
||||||
|
data: string
|
||||||
|
wait?: boolean
|
||||||
|
format?: 'ndjson' | 'json'
|
||||||
|
compression?: 'none' | 'gzip'
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Response from sending events to Tinybird
|
||||||
|
*/
|
||||||
|
export interface TinybirdEventsResponse extends ToolResponse {
|
||||||
|
output: {
|
||||||
|
successful_rows: number
|
||||||
|
quarantined_rows: number
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Parameters for querying Tinybird
|
||||||
|
*/
|
||||||
|
export interface TinybirdQueryParams extends TinybirdBaseParams {
|
||||||
|
base_url: string
|
||||||
|
query: string
|
||||||
|
pipeline?: string
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Response from querying Tinybird
|
||||||
|
*/
|
||||||
|
export interface TinybirdQueryResponse extends ToolResponse {
|
||||||
|
output: {
|
||||||
|
data: unknown[] | string
|
||||||
|
rows?: number
|
||||||
|
statistics?: {
|
||||||
|
elapsed: number
|
||||||
|
rows_read: number
|
||||||
|
bytes_read: number
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Union type for all possible Tinybird responses
|
||||||
|
*/
|
||||||
|
export type TinybirdResponse = TinybirdEventsResponse | TinybirdQueryResponse
|
||||||
@@ -3,6 +3,7 @@ import { getBaseUrl } from '@/lib/core/utils/urls'
|
|||||||
import { AGENT, isCustomTool } from '@/executor/constants'
|
import { AGENT, isCustomTool } from '@/executor/constants'
|
||||||
import { useCustomToolsStore } from '@/stores/custom-tools'
|
import { useCustomToolsStore } from '@/stores/custom-tools'
|
||||||
import { useEnvironmentStore } from '@/stores/settings/environment'
|
import { useEnvironmentStore } from '@/stores/settings/environment'
|
||||||
|
import { extractErrorMessage } from '@/tools/error-extractors'
|
||||||
import { tools } from '@/tools/registry'
|
import { tools } from '@/tools/registry'
|
||||||
import type { TableRow, ToolConfig, ToolResponse } from '@/tools/types'
|
import type { TableRow, ToolConfig, ToolResponse } from '@/tools/types'
|
||||||
|
|
||||||
@@ -162,14 +163,22 @@ export async function executeRequest(
|
|||||||
const externalResponse = await fetch(url, { method, headers, body })
|
const externalResponse = await fetch(url, { method, headers, body })
|
||||||
|
|
||||||
if (!externalResponse.ok) {
|
if (!externalResponse.ok) {
|
||||||
let errorContent
|
let errorData: any
|
||||||
try {
|
try {
|
||||||
errorContent = await externalResponse.json()
|
errorData = await externalResponse.json()
|
||||||
} catch (_e) {
|
} catch (_e) {
|
||||||
errorContent = { message: externalResponse.statusText }
|
try {
|
||||||
|
errorData = await externalResponse.text()
|
||||||
|
} catch (_e2) {
|
||||||
|
errorData = null
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const error = errorContent.message || `${toolId} API error: ${externalResponse.statusText}`
|
const error = extractErrorMessage({
|
||||||
|
status: externalResponse.status,
|
||||||
|
statusText: externalResponse.statusText,
|
||||||
|
data: errorData,
|
||||||
|
})
|
||||||
logger.error(`${toolId} error:`, { error })
|
logger.error(`${toolId} error:`, { error })
|
||||||
throw new Error(error)
|
throw new Error(error)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -96,23 +96,3 @@ export function buildMeetingOutputs(): Record<string, TriggerOutput> {
|
|||||||
},
|
},
|
||||||
} as Record<string, TriggerOutput>
|
} as Record<string, TriggerOutput>
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Build output schema for generic webhook events
|
|
||||||
*/
|
|
||||||
export function buildGenericOutputs(): Record<string, TriggerOutput> {
|
|
||||||
return {
|
|
||||||
payload: {
|
|
||||||
type: 'object',
|
|
||||||
description: 'Raw webhook payload',
|
|
||||||
},
|
|
||||||
headers: {
|
|
||||||
type: 'object',
|
|
||||||
description: 'Request headers',
|
|
||||||
},
|
|
||||||
timestamp: {
|
|
||||||
type: 'string',
|
|
||||||
description: 'ISO8601 received timestamp',
|
|
||||||
},
|
|
||||||
} as Record<string, TriggerOutput>
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
import { CirclebackIcon } from '@/components/icons'
|
import { CirclebackIcon } from '@/components/icons'
|
||||||
import type { TriggerConfig } from '@/triggers/types'
|
import type { TriggerConfig } from '@/triggers/types'
|
||||||
import { buildGenericOutputs, circlebackSetupInstructions, circlebackTriggerOptions } from './utils'
|
import { buildMeetingOutputs, circlebackSetupInstructions, circlebackTriggerOptions } from './utils'
|
||||||
|
|
||||||
export const circlebackWebhookTrigger: TriggerConfig = {
|
export const circlebackWebhookTrigger: TriggerConfig = {
|
||||||
id: 'circleback_webhook',
|
id: 'circleback_webhook',
|
||||||
@@ -74,7 +74,7 @@ export const circlebackWebhookTrigger: TriggerConfig = {
|
|||||||
},
|
},
|
||||||
],
|
],
|
||||||
|
|
||||||
outputs: buildGenericOutputs(),
|
outputs: buildMeetingOutputs(),
|
||||||
|
|
||||||
webhook: {
|
webhook: {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
|
|||||||
@@ -31,8 +31,14 @@ export const TRIGGER_PERSISTED_SUBBLOCK_IDS: string[] = [
|
|||||||
/**
|
/**
|
||||||
* Trigger-related subblock IDs that represent runtime metadata. They should remain
|
* Trigger-related subblock IDs that represent runtime metadata. They should remain
|
||||||
* in the workflow state but must not be modified or cleared by diff operations.
|
* in the workflow state but must not be modified or cleared by diff operations.
|
||||||
|
*
|
||||||
|
* Note: 'triggerConfig' is included because it's an aggregate of individual trigger
|
||||||
|
* field subblocks. Those individual fields are compared separately, so comparing
|
||||||
|
* triggerConfig would be redundant. Additionally, the client populates triggerConfig
|
||||||
|
* with default values from the trigger definition on load, which aren't present in
|
||||||
|
* the deployed state, causing false positive change detection.
|
||||||
*/
|
*/
|
||||||
export const TRIGGER_RUNTIME_SUBBLOCK_IDS: string[] = ['webhookId', 'triggerPath']
|
export const TRIGGER_RUNTIME_SUBBLOCK_IDS: string[] = ['webhookId', 'triggerPath', 'triggerConfig']
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Maximum number of consecutive failures before a trigger (schedule/webhook) is auto-disabled.
|
* Maximum number of consecutive failures before a trigger (schedule/webhook) is auto-disabled.
|
||||||
|
|||||||
@@ -116,6 +116,11 @@ export const githubIssueClosedTrigger: TriggerConfig = {
|
|||||||
],
|
],
|
||||||
|
|
||||||
outputs: {
|
outputs: {
|
||||||
|
event_type: {
|
||||||
|
type: 'string',
|
||||||
|
description:
|
||||||
|
'GitHub event type from X-GitHub-Event header (e.g., issues, pull_request, push)',
|
||||||
|
},
|
||||||
action: {
|
action: {
|
||||||
type: 'string',
|
type: 'string',
|
||||||
description: 'Action performed (opened, closed, reopened, edited, etc.)',
|
description: 'Action performed (opened, closed, reopened, edited, etc.)',
|
||||||
|
|||||||
@@ -117,6 +117,10 @@ export const githubIssueCommentTrigger: TriggerConfig = {
|
|||||||
],
|
],
|
||||||
|
|
||||||
outputs: {
|
outputs: {
|
||||||
|
event_type: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'GitHub event type from X-GitHub-Event header (e.g., issue_comment)',
|
||||||
|
},
|
||||||
action: {
|
action: {
|
||||||
type: 'string',
|
type: 'string',
|
||||||
description: 'Action performed (created, edited, deleted)',
|
description: 'Action performed (created, edited, deleted)',
|
||||||
|
|||||||
@@ -137,6 +137,11 @@ export const githubIssueOpenedTrigger: TriggerConfig = {
|
|||||||
],
|
],
|
||||||
|
|
||||||
outputs: {
|
outputs: {
|
||||||
|
event_type: {
|
||||||
|
type: 'string',
|
||||||
|
description:
|
||||||
|
'GitHub event type from X-GitHub-Event header (e.g., issues, pull_request, push)',
|
||||||
|
},
|
||||||
action: {
|
action: {
|
||||||
type: 'string',
|
type: 'string',
|
||||||
description: 'Action performed (opened, closed, reopened, edited, etc.)',
|
description: 'Action performed (opened, closed, reopened, edited, etc.)',
|
||||||
|
|||||||
@@ -117,6 +117,10 @@ export const githubPRClosedTrigger: TriggerConfig = {
|
|||||||
],
|
],
|
||||||
|
|
||||||
outputs: {
|
outputs: {
|
||||||
|
event_type: {
|
||||||
|
type: 'string',
|
||||||
|
description: 'GitHub event type from X-GitHub-Event header (e.g., pull_request)',
|
||||||
|
},
|
||||||
action: {
|
action: {
|
||||||
type: 'string',
|
type: 'string',
|
||||||
description: 'Action performed (opened, closed, synchronize, reopened, edited, etc.)',
|
description: 'Action performed (opened, closed, synchronize, reopened, edited, etc.)',
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user