Files
openclaw/test/provider-timeout.e2e.test.ts
2026-02-16 14:59:30 +00:00

231 lines
7.4 KiB
TypeScript

import { randomUUID } from "node:crypto";
import fs from "node:fs/promises";
import os from "node:os";
import path from "node:path";
import { describe, expect, it } from "vitest";
import { startGatewayWithClient } from "../src/gateway/test-helpers.e2e.js";
import { buildOpenAiResponsesProviderConfig } from "../src/gateway/test-openai-responses-model.js";
type OpenAIResponseStreamEvent =
| { type: "response.output_item.added"; item: Record<string, unknown> }
| { type: "response.output_item.done"; item: Record<string, unknown> }
| {
type: "response.completed";
response: {
status: "completed";
usage: {
input_tokens: number;
output_tokens: number;
total_tokens: number;
};
};
};
function buildOpenAIResponsesSse(text: string): Response {
const events: OpenAIResponseStreamEvent[] = [
{
type: "response.output_item.added",
item: {
type: "message",
id: "msg_test_1",
role: "assistant",
content: [],
status: "in_progress",
},
},
{
type: "response.output_item.done",
item: {
type: "message",
id: "msg_test_1",
role: "assistant",
status: "completed",
content: [{ type: "output_text", text, annotations: [] }],
},
},
{
type: "response.completed",
response: {
status: "completed",
usage: { input_tokens: 10, output_tokens: 10, total_tokens: 20 },
},
},
];
const sse = `${events.map((e) => `data: ${JSON.stringify(e)}\n\n`).join("")}data: [DONE]\n\n`;
const encoder = new TextEncoder();
const body = new ReadableStream<Uint8Array>({
start(controller) {
controller.enqueue(encoder.encode(sse));
controller.close();
},
});
return new Response(body, {
status: 200,
headers: { "content-type": "text/event-stream" },
});
}
function extractPayloadText(result: unknown): string {
const record = result as Record<string, unknown>;
const payloads = Array.isArray(record.payloads) ? record.payloads : [];
const texts = payloads
.map((p) => (p && typeof p === "object" ? (p as Record<string, unknown>).text : undefined))
.filter((t): t is string => typeof t === "string" && t.trim().length > 0);
return texts.join("\n").trim();
}
describe("provider timeouts (e2e)", () => {
it(
"falls back when the primary provider aborts with a timeout-like AbortError",
{ timeout: 60_000 },
async () => {
const prev = {
home: process.env.HOME,
configPath: process.env.OPENCLAW_CONFIG_PATH,
token: process.env.OPENCLAW_GATEWAY_TOKEN,
skipChannels: process.env.OPENCLAW_SKIP_CHANNELS,
skipGmail: process.env.OPENCLAW_SKIP_GMAIL_WATCHER,
skipCron: process.env.OPENCLAW_SKIP_CRON,
skipCanvas: process.env.OPENCLAW_SKIP_CANVAS_HOST,
};
const originalFetch = globalThis.fetch;
const primaryBaseUrl = "https://primary.example/v1";
const fallbackBaseUrl = "https://fallback.example/v1";
const counts = { primary: 0, fallback: 0 };
const fetchImpl = async (input: RequestInfo | URL, init?: RequestInit): Promise<Response> => {
const url =
typeof input === "string" ? input : input instanceof URL ? input.toString() : input.url;
if (url.startsWith(`${primaryBaseUrl}/responses`)) {
counts.primary += 1;
const err = new Error("request was aborted");
err.name = "AbortError";
throw err;
}
if (url.startsWith(`${fallbackBaseUrl}/responses`)) {
counts.fallback += 1;
return buildOpenAIResponsesSse("fallback-ok");
}
if (!originalFetch) {
throw new Error(`fetch is not available (url=${url})`);
}
return await originalFetch(input, init);
};
(globalThis as unknown as { fetch: unknown }).fetch = fetchImpl;
const tempHome = await fs.mkdtemp(path.join(os.tmpdir(), "openclaw-timeout-e2e-"));
process.env.HOME = tempHome;
process.env.OPENCLAW_SKIP_CHANNELS = "1";
process.env.OPENCLAW_SKIP_GMAIL_WATCHER = "1";
process.env.OPENCLAW_SKIP_CRON = "1";
process.env.OPENCLAW_SKIP_CANVAS_HOST = "1";
const token = `test-${randomUUID()}`;
process.env.OPENCLAW_GATEWAY_TOKEN = token;
const configDir = path.join(tempHome, ".openclaw");
await fs.mkdir(configDir, { recursive: true });
const configPath = path.join(configDir, "openclaw.json");
const cfg = {
agents: {
defaults: {
model: {
primary: "primary/gpt-5.2",
fallbacks: ["fallback/gpt-5.2"],
},
},
},
models: {
mode: "replace",
providers: {
primary: buildOpenAiResponsesProviderConfig(primaryBaseUrl),
fallback: buildOpenAiResponsesProviderConfig(fallbackBaseUrl),
},
},
gateway: { auth: { token } },
};
const { server, client } = await startGatewayWithClient({
cfg,
configPath,
token,
clientDisplayName: "vitest-timeout-fallback",
});
try {
const sessionKey = "agent:dev:timeout-fallback";
await client.request("sessions.patch", {
key: sessionKey,
model: "primary/gpt-5.2",
});
const runId = randomUUID();
const payload = await client.request<{
status?: unknown;
result?: unknown;
}>(
"agent",
{
sessionKey,
idempotencyKey: `idem-${runId}`,
message: "say fallback-ok",
deliver: false,
},
{ expectFinal: true },
);
expect(payload?.status).toBe("ok");
const text = extractPayloadText(payload?.result);
expect(text).toContain("fallback-ok");
expect(counts.primary).toBeGreaterThan(0);
expect(counts.fallback).toBeGreaterThan(0);
} finally {
client.stop();
await server.close({ reason: "timeout fallback test complete" });
await fs.rm(tempHome, { recursive: true, force: true });
(globalThis as unknown as { fetch: unknown }).fetch = originalFetch;
if (prev.home === undefined) {
delete process.env.HOME;
} else {
process.env.HOME = prev.home;
}
if (prev.configPath === undefined) {
delete process.env.OPENCLAW_CONFIG_PATH;
} else {
process.env.OPENCLAW_CONFIG_PATH = prev.configPath;
}
if (prev.token === undefined) {
delete process.env.OPENCLAW_GATEWAY_TOKEN;
} else {
process.env.OPENCLAW_GATEWAY_TOKEN = prev.token;
}
if (prev.skipChannels === undefined) {
delete process.env.OPENCLAW_SKIP_CHANNELS;
} else {
process.env.OPENCLAW_SKIP_CHANNELS = prev.skipChannels;
}
if (prev.skipGmail === undefined) {
delete process.env.OPENCLAW_SKIP_GMAIL_WATCHER;
} else {
process.env.OPENCLAW_SKIP_GMAIL_WATCHER = prev.skipGmail;
}
if (prev.skipCron === undefined) {
delete process.env.OPENCLAW_SKIP_CRON;
} else {
process.env.OPENCLAW_SKIP_CRON = prev.skipCron;
}
if (prev.skipCanvas === undefined) {
delete process.env.OPENCLAW_SKIP_CANVAS_HOST;
} else {
process.env.OPENCLAW_SKIP_CANVAS_HOST = prev.skipCanvas;
}
}
},
);
});