mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-09 02:18:25 +00:00
test: move gateway server integration suites out of e2e
This commit is contained in:
430
src/gateway/server.chat.gateway-server-chat-b.test.ts
Normal file
430
src/gateway/server.chat.gateway-server-chat-b.test.ts
Normal file
@@ -0,0 +1,430 @@
|
||||
import fs from "node:fs/promises";
|
||||
import os from "node:os";
|
||||
import path from "node:path";
|
||||
import { describe, expect, test, vi } from "vitest";
|
||||
import type { GetReplyOptions } from "../auto-reply/types.js";
|
||||
import { __setMaxChatHistoryMessagesBytesForTest } from "./server-constants.js";
|
||||
import {
|
||||
connectOk,
|
||||
getReplyFromConfig,
|
||||
installGatewayTestHooks,
|
||||
onceMessage,
|
||||
rpcReq,
|
||||
startServerWithClient,
|
||||
testState,
|
||||
writeSessionStore,
|
||||
} from "./test-helpers.js";
|
||||
|
||||
installGatewayTestHooks({ scope: "suite" });
|
||||
|
||||
const sendReq = (
|
||||
ws: { send: (payload: string) => void },
|
||||
id: string,
|
||||
method: string,
|
||||
params: unknown,
|
||||
) => {
|
||||
ws.send(
|
||||
JSON.stringify({
|
||||
type: "req",
|
||||
id,
|
||||
method,
|
||||
params,
|
||||
}),
|
||||
);
|
||||
};
|
||||
|
||||
async function withGatewayChatHarness(
|
||||
run: (ctx: {
|
||||
ws: Awaited<ReturnType<typeof startServerWithClient>>["ws"];
|
||||
createSessionDir: () => Promise<string>;
|
||||
}) => Promise<void>,
|
||||
) {
|
||||
const tempDirs: string[] = [];
|
||||
const { server, ws } = await startServerWithClient();
|
||||
const createSessionDir = async () => {
|
||||
const sessionDir = await fs.mkdtemp(path.join(os.tmpdir(), "openclaw-gw-"));
|
||||
tempDirs.push(sessionDir);
|
||||
testState.sessionStorePath = path.join(sessionDir, "sessions.json");
|
||||
return sessionDir;
|
||||
};
|
||||
|
||||
try {
|
||||
await run({ ws, createSessionDir });
|
||||
} finally {
|
||||
__setMaxChatHistoryMessagesBytesForTest();
|
||||
testState.sessionStorePath = undefined;
|
||||
ws.close();
|
||||
await server.close();
|
||||
await Promise.all(tempDirs.map((dir) => fs.rm(dir, { recursive: true, force: true })));
|
||||
}
|
||||
}
|
||||
|
||||
async function writeMainSessionStore() {
|
||||
await writeSessionStore({
|
||||
entries: {
|
||||
main: { sessionId: "sess-main", updatedAt: Date.now() },
|
||||
},
|
||||
});
|
||||
}
|
||||
|
||||
async function writeMainSessionTranscript(sessionDir: string, lines: string[]) {
|
||||
await fs.writeFile(path.join(sessionDir, "sess-main.jsonl"), `${lines.join("\n")}\n`, "utf-8");
|
||||
}
|
||||
|
||||
async function fetchHistoryMessages(
|
||||
ws: Awaited<ReturnType<typeof startServerWithClient>>["ws"],
|
||||
): Promise<unknown[]> {
|
||||
const historyRes = await rpcReq<{ messages?: unknown[] }>(ws, "chat.history", {
|
||||
sessionKey: "main",
|
||||
limit: 1000,
|
||||
});
|
||||
expect(historyRes.ok).toBe(true);
|
||||
return historyRes.payload?.messages ?? [];
|
||||
}
|
||||
|
||||
describe("gateway server chat", () => {
|
||||
test("smoke: caps history payload and preserves routing metadata", async () => {
|
||||
await withGatewayChatHarness(async ({ ws, createSessionDir }) => {
|
||||
const historyMaxBytes = 192 * 1024;
|
||||
__setMaxChatHistoryMessagesBytesForTest(historyMaxBytes);
|
||||
await connectOk(ws);
|
||||
|
||||
const sessionDir = await createSessionDir();
|
||||
await writeMainSessionStore();
|
||||
|
||||
const bigText = "x".repeat(4_000);
|
||||
const historyLines: string[] = [];
|
||||
for (let i = 0; i < 60; i += 1) {
|
||||
historyLines.push(
|
||||
JSON.stringify({
|
||||
message: {
|
||||
role: "user",
|
||||
content: [{ type: "text", text: `${i}:${bigText}` }],
|
||||
timestamp: Date.now() + i,
|
||||
},
|
||||
}),
|
||||
);
|
||||
}
|
||||
await writeMainSessionTranscript(sessionDir, historyLines);
|
||||
const messages = await fetchHistoryMessages(ws);
|
||||
const bytes = Buffer.byteLength(JSON.stringify(messages), "utf8");
|
||||
expect(bytes).toBeLessThanOrEqual(historyMaxBytes);
|
||||
expect(messages.length).toBeLessThan(60);
|
||||
|
||||
await writeSessionStore({
|
||||
entries: {
|
||||
main: {
|
||||
sessionId: "sess-main",
|
||||
updatedAt: Date.now(),
|
||||
lastChannel: "whatsapp",
|
||||
lastTo: "+1555",
|
||||
},
|
||||
},
|
||||
});
|
||||
|
||||
const sendRes = await rpcReq(ws, "chat.send", {
|
||||
sessionKey: "main",
|
||||
message: "hello",
|
||||
idempotencyKey: "idem-route",
|
||||
});
|
||||
expect(sendRes.ok).toBe(true);
|
||||
|
||||
const sessionStorePath = testState.sessionStorePath;
|
||||
if (!sessionStorePath) {
|
||||
throw new Error("expected session store path");
|
||||
}
|
||||
const stored = JSON.parse(await fs.readFile(sessionStorePath, "utf-8")) as Record<
|
||||
string,
|
||||
{ lastChannel?: string; lastTo?: string } | undefined
|
||||
>;
|
||||
expect(stored["agent:main:main"]?.lastChannel).toBe("whatsapp");
|
||||
expect(stored["agent:main:main"]?.lastTo).toBe("+1555");
|
||||
});
|
||||
});
|
||||
|
||||
test("chat.send does not force-disable block streaming", async () => {
|
||||
await withGatewayChatHarness(async ({ ws, createSessionDir }) => {
|
||||
const spy = getReplyFromConfig;
|
||||
await connectOk(ws);
|
||||
|
||||
await createSessionDir();
|
||||
await writeMainSessionStore();
|
||||
testState.agentConfig = { blockStreamingDefault: "on" };
|
||||
try {
|
||||
spy.mockClear();
|
||||
let capturedOpts: GetReplyOptions | undefined;
|
||||
spy.mockImplementationOnce(async (_ctx: unknown, opts?: GetReplyOptions) => {
|
||||
capturedOpts = opts;
|
||||
return undefined;
|
||||
});
|
||||
|
||||
const sendRes = await rpcReq(ws, "chat.send", {
|
||||
sessionKey: "main",
|
||||
message: "hello",
|
||||
idempotencyKey: "idem-block-streaming",
|
||||
});
|
||||
expect(sendRes.ok).toBe(true);
|
||||
|
||||
await vi.waitFor(
|
||||
() => {
|
||||
expect(spy.mock.calls.length).toBeGreaterThan(0);
|
||||
},
|
||||
{ timeout: 2_000, interval: 10 },
|
||||
);
|
||||
|
||||
expect(capturedOpts?.disableBlockStreaming).toBeUndefined();
|
||||
} finally {
|
||||
testState.agentConfig = undefined;
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
test("chat.history hard-caps single oversized nested payloads", async () => {
|
||||
await withGatewayChatHarness(async ({ ws, createSessionDir }) => {
|
||||
const historyMaxBytes = 64 * 1024;
|
||||
__setMaxChatHistoryMessagesBytesForTest(historyMaxBytes);
|
||||
await connectOk(ws);
|
||||
|
||||
const sessionDir = await createSessionDir();
|
||||
await writeMainSessionStore();
|
||||
|
||||
const hugeNestedText = "n".repeat(450_000);
|
||||
const oversizedLine = JSON.stringify({
|
||||
message: {
|
||||
role: "assistant",
|
||||
timestamp: Date.now(),
|
||||
content: [
|
||||
{
|
||||
type: "tool_result",
|
||||
toolUseId: "tool-1",
|
||||
output: {
|
||||
nested: {
|
||||
payload: hugeNestedText,
|
||||
},
|
||||
},
|
||||
},
|
||||
],
|
||||
},
|
||||
});
|
||||
await writeMainSessionTranscript(sessionDir, [oversizedLine]);
|
||||
const messages = await fetchHistoryMessages(ws);
|
||||
expect(messages.length).toBe(1);
|
||||
|
||||
const serialized = JSON.stringify(messages);
|
||||
const bytes = Buffer.byteLength(serialized, "utf8");
|
||||
expect(bytes).toBeLessThanOrEqual(historyMaxBytes);
|
||||
expect(serialized).toContain("[chat.history omitted: message too large]");
|
||||
expect(serialized.includes(hugeNestedText.slice(0, 256))).toBe(false);
|
||||
});
|
||||
});
|
||||
|
||||
test("chat.history keeps recent small messages when latest message is oversized", async () => {
|
||||
await withGatewayChatHarness(async ({ ws, createSessionDir }) => {
|
||||
const historyMaxBytes = 64 * 1024;
|
||||
__setMaxChatHistoryMessagesBytesForTest(historyMaxBytes);
|
||||
await connectOk(ws);
|
||||
|
||||
const sessionDir = await createSessionDir();
|
||||
await writeMainSessionStore();
|
||||
|
||||
const baseText = "s".repeat(1_200);
|
||||
const lines: string[] = [];
|
||||
for (let i = 0; i < 30; i += 1) {
|
||||
lines.push(
|
||||
JSON.stringify({
|
||||
message: {
|
||||
role: "user",
|
||||
timestamp: Date.now() + i,
|
||||
content: [{ type: "text", text: `small-${i}:${baseText}` }],
|
||||
},
|
||||
}),
|
||||
);
|
||||
}
|
||||
|
||||
const hugeNestedText = "z".repeat(450_000);
|
||||
lines.push(
|
||||
JSON.stringify({
|
||||
message: {
|
||||
role: "assistant",
|
||||
timestamp: Date.now() + 1_000,
|
||||
content: [
|
||||
{
|
||||
type: "tool_result",
|
||||
toolUseId: "tool-1",
|
||||
output: {
|
||||
nested: {
|
||||
payload: hugeNestedText,
|
||||
},
|
||||
},
|
||||
},
|
||||
],
|
||||
},
|
||||
}),
|
||||
);
|
||||
|
||||
await writeMainSessionTranscript(sessionDir, lines);
|
||||
const messages = await fetchHistoryMessages(ws);
|
||||
const serialized = JSON.stringify(messages);
|
||||
const bytes = Buffer.byteLength(serialized, "utf8");
|
||||
|
||||
expect(bytes).toBeLessThanOrEqual(historyMaxBytes);
|
||||
expect(messages.length).toBeGreaterThan(1);
|
||||
expect(serialized).toContain("small-29:");
|
||||
expect(serialized).toContain("[chat.history omitted: message too large]");
|
||||
expect(serialized.includes(hugeNestedText.slice(0, 256))).toBe(false);
|
||||
});
|
||||
});
|
||||
|
||||
test("chat.history strips inline directives from displayed message text", async () => {
|
||||
await withGatewayChatHarness(async ({ ws, createSessionDir }) => {
|
||||
await connectOk(ws);
|
||||
|
||||
const sessionDir = await createSessionDir();
|
||||
await writeMainSessionStore();
|
||||
|
||||
const lines = [
|
||||
JSON.stringify({
|
||||
message: {
|
||||
role: "assistant",
|
||||
content: [
|
||||
{ type: "text", text: "Hello [[reply_to_current]] world [[audio_as_voice]]" },
|
||||
],
|
||||
timestamp: Date.now(),
|
||||
},
|
||||
}),
|
||||
JSON.stringify({
|
||||
message: {
|
||||
role: "assistant",
|
||||
content: "A [[reply_to:abc-123]] B",
|
||||
timestamp: Date.now() + 1,
|
||||
},
|
||||
}),
|
||||
JSON.stringify({
|
||||
message: {
|
||||
role: "assistant",
|
||||
text: "[[ reply_to : 456 ]] C",
|
||||
timestamp: Date.now() + 2,
|
||||
},
|
||||
}),
|
||||
JSON.stringify({
|
||||
message: {
|
||||
role: "assistant",
|
||||
content: [{ type: "text", text: " keep padded " }],
|
||||
timestamp: Date.now() + 3,
|
||||
},
|
||||
}),
|
||||
];
|
||||
await writeMainSessionTranscript(sessionDir, lines);
|
||||
const messages = await fetchHistoryMessages(ws);
|
||||
expect(messages.length).toBe(4);
|
||||
|
||||
const serialized = JSON.stringify(messages);
|
||||
expect(serialized.includes("[[reply_to")).toBe(false);
|
||||
expect(serialized.includes("[[audio_as_voice]]")).toBe(false);
|
||||
|
||||
const first = messages[0] as { content?: Array<{ text?: string }> };
|
||||
const second = messages[1] as { content?: string };
|
||||
const third = messages[2] as { text?: string };
|
||||
const fourth = messages[3] as { content?: Array<{ text?: string }> };
|
||||
|
||||
expect(first.content?.[0]?.text?.replace(/\s+/g, " ").trim()).toBe("Hello world");
|
||||
expect(second.content?.replace(/\s+/g, " ").trim()).toBe("A B");
|
||||
expect(third.text?.replace(/\s+/g, " ").trim()).toBe("C");
|
||||
expect(fourth.content?.[0]?.text).toBe(" keep padded ");
|
||||
});
|
||||
});
|
||||
|
||||
test("smoke: supports abort and idempotent completion", async () => {
|
||||
await withGatewayChatHarness(async ({ ws, createSessionDir }) => {
|
||||
const spy = getReplyFromConfig;
|
||||
let aborted = false;
|
||||
await connectOk(ws);
|
||||
|
||||
await createSessionDir();
|
||||
await writeMainSessionStore();
|
||||
|
||||
spy.mockClear();
|
||||
spy.mockImplementationOnce(async (_ctx, opts) => {
|
||||
opts?.onAgentRunStart?.(opts.runId ?? "idem-abort-1");
|
||||
const signal = opts?.abortSignal;
|
||||
await new Promise<void>((resolve) => {
|
||||
if (!signal || signal.aborted) {
|
||||
aborted = Boolean(signal?.aborted);
|
||||
resolve();
|
||||
return;
|
||||
}
|
||||
signal.addEventListener(
|
||||
"abort",
|
||||
() => {
|
||||
aborted = true;
|
||||
resolve();
|
||||
},
|
||||
{ once: true },
|
||||
);
|
||||
});
|
||||
return undefined;
|
||||
});
|
||||
|
||||
const sendResP = onceMessage(ws, (o) => o.type === "res" && o.id === "send-abort-1", 8_000);
|
||||
sendReq(ws, "send-abort-1", "chat.send", {
|
||||
sessionKey: "main",
|
||||
message: "hello",
|
||||
idempotencyKey: "idem-abort-1",
|
||||
timeoutMs: 30_000,
|
||||
});
|
||||
|
||||
const sendRes = await sendResP;
|
||||
expect(sendRes.ok).toBe(true);
|
||||
await vi.waitFor(
|
||||
() => {
|
||||
expect(spy.mock.calls.length).toBeGreaterThan(0);
|
||||
},
|
||||
{ timeout: 2_000, interval: 10 },
|
||||
);
|
||||
|
||||
const inFlight = await rpcReq<{ status?: string }>(ws, "chat.send", {
|
||||
sessionKey: "main",
|
||||
message: "hello",
|
||||
idempotencyKey: "idem-abort-1",
|
||||
});
|
||||
expect(inFlight.ok).toBe(true);
|
||||
expect(["started", "in_flight", "ok"]).toContain(inFlight.payload?.status ?? "");
|
||||
|
||||
const abortRes = await rpcReq<{ aborted?: boolean }>(ws, "chat.abort", {
|
||||
sessionKey: "main",
|
||||
runId: "idem-abort-1",
|
||||
});
|
||||
expect(abortRes.ok).toBe(true);
|
||||
expect(abortRes.payload?.aborted).toBe(true);
|
||||
await vi.waitFor(
|
||||
() => {
|
||||
expect(aborted).toBe(true);
|
||||
},
|
||||
{ timeout: 2_000, interval: 10 },
|
||||
);
|
||||
|
||||
spy.mockClear();
|
||||
spy.mockResolvedValueOnce(undefined);
|
||||
|
||||
const completeRes = await rpcReq<{ status?: string }>(ws, "chat.send", {
|
||||
sessionKey: "main",
|
||||
message: "hello",
|
||||
idempotencyKey: "idem-complete-1",
|
||||
});
|
||||
expect(completeRes.ok).toBe(true);
|
||||
|
||||
await vi.waitFor(
|
||||
async () => {
|
||||
const again = await rpcReq<{ status?: string }>(ws, "chat.send", {
|
||||
sessionKey: "main",
|
||||
message: "hello",
|
||||
idempotencyKey: "idem-complete-1",
|
||||
});
|
||||
expect(again.ok).toBe(true);
|
||||
expect(again.payload?.status).toBe("ok");
|
||||
},
|
||||
{ timeout: 2_000, interval: 10 },
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
Reference in New Issue
Block a user