perf(test): prebuild download archives and cache apply module

This commit is contained in:
Peter Steinberger
2026-02-22 17:16:04 +00:00
parent 2962e5a383
commit c964d21d74
2 changed files with 24 additions and 44 deletions

View File

@@ -1,6 +1,6 @@
import fs from "node:fs/promises";
import path from "node:path";
import { afterEach, beforeEach, describe, expect, it, vi } from "vitest";
import { afterEach, beforeAll, beforeEach, describe, expect, it, vi } from "vitest";
import { resolveApiKeyForProvider } from "../agents/model-auth.js";
import type { MsgContext } from "../auto-reply/templating.js";
import type { OpenClawConfig } from "../config/config.js";
@@ -29,9 +29,7 @@ vi.mock("../process/exec.js", () => ({
runExec: vi.fn(),
}));
async function loadApply() {
return await import("./apply.js");
}
let applyMediaUnderstanding: typeof import("./apply.js").applyMediaUnderstanding;
const TEMP_MEDIA_PREFIX = "openclaw-media-";
const tempMediaDirs: string[] = [];
@@ -137,7 +135,6 @@ async function applyWithDisabledMedia(params: {
mediaType?: string;
cfg?: OpenClawConfig;
}) {
const { applyMediaUnderstanding } = await loadApply();
const ctx: MsgContext = {
Body: params.body,
MediaPath: params.mediaPath,
@@ -164,6 +161,10 @@ describe("applyMediaUnderstanding", () => {
const mockedResolveApiKey = vi.mocked(resolveApiKeyForProvider);
const mockedFetchRemoteMedia = vi.mocked(fetchRemoteMedia);
beforeAll(async () => {
({ applyMediaUnderstanding } = await import("./apply.js"));
});
beforeEach(() => {
mockedResolveApiKey.mockClear();
mockedFetchRemoteMedia.mockClear();
@@ -183,7 +184,6 @@ describe("applyMediaUnderstanding", () => {
});
it("sets Transcript and replaces Body when audio transcription succeeds", async () => {
const { applyMediaUnderstanding } = await loadApply();
const ctx = await createAudioCtx();
const result = await applyMediaUnderstanding({
ctx,
@@ -202,7 +202,6 @@ describe("applyMediaUnderstanding", () => {
});
it("skips file blocks for text-like audio when transcription succeeds", async () => {
const { applyMediaUnderstanding } = await loadApply();
const ctx = await createAudioCtx({
fileName: "data.mp3",
mediaType: "audio/mpeg",
@@ -221,7 +220,6 @@ describe("applyMediaUnderstanding", () => {
});
it("keeps caption for command parsing when audio has user text", async () => {
const { applyMediaUnderstanding } = await loadApply();
const ctx = await createAudioCtx({
body: "<media:audio> /capture status",
});
@@ -241,7 +239,6 @@ describe("applyMediaUnderstanding", () => {
});
it("handles URL-only attachments for audio transcription", async () => {
const { applyMediaUnderstanding } = await loadApply();
const ctx: MsgContext = {
Body: "<media:audio>",
MediaUrl: "https://example.com/note.ogg",
@@ -281,7 +278,6 @@ describe("applyMediaUnderstanding", () => {
});
it("skips audio transcription when attachment exceeds maxBytes", async () => {
const { applyMediaUnderstanding } = await loadApply();
const ctx = await createAudioCtx({
fileName: "large.wav",
mediaType: "audio/wav",
@@ -312,7 +308,6 @@ describe("applyMediaUnderstanding", () => {
});
it("falls back to CLI model when provider fails", async () => {
const { applyMediaUnderstanding } = await loadApply();
const ctx = await createAudioCtx();
const cfg: OpenClawConfig = {
tools: {
@@ -357,7 +352,6 @@ describe("applyMediaUnderstanding", () => {
});
it("uses CLI image understanding and preserves caption for commands", async () => {
const { applyMediaUnderstanding } = await loadApply();
const imagePath = await createTempMediaFile({
fileName: "photo.jpg",
content: "image-bytes",
@@ -405,7 +399,6 @@ describe("applyMediaUnderstanding", () => {
});
it("uses shared media models list when capability config is missing", async () => {
const { applyMediaUnderstanding } = await loadApply();
const imagePath = await createTempMediaFile({
fileName: "shared.jpg",
content: "image-bytes",
@@ -447,7 +440,6 @@ describe("applyMediaUnderstanding", () => {
});
it("uses active model when enabled and models are missing", async () => {
const { applyMediaUnderstanding } = await loadApply();
const audioPath = await createTempMediaFile({
fileName: "fallback.ogg",
content: Buffer.from([0, 255, 0, 1, 2, 3, 4, 5, 6]),
@@ -485,7 +477,6 @@ describe("applyMediaUnderstanding", () => {
});
it("handles multiple audio attachments when attachment mode is all", async () => {
const { applyMediaUnderstanding } = await loadApply();
const dir = await createTempMediaDir();
const audioBytes = Buffer.from([200, 201, 202, 203, 204, 205, 206, 207, 208]);
const audioPathA = path.join(dir, "note-a.ogg");
@@ -529,7 +520,6 @@ describe("applyMediaUnderstanding", () => {
});
it("orders mixed media outputs as image, audio, video", async () => {
const { applyMediaUnderstanding } = await loadApply();
const dir = await createTempMediaDir();
const imagePath = path.join(dir, "photo.jpg");
const audioPath = path.join(dir, "note.ogg");