chore: Enable "curly" rule to avoid single-statement if confusion/errors.

This commit is contained in:
cpojer
2026-01-31 16:19:20 +09:00
parent 009b16fab8
commit 5ceff756e1
1266 changed files with 27871 additions and 9393 deletions

View File

@@ -56,7 +56,9 @@ function getOpenAiHeaders(
}
function splitOpenAiBatchRequests(requests: OpenAiBatchRequest[]): OpenAiBatchRequest[][] {
if (requests.length <= OPENAI_BATCH_MAX_REQUESTS) return [requests];
if (requests.length <= OPENAI_BATCH_MAX_REQUESTS) {
return [requests];
}
const groups: OpenAiBatchRequest[][] = [];
for (let i = 0; i < requests.length; i += OPENAI_BATCH_MAX_REQUESTS) {
groups.push(requests.slice(i, i + OPENAI_BATCH_MAX_REQUESTS));
@@ -163,7 +165,9 @@ async function fetchOpenAiFileContent(params: {
}
function parseOpenAiBatchOutput(text: string): OpenAiBatchOutputLine[] {
if (!text.trim()) return [];
if (!text.trim()) {
return [];
}
return text
.split("\n")
.map((line) => line.trim())
@@ -242,7 +246,9 @@ async function waitForOpenAiBatch(params: {
}
async function runWithConcurrency<T>(tasks: Array<() => Promise<T>>, limit: number): Promise<T[]> {
if (tasks.length === 0) return [];
if (tasks.length === 0) {
return [];
}
const resolvedLimit = Math.max(1, Math.min(limit, tasks.length));
const results: T[] = Array.from({ length: tasks.length });
let next = 0;
@@ -250,10 +256,14 @@ async function runWithConcurrency<T>(tasks: Array<() => Promise<T>>, limit: numb
const workers = Array.from({ length: resolvedLimit }, async () => {
while (true) {
if (firstError) return;
if (firstError) {
return;
}
const index = next;
next += 1;
if (index >= tasks.length) return;
if (index >= tasks.length) {
return;
}
try {
results[index] = await tasks[index]();
} catch (err) {
@@ -264,7 +274,9 @@ async function runWithConcurrency<T>(tasks: Array<() => Promise<T>>, limit: numb
});
await Promise.allSettled(workers);
if (firstError) throw firstError;
if (firstError) {
throw firstError;
}
return results;
}
@@ -278,7 +290,9 @@ export async function runOpenAiEmbeddingBatches(params: {
concurrency: number;
debug?: (message: string, data?: Record<string, unknown>) => void;
}): Promise<Map<string, number[]>> {
if (params.requests.length === 0) return new Map();
if (params.requests.length === 0) {
return new Map();
}
const groups = splitOpenAiBatchRequests(params.requests);
const byCustomId = new Map<string, number[]>();
@@ -335,7 +349,9 @@ export async function runOpenAiEmbeddingBatches(params: {
for (const line of outputLines) {
const customId = line.custom_id;
if (!customId) continue;
if (!customId) {
continue;
}
remaining.delete(customId);
if (line.error?.message) {
errors.push(`${customId}: ${line.error.message}`);