import { randomBytes } from "crypto"; import fs from "fs/promises"; import path from "path"; export type TokenUsageRecord = { id: string; label: string; tokensUsed: number; tokenLimit?: number; inputTokens?: number; outputTokens?: number; cacheReadTokens?: number; cacheWriteTokens?: number; model?: string; provider?: string; runId?: string; sessionId?: string; sessionKey?: string; createdAt: string; }; function makeId() { return `usage_${Date.now().toString(36)}_${randomBytes(4).toString("hex")}`; } async function readJsonArray(file: string): Promise { try { const raw = await fs.readFile(file, "utf-8"); const parsed = JSON.parse(raw); if (!Array.isArray(parsed)) { // Valid JSON but unexpected shape (object, number, string, …). // Returning [] here would cause appendRecord to overwrite the file // with only the new entry, silently deleting prior data. throw Object.assign( new Error( `token-usage.json contains valid JSON but is not an array (got ${typeof parsed})`, ), { code: "ERR_UNEXPECTED_TOKEN_LOG_SHAPE" }, ); } return parsed as TokenUsageRecord[]; } catch (err) { // File does not exist yet — start with an empty array. if ((err as NodeJS.ErrnoException).code === "ENOENT") { return []; } // Any other error (malformed JSON, permission denied, partial write, …) // must propagate so appendRecord aborts and the existing file is not // silently overwritten with only the new entry. throw err; } } // --------------------------------------------------------------------------- // Cross-process file lock // // The in-memory writeQueues Map serialises writes within a single Node // process, but two concurrent OpenClaw processes targeting the same // workspaceDir can still race: both read the same snapshot before either // writes. We guard against that with an advisory lock file (.lock) using // O_EXCL (create-exclusive), which is atomic on POSIX filesystems. // // Lock acquisition retries with a fixed interval up to LOCK_TIMEOUT_MS. // On timeout ERR_LOCK_TIMEOUT is thrown; the lock file is intentionally left // untouched so that an active holder's mutual exclusion is never broken. // --------------------------------------------------------------------------- const LOCK_RETRY_MS = 50; const LOCK_TIMEOUT_MS = 5_000; async function withFileLock(lockPath: string, fn: () => Promise): Promise { const deadline = Date.now() + LOCK_TIMEOUT_MS; while (Date.now() < deadline) { let fh: fs.FileHandle | undefined; try { // wx = O_WRONLY | O_CREAT | O_EXCL — fails if the file already exists fh = await fs.open(lockPath, "wx"); await fh.close(); fh = undefined; try { return await fn(); } finally { await fs.unlink(lockPath).catch(() => {}); } } catch (err) { await fh?.close().catch(() => {}); if ((err as NodeJS.ErrnoException).code !== "EEXIST") { throw err; } // Another process holds the lock — wait and retry. await new Promise((r) => setTimeout(r, LOCK_RETRY_MS)); } } // Timed out without acquiring the lock. We deliberately do NOT delete the // lock file here: the holder may still be active (slow disk, large file), // and removing a live lock would break mutual exclusion and allow concurrent // read-modify-write cycles to overwrite each other's entries. throw Object.assign(new Error(`Could not acquire lock ${lockPath} within ${LOCK_TIMEOUT_MS}ms`), { code: "ERR_LOCK_TIMEOUT", }); } async function appendRecord(file: string, entry: TokenUsageRecord): Promise { const lockPath = `${file}.lock`; await withFileLock(lockPath, async () => { const records = await readJsonArray(file); records.push(entry); await fs.writeFile(file, JSON.stringify(records, null, 2)); }); } // Per-file write queue: serialises concurrent recordTokenUsage() calls within // the same process so they do not all contend on the cross-process file lock. const writeQueues = new Map>(); export async function recordTokenUsage(params: { workspaceDir: string; runId?: string; sessionId?: string; sessionKey?: string; provider?: string; model?: string; label: string; usage?: { input?: number; output?: number; cacheRead?: number; cacheWrite?: number; total?: number; }; }) { const usage = params.usage; if (!usage) { return; } const total = usage.total ?? (usage.input ?? 0) + (usage.output ?? 0) + (usage.cacheRead ?? 0) + (usage.cacheWrite ?? 0); if (!total || total <= 0) { return; } const memoryDir = path.join(params.workspaceDir, "memory"); const file = path.join(memoryDir, "token-usage.json"); await fs.mkdir(memoryDir, { recursive: true }); const entry: TokenUsageRecord = { id: makeId(), label: params.label, tokensUsed: Math.trunc(total), ...(usage.input != null && usage.input > 0 && { inputTokens: Math.trunc(usage.input) }), ...(usage.output != null && usage.output > 0 && { outputTokens: Math.trunc(usage.output) }), ...(usage.cacheRead != null && usage.cacheRead > 0 && { cacheReadTokens: Math.trunc(usage.cacheRead) }), ...(usage.cacheWrite != null && usage.cacheWrite > 0 && { cacheWriteTokens: Math.trunc(usage.cacheWrite) }), model: params.model, provider: params.provider, runId: params.runId, sessionId: params.sessionId, sessionKey: params.sessionKey, createdAt: new Date().toISOString(), }; const queued = writeQueues.get(file) ?? Promise.resolve(); const next = queued.then(() => appendRecord(file, entry)); writeQueues.set( file, next.catch(() => {}), ); await next; }