mirror of
https://github.com/ksyasuda/SubMiner.git
synced 2026-03-23 00:11:28 -07:00
chore: add shared log-file source for diagnostics
This commit is contained in:
@@ -1,5 +0,0 @@
|
||||
type: changed
|
||||
area: subtitle sidebar
|
||||
|
||||
- Added subtitle sidebar state and behavior updates, including startup-auto-open controls and resume positioning improvements.
|
||||
- Fixed subtitle prefetch and embedded overlay passthrough sync between sidebar and overlay subtitle rendering.
|
||||
65
src/shared/log-files.test.ts
Normal file
65
src/shared/log-files.test.ts
Normal file
@@ -0,0 +1,65 @@
|
||||
import test from 'node:test';
|
||||
import assert from 'node:assert/strict';
|
||||
import fs from 'node:fs';
|
||||
import os from 'node:os';
|
||||
import path from 'node:path';
|
||||
import {
|
||||
appendLogLine,
|
||||
pruneLogFiles,
|
||||
resolveDefaultLogFilePath,
|
||||
} from './log-files';
|
||||
|
||||
test('resolveDefaultLogFilePath uses app prefix by default', () => {
|
||||
const resolved = resolveDefaultLogFilePath('app', {
|
||||
platform: 'linux',
|
||||
homeDir: '/home/tester',
|
||||
});
|
||||
|
||||
assert.equal(
|
||||
resolved,
|
||||
path.join(
|
||||
'/home/tester',
|
||||
'.config',
|
||||
'SubMiner',
|
||||
'logs',
|
||||
`app-${new Date().toISOString().slice(0, 10)}.log`,
|
||||
),
|
||||
);
|
||||
});
|
||||
|
||||
test('pruneLogFiles removes logs older than retention window', () => {
|
||||
const logsDir = fs.mkdtempSync(path.join(os.tmpdir(), 'subminer-log-prune-'));
|
||||
const stalePath = path.join(logsDir, 'app-old.log');
|
||||
const freshPath = path.join(logsDir, 'app-fresh.log');
|
||||
fs.writeFileSync(stalePath, 'stale\n', 'utf8');
|
||||
fs.writeFileSync(freshPath, 'fresh\n', 'utf8');
|
||||
const now = new Date('2026-03-22T12:00:00.000Z');
|
||||
fs.utimesSync(stalePath, new Date('2026-03-01T12:00:00.000Z'), new Date('2026-03-01T12:00:00.000Z'));
|
||||
fs.utimesSync(freshPath, new Date('2026-03-21T12:00:00.000Z'), new Date('2026-03-21T12:00:00.000Z'));
|
||||
|
||||
try {
|
||||
pruneLogFiles(logsDir, { retentionDays: 7, now });
|
||||
|
||||
assert.equal(fs.existsSync(stalePath), false);
|
||||
assert.equal(fs.existsSync(freshPath), true);
|
||||
} finally {
|
||||
fs.rmSync(logsDir, { recursive: true, force: true });
|
||||
}
|
||||
});
|
||||
|
||||
test('appendLogLine trims oversized logs to newest bytes', () => {
|
||||
const logsDir = fs.mkdtempSync(path.join(os.tmpdir(), 'subminer-log-trim-'));
|
||||
const logPath = path.join(logsDir, 'app.log');
|
||||
|
||||
try {
|
||||
appendLogLine(logPath, '012345678901234567890123456789', { maxBytes: 48, retentionDays: 30 });
|
||||
appendLogLine(logPath, 'abcdefghijabcdefghijabcdefghij', { maxBytes: 48, retentionDays: 30 });
|
||||
|
||||
const content = fs.readFileSync(logPath, 'utf8');
|
||||
assert.match(content, /\[truncated older log content\]/);
|
||||
assert.match(content, /abcdefghij/);
|
||||
assert.ok(Buffer.byteLength(content) <= 48);
|
||||
} finally {
|
||||
fs.rmSync(logsDir, { recursive: true, force: true });
|
||||
}
|
||||
});
|
||||
123
src/shared/log-files.ts
Normal file
123
src/shared/log-files.ts
Normal file
@@ -0,0 +1,123 @@
|
||||
import fs from 'node:fs';
|
||||
import os from 'node:os';
|
||||
import path from 'node:path';
|
||||
|
||||
export type LogKind = 'app' | 'launcher' | 'mpv';
|
||||
|
||||
export const DEFAULT_LOG_RETENTION_DAYS = 7;
|
||||
export const DEFAULT_LOG_MAX_BYTES = 10 * 1024 * 1024;
|
||||
|
||||
const TRUNCATED_MARKER = '[truncated older log content]\n';
|
||||
const prunedDirectories = new Set<string>();
|
||||
|
||||
export function resolveLogBaseDir(options?: {
|
||||
platform?: NodeJS.Platform;
|
||||
homeDir?: string;
|
||||
appDataDir?: string;
|
||||
}): string {
|
||||
const platform = options?.platform ?? process.platform;
|
||||
const homeDir = options?.homeDir ?? os.homedir();
|
||||
return platform === 'win32'
|
||||
? path.join(options?.appDataDir?.trim() || path.join(homeDir, 'AppData', 'Roaming'), 'SubMiner')
|
||||
: path.join(homeDir, '.config', 'SubMiner');
|
||||
}
|
||||
|
||||
export function resolveDefaultLogFilePath(
|
||||
kind: LogKind = 'app',
|
||||
options?: {
|
||||
platform?: NodeJS.Platform;
|
||||
homeDir?: string;
|
||||
appDataDir?: string;
|
||||
now?: Date;
|
||||
},
|
||||
): string {
|
||||
const date = (options?.now ?? new Date()).toISOString().slice(0, 10);
|
||||
return path.join(resolveLogBaseDir(options), 'logs', `${kind}-${date}.log`);
|
||||
}
|
||||
|
||||
export function pruneLogFiles(
|
||||
logsDir: string,
|
||||
options?: {
|
||||
retentionDays?: number;
|
||||
now?: Date;
|
||||
},
|
||||
): void {
|
||||
const retentionDays = options?.retentionDays ?? DEFAULT_LOG_RETENTION_DAYS;
|
||||
if (!Number.isFinite(retentionDays) || retentionDays <= 0) return;
|
||||
|
||||
let entries: string[];
|
||||
try {
|
||||
entries = fs.readdirSync(logsDir);
|
||||
} catch {
|
||||
return;
|
||||
}
|
||||
|
||||
const cutoffMs = (options?.now ?? new Date()).getTime() - retentionDays * 24 * 60 * 60 * 1000;
|
||||
for (const entry of entries) {
|
||||
const candidate = path.join(logsDir, entry);
|
||||
let stats: fs.Stats;
|
||||
try {
|
||||
stats = fs.statSync(candidate);
|
||||
} catch {
|
||||
continue;
|
||||
}
|
||||
if (!stats.isFile() || !entry.endsWith('.log') || stats.mtimeMs >= cutoffMs) continue;
|
||||
try {
|
||||
fs.rmSync(candidate, { force: true });
|
||||
} catch {
|
||||
// ignore cleanup failures
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
function maybePruneLogDirectory(logPath: string, retentionDays: number): void {
|
||||
const logsDir = path.dirname(logPath);
|
||||
const key = `${logsDir}:${new Date().toISOString().slice(0, 10)}:${retentionDays}`;
|
||||
if (prunedDirectories.has(key)) return;
|
||||
pruneLogFiles(logsDir, { retentionDays });
|
||||
prunedDirectories.add(key);
|
||||
}
|
||||
|
||||
function trimLogFileToMaxBytes(logPath: string, maxBytes: number): void {
|
||||
if (!Number.isFinite(maxBytes) || maxBytes <= 0) return;
|
||||
|
||||
let stats: fs.Stats;
|
||||
try {
|
||||
stats = fs.statSync(logPath);
|
||||
} catch {
|
||||
return;
|
||||
}
|
||||
if (stats.size <= maxBytes) return;
|
||||
|
||||
try {
|
||||
const buffer = fs.readFileSync(logPath);
|
||||
const marker = Buffer.from(TRUNCATED_MARKER, 'utf8');
|
||||
const tailBudget = Math.max(0, maxBytes - marker.length);
|
||||
const tail =
|
||||
tailBudget > 0 ? buffer.subarray(Math.max(0, buffer.length - tailBudget)) : Buffer.alloc(0);
|
||||
fs.writeFileSync(logPath, Buffer.concat([marker, tail]));
|
||||
} catch {
|
||||
// ignore trim failures
|
||||
}
|
||||
}
|
||||
|
||||
export function appendLogLine(
|
||||
logPath: string,
|
||||
line: string,
|
||||
options?: {
|
||||
retentionDays?: number;
|
||||
maxBytes?: number;
|
||||
},
|
||||
): void {
|
||||
const retentionDays = options?.retentionDays ?? DEFAULT_LOG_RETENTION_DAYS;
|
||||
const maxBytes = options?.maxBytes ?? DEFAULT_LOG_MAX_BYTES;
|
||||
|
||||
try {
|
||||
fs.mkdirSync(path.dirname(logPath), { recursive: true });
|
||||
maybePruneLogDirectory(logPath, retentionDays);
|
||||
fs.appendFileSync(logPath, `${line}\n`, { encoding: 'utf8' });
|
||||
trimLogFileToMaxBytes(logPath, maxBytes);
|
||||
} catch {
|
||||
// never break runtime due to logging sink failures
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user