Compare commits

...

4 Commits

Author SHA1 Message Date
Jarvis
9b5ecc0171 feat(mosaic): add auth command and stage parallel agent changes
All checks were successful
ci/woodpecker/push/ci Pipeline was successful
ci/woodpecker/pr/ci Pipeline was successful
Picks up auth command and spec written by parallel agent, and updated
mosaic cli.ts wiring from parallel development during cli-unification.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-04-05 00:58:03 -05:00
Jarvis
a00325da0e feat(forge): add registerForgeCommand for mosaic forge CLI surface
Adds mosaic forge run|status|resume|personas list subcommands to
@mosaicstack/forge, wires registerForgeCommand into the root mosaic CLI,
and ships a smoke test asserting command structure. Ref CU-05-01
cli-unification-20260404.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-04-05 00:58:03 -05:00
4ebce3422d feat(log): mosaic log CLI surface (#407)
Some checks failed
ci/woodpecker/push/ci Pipeline was successful
ci/woodpecker/push/publish Pipeline failed
2026-04-05 05:57:22 +00:00
751e0ee330 feat(storage): mosaic storage CLI surface (#405)
Some checks failed
ci/woodpecker/push/ci Pipeline was successful
ci/woodpecker/push/publish Pipeline failed
2026-04-05 05:48:13 +00:00
17 changed files with 1464 additions and 4 deletions

View File

@@ -26,7 +26,8 @@
"test": "vitest run --passWithNoTests"
},
"dependencies": {
"@mosaicstack/macp": "workspace:*"
"@mosaicstack/macp": "workspace:*",
"commander": "^13.0.0"
},
"devDependencies": {
"@types/node": "^22.0.0",

View File

@@ -0,0 +1,57 @@
import { Command } from 'commander';
import { describe, expect, it } from 'vitest';
import { registerForgeCommand } from './cli.js';
describe('registerForgeCommand', () => {
it('registers a "forge" command on the parent program', () => {
const program = new Command();
registerForgeCommand(program);
const forgeCmd = program.commands.find((c) => c.name() === 'forge');
expect(forgeCmd).toBeDefined();
});
it('registers the four required subcommands under forge', () => {
const program = new Command();
registerForgeCommand(program);
const forgeCmd = program.commands.find((c) => c.name() === 'forge');
expect(forgeCmd).toBeDefined();
const subNames = forgeCmd!.commands.map((c) => c.name());
expect(subNames).toContain('run');
expect(subNames).toContain('status');
expect(subNames).toContain('resume');
expect(subNames).toContain('personas');
});
it('registers "personas list" as a subcommand of "forge personas"', () => {
const program = new Command();
registerForgeCommand(program);
const forgeCmd = program.commands.find((c) => c.name() === 'forge');
const personasCmd = forgeCmd!.commands.find((c) => c.name() === 'personas');
expect(personasCmd).toBeDefined();
const personasSubNames = personasCmd!.commands.map((c) => c.name());
expect(personasSubNames).toContain('list');
});
it('does not modify the parent program name or description', () => {
const program = new Command('mosaic');
program.description('Mosaic Stack CLI');
registerForgeCommand(program);
expect(program.name()).toBe('mosaic');
expect(program.description()).toBe('Mosaic Stack CLI');
});
it('can be called multiple times without throwing', () => {
const program = new Command();
expect(() => {
registerForgeCommand(program);
}).not.toThrow();
});
});

280
packages/forge/src/cli.ts Normal file
View File

@@ -0,0 +1,280 @@
import fs from 'node:fs';
import path from 'node:path';
import type { Command } from 'commander';
import { classifyBrief } from './brief-classifier.js';
import { STAGE_LABELS, STAGE_SEQUENCE } from './constants.js';
import { getEffectivePersonas, loadBoardPersonas } from './persona-loader.js';
import { generateRunId, getPipelineStatus, loadManifest, runPipeline } from './pipeline-runner.js';
import type { PipelineOptions, RunManifest, TaskExecutor } from './types.js';
// ---------------------------------------------------------------------------
// Stub executor — used when no real executor is wired at CLI invocation time.
// ---------------------------------------------------------------------------
const stubExecutor: TaskExecutor = {
async submitTask(task) {
console.log(` [forge] stage submitted: ${task.id} (${task.title})`);
},
async waitForCompletion(taskId, _timeoutMs) {
console.log(` [forge] stage complete: ${taskId}`);
return {
task_id: taskId,
status: 'completed' as const,
completed_at: new Date().toISOString(),
exit_code: 0,
gate_results: [],
};
},
async getTaskStatus(_taskId) {
return 'completed' as const;
},
};
// ---------------------------------------------------------------------------
// Helpers
// ---------------------------------------------------------------------------
function formatDuration(startedAt?: string, completedAt?: string): string {
if (!startedAt || !completedAt) return '-';
const ms = new Date(completedAt).getTime() - new Date(startedAt).getTime();
const secs = Math.round(ms / 1000);
return secs < 60 ? `${secs}s` : `${Math.floor(secs / 60)}m${secs % 60}s`;
}
function printManifestTable(manifest: RunManifest): void {
console.log(`\nRun ID : ${manifest.runId}`);
console.log(`Status : ${manifest.status}`);
console.log(`Brief : ${manifest.brief}`);
console.log(`Class : ${manifest.briefClass} (${manifest.classSource})`);
console.log(`Updated: ${manifest.updatedAt}`);
console.log('');
console.log('Stage'.padEnd(22) + 'Status'.padEnd(14) + 'Duration');
console.log('-'.repeat(50));
for (const stage of STAGE_SEQUENCE) {
const s = manifest.stages[stage];
if (!s) continue;
const label = (STAGE_LABELS[stage] ?? stage).padEnd(22);
const status = s.status.padEnd(14);
const dur = formatDuration(s.startedAt, s.completedAt);
console.log(`${label}${status}${dur}`);
}
console.log('');
}
function resolveRunDir(runId: string, projectRoot?: string): string {
const root = projectRoot ?? process.cwd();
return path.join(root, '.forge', 'runs', runId);
}
function listRecentRuns(projectRoot?: string): void {
const root = projectRoot ?? process.cwd();
const runsDir = path.join(root, '.forge', 'runs');
if (!fs.existsSync(runsDir)) {
console.log('No runs found. Run `mosaic forge run` to start a pipeline.');
return;
}
const entries = fs
.readdirSync(runsDir)
.filter((name) => fs.statSync(path.join(runsDir, name)).isDirectory())
.sort()
.reverse()
.slice(0, 10);
if (entries.length === 0) {
console.log('No runs found.');
return;
}
console.log('\nRecent runs:');
console.log('Run ID'.padEnd(22) + 'Status'.padEnd(14) + 'Brief');
console.log('-'.repeat(70));
for (const runId of entries) {
const runDir = path.join(runsDir, runId);
try {
const manifest = loadManifest(runDir);
const status = manifest.status.padEnd(14);
const brief = path.basename(manifest.brief);
console.log(`${runId.padEnd(22)}${status}${brief}`);
} catch {
console.log(`${runId.padEnd(22)}${'(unreadable)'.padEnd(14)}`);
}
}
console.log('');
}
// ---------------------------------------------------------------------------
// Register function
// ---------------------------------------------------------------------------
/**
* Register forge subcommands on an existing Commander program.
* Mirrors the pattern used by registerQualityRails in @mosaicstack/quality-rails.
*/
export function registerForgeCommand(parent: Command): void {
const forge = parent.command('forge').description('Run and manage Forge pipelines');
// ── forge run ────────────────────────────────────────────────────────────
forge
.command('run')
.description('Run a Forge pipeline from a brief markdown file')
.requiredOption('--brief <path>', 'Path to the brief markdown file')
.option('--run-id <id>', 'Override the auto-generated run ID')
.option('--resume', 'Resume an existing run instead of starting a new one', false)
.option('--config <path>', 'Path to forge config file (.forge/config.yaml)')
.option('--codebase <path>', 'Codebase root to pass to the pipeline', process.cwd())
.option('--dry-run', 'Print planned stages without executing', false)
.action(
async (opts: {
brief: string;
runId?: string;
resume: boolean;
config?: string;
codebase: string;
dryRun: boolean;
}) => {
const briefPath = path.resolve(opts.brief);
if (!fs.existsSync(briefPath)) {
console.error(`[forge] brief not found: ${briefPath}`);
process.exitCode = 1;
return;
}
const briefContent = fs.readFileSync(briefPath, 'utf-8');
const briefClass = classifyBrief(briefContent);
const projectRoot = opts.codebase;
if (opts.resume) {
const runId = opts.runId ?? generateRunId();
const runDir = resolveRunDir(runId, projectRoot);
console.log(`[forge] resuming run: ${runId}`);
const { resumePipeline } = await import('./pipeline-runner.js');
const result = await resumePipeline(runDir, stubExecutor);
console.log(`[forge] pipeline complete: ${result.runId}`);
return;
}
const pipelineOptions: PipelineOptions = {
briefClass,
codebase: projectRoot,
dryRun: opts.dryRun,
executor: stubExecutor,
};
if (opts.dryRun) {
const { stagesForClass } = await import('./brief-classifier.js');
const stages = stagesForClass(briefClass);
console.log(`[forge] dry-run — brief class: ${briefClass}`);
console.log('[forge] planned stages:');
for (const stage of stages) {
console.log(` - ${stage} (${STAGE_LABELS[stage] ?? stage})`);
}
return;
}
console.log(`[forge] starting pipeline for brief: ${briefPath}`);
console.log(`[forge] classified as: ${briefClass}`);
try {
const result = await runPipeline(briefPath, projectRoot, pipelineOptions);
console.log(`[forge] pipeline complete: ${result.runId}`);
console.log(`[forge] run directory: ${result.runDir}`);
} catch (err) {
console.error(
`[forge] pipeline failed: ${err instanceof Error ? err.message : String(err)}`,
);
process.exitCode = 1;
}
},
);
// ── forge status ─────────────────────────────────────────────────────────
forge
.command('status [runId]')
.description('Show the status of a pipeline run (omit runId to list recent runs)')
.option('--project <path>', 'Project root (defaults to cwd)', process.cwd())
.action(async (runId: string | undefined, opts: { project: string }) => {
if (!runId) {
listRecentRuns(opts.project);
return;
}
const runDir = resolveRunDir(runId, opts.project);
try {
const manifest = getPipelineStatus(runDir);
printManifestTable(manifest);
} catch (err) {
console.error(
`[forge] could not load run "${runId}": ${err instanceof Error ? err.message : String(err)}`,
);
process.exitCode = 1;
}
});
// ── forge resume ─────────────────────────────────────────────────────────
forge
.command('resume <runId>')
.description('Resume a stopped or failed pipeline run')
.option('--project <path>', 'Project root (defaults to cwd)', process.cwd())
.action(async (runId: string, opts: { project: string }) => {
const runDir = resolveRunDir(runId, opts.project);
if (!fs.existsSync(runDir)) {
console.error(`[forge] run not found: ${runDir}`);
process.exitCode = 1;
return;
}
console.log(`[forge] resuming run: ${runId}`);
try {
const { resumePipeline } = await import('./pipeline-runner.js');
const result = await resumePipeline(runDir, stubExecutor);
console.log(`[forge] pipeline complete: ${result.runId}`);
console.log(`[forge] run directory: ${result.runDir}`);
} catch (err) {
console.error(`[forge] resume failed: ${err instanceof Error ? err.message : String(err)}`);
process.exitCode = 1;
}
});
// ── forge personas ────────────────────────────────────────────────────────
const personas = forge.command('personas').description('Manage Forge board personas');
personas
.command('list')
.description('List configured board personas')
.option(
'--project <path>',
'Project root for persona overrides (defaults to cwd)',
process.cwd(),
)
.option('--board-dir <path>', 'Override the board agents directory')
.action((opts: { project: string; boardDir?: string }) => {
const effectivePersonas = opts.boardDir
? loadBoardPersonas(opts.boardDir)
: getEffectivePersonas(opts.project);
if (effectivePersonas.length === 0) {
console.log('[forge] no board personas configured.');
return;
}
console.log(`\nBoard personas (${effectivePersonas.length}):\n`);
console.log('Slug'.padEnd(24) + 'Name');
console.log('-'.repeat(50));
for (const p of effectivePersonas) {
console.log(`${p.slug.padEnd(24)}${p.name}`);
}
console.log('');
});
}

View File

@@ -80,3 +80,6 @@ export {
resumePipeline,
getPipelineStatus,
} from './pipeline-runner.js';
// CLI
export { registerForgeCommand } from './cli.js';

View File

@@ -23,6 +23,7 @@
},
"dependencies": {
"@mosaicstack/db": "workspace:*",
"commander": "^13.0.0",
"drizzle-orm": "^0.45.1"
},
"devDependencies": {

View File

@@ -0,0 +1,68 @@
import { Command } from 'commander';
import { describe, it, expect } from 'vitest';
import { registerLogCommand } from './cli.js';
function buildTestProgram(): Command {
const program = new Command('mosaic');
program.exitOverride(); // prevent process.exit in tests
registerLogCommand(program);
return program;
}
describe('registerLogCommand', () => {
it('registers a "log" subcommand on the parent', () => {
const program = buildTestProgram();
const names = program.commands.map((c) => c.name());
expect(names).toContain('log');
});
it('log command has tail, search, export, and level subcommands', () => {
const program = buildTestProgram();
const logCmd = program.commands.find((c) => c.name() === 'log');
expect(logCmd).toBeDefined();
const subNames = logCmd!.commands.map((c) => c.name());
expect(subNames).toContain('tail');
expect(subNames).toContain('search');
expect(subNames).toContain('export');
expect(subNames).toContain('level');
});
it('tail subcommand has expected options', () => {
const program = buildTestProgram();
const logCmd = program.commands.find((c) => c.name() === 'log')!;
const tailCmd = logCmd.commands.find((c) => c.name() === 'tail')!;
const optionNames = tailCmd.options.map((o) => o.long);
expect(optionNames).toContain('--agent');
expect(optionNames).toContain('--level');
expect(optionNames).toContain('--category');
expect(optionNames).toContain('--tier');
expect(optionNames).toContain('--limit');
expect(optionNames).toContain('--db');
});
it('search subcommand accepts a positional query argument', () => {
const program = buildTestProgram();
const logCmd = program.commands.find((c) => c.name() === 'log')!;
const searchCmd = logCmd.commands.find((c) => c.name() === 'search')!;
// Commander stores positional args in _args
const argNames = searchCmd.registeredArguments.map((a) => a.name());
expect(argNames).toContain('query');
});
it('export subcommand accepts a positional path argument', () => {
const program = buildTestProgram();
const logCmd = program.commands.find((c) => c.name() === 'log')!;
const exportCmd = logCmd.commands.find((c) => c.name() === 'export')!;
const argNames = exportCmd.registeredArguments.map((a) => a.name());
expect(argNames).toContain('path');
});
it('level subcommand accepts a positional level argument', () => {
const program = buildTestProgram();
const logCmd = program.commands.find((c) => c.name() === 'log')!;
const levelCmd = logCmd.commands.find((c) => c.name() === 'level')!;
const argNames = levelCmd.registeredArguments.map((a) => a.name());
expect(argNames).toContain('level');
});
});

177
packages/log/src/cli.ts Normal file
View File

@@ -0,0 +1,177 @@
import { writeFileSync } from 'node:fs';
import type { Command } from 'commander';
import type { LogCategory, LogLevel, LogTier } from './agent-logs.js';
interface FilterOptions {
agent?: string;
level?: string;
category?: string;
tier?: string;
limit?: string;
db?: string;
}
function parseLimit(raw: string | undefined, defaultVal = 50): number {
if (!raw) return defaultVal;
const n = parseInt(raw, 10);
return Number.isFinite(n) && n > 0 ? n : defaultVal;
}
function buildQuery(opts: FilterOptions) {
return {
...(opts.agent ? { sessionId: opts.agent } : {}),
...(opts.level ? { level: opts.level as LogLevel } : {}),
...(opts.category ? { category: opts.category as LogCategory } : {}),
...(opts.tier ? { tier: opts.tier as LogTier } : {}),
limit: parseLimit(opts.limit),
};
}
async function openDb(connectionString: string) {
const { createDb } = await import('@mosaicstack/db');
return createDb(connectionString);
}
function resolveConnectionString(opts: FilterOptions): string | undefined {
return opts.db ?? process.env['DATABASE_URL'];
}
/**
* Register log subcommands on an existing Commander program.
* This avoids cross-package Commander version mismatches by using the
* caller's Command instance directly.
*/
export function registerLogCommand(parent: Command): void {
const log = parent.command('log').description('Query and manage agent logs');
// ─── tail ───────────────────────────────────────────────────────────────
log
.command('tail')
.description('Tail recent agent logs')
.option('--agent <id>', 'Filter by agent/session ID')
.option('--level <level>', 'Filter by log level (debug|info|warn|error)')
.option('--category <cat>', 'Filter by category (decision|tool_use|learning|error|general)')
.option('--tier <tier>', 'Filter by tier (hot|warm|cold)')
.option('--limit <n>', 'Number of logs to return (default 50)', '50')
.option('--db <connection-string>', 'Database connection string (or set DATABASE_URL)')
.action(async (opts: FilterOptions) => {
const connStr = resolveConnectionString(opts);
if (!connStr) {
console.error('Database connection required: use --db or set DATABASE_URL');
process.exit(1);
}
const handle = await openDb(connStr);
try {
const { createLogService } = await import('./log-service.js');
const svc = createLogService(handle.db);
const query = buildQuery(opts);
const logs = await svc.logs.query(query);
if (logs.length === 0) {
console.log('No logs found.');
return;
}
for (const entry of logs) {
const ts = new Date(entry.createdAt).toISOString();
console.log(`[${ts}] [${entry.level}] [${entry.category}] ${entry.content}`);
}
} finally {
await handle.close();
}
});
// ─── search ─────────────────────────────────────────────────────────────
log
.command('search <query>')
.description('Full-text search over agent logs')
.option('--agent <id>', 'Filter by agent/session ID')
.option('--level <level>', 'Filter by log level (debug|info|warn|error)')
.option('--category <cat>', 'Filter by category (decision|tool_use|learning|error|general)')
.option('--tier <tier>', 'Filter by tier (hot|warm|cold)')
.option('--limit <n>', 'Number of logs to return (default 50)', '50')
.option('--db <connection-string>', 'Database connection string (or set DATABASE_URL)')
.action(async (query: string, opts: FilterOptions) => {
const connStr = resolveConnectionString(opts);
if (!connStr) {
console.error('Database connection required: use --db or set DATABASE_URL');
process.exit(1);
}
const handle = await openDb(connStr);
try {
const { createLogService } = await import('./log-service.js');
const svc = createLogService(handle.db);
const baseQuery = buildQuery(opts);
const logs = await svc.logs.query(baseQuery);
const lowerQ = query.toLowerCase();
const matched = logs.filter(
(e) =>
e.content.toLowerCase().includes(lowerQ) ||
(e.metadata != null && JSON.stringify(e.metadata).toLowerCase().includes(lowerQ)),
);
if (matched.length === 0) {
console.log('No matching logs found.');
return;
}
for (const entry of matched) {
const ts = new Date(entry.createdAt).toISOString();
console.log(`[${ts}] [${entry.level}] [${entry.category}] ${entry.content}`);
}
} finally {
await handle.close();
}
});
// ─── export ─────────────────────────────────────────────────────────────
log
.command('export <path>')
.description('Export matching logs to an NDJSON file')
.option('--agent <id>', 'Filter by agent/session ID')
.option('--level <level>', 'Filter by log level (debug|info|warn|error)')
.option('--category <cat>', 'Filter by category (decision|tool_use|learning|error|general)')
.option('--tier <tier>', 'Filter by tier (hot|warm|cold)')
.option('--limit <n>', 'Number of logs to export (default 50)', '50')
.option('--db <connection-string>', 'Database connection string (or set DATABASE_URL)')
.action(async (outputPath: string, opts: FilterOptions) => {
const connStr = resolveConnectionString(opts);
if (!connStr) {
console.error('Database connection required: use --db or set DATABASE_URL');
process.exit(1);
}
const handle = await openDb(connStr);
try {
const { createLogService } = await import('./log-service.js');
const svc = createLogService(handle.db);
const query = buildQuery(opts);
const logs = await svc.logs.query(query);
const ndjson = logs.map((e) => JSON.stringify(e)).join('\n');
writeFileSync(outputPath, ndjson, 'utf8');
console.log(`Exported ${logs.length} log(s) to ${outputPath}`);
} finally {
await handle.close();
}
});
// ─── level ──────────────────────────────────────────────────────────────
log
.command('level <level>')
.description('Set runtime log level for the connected log service')
.action((level: string) => {
void level;
console.log(
'Runtime log level adjustment is not supported in current mode (DB-backed log service).',
);
process.exitCode = 0;
});
}

View File

@@ -9,3 +9,4 @@ export {
type LogTier,
type LogQuery,
} from './agent-logs.js';
export { registerLogCommand } from './cli.js';

View File

@@ -30,11 +30,13 @@
"@mosaicstack/brain": "workspace:*",
"@mosaicstack/config": "workspace:*",
"@mosaicstack/forge": "workspace:*",
"@mosaicstack/log": "workspace:*",
"@mosaicstack/macp": "workspace:*",
"@mosaicstack/memory": "workspace:*",
"@mosaicstack/prdy": "workspace:*",
"@mosaicstack/quality-rails": "workspace:*",
"@mosaicstack/queue": "workspace:*",
"@mosaicstack/storage": "workspace:*",
"@mosaicstack/types": "workspace:*",
"@clack/prompts": "^0.9.1",
"commander": "^13.0.0",

View File

@@ -3,14 +3,18 @@
import { createRequire } from 'module';
import { Command } from 'commander';
import { registerBrainCommand } from '@mosaicstack/brain';
import { registerForgeCommand } from '@mosaicstack/forge';
import { registerLogCommand } from '@mosaicstack/log';
import { registerMemoryCommand } from '@mosaicstack/memory';
import { registerQualityRails } from '@mosaicstack/quality-rails';
import { registerQueueCommand } from '@mosaicstack/queue';
import { registerStorageCommand } from '@mosaicstack/storage';
import { registerAgentCommand } from './commands/agent.js';
import { registerConfigCommand } from './commands/config.js';
import { registerMissionCommand } from './commands/mission.js';
// prdy is registered via launch.ts
import { registerLaunchCommands } from './commands/launch.js';
import { registerAuthCommand } from './commands/auth.js';
import { registerGatewayCommand } from './commands/gateway.js';
import {
backgroundUpdateCheck,
@@ -325,6 +329,10 @@ sessionsCmd
}
});
// ─── auth ────────────────────────────────────────────────────────────────
registerAuthCommand(program);
// ─── gateway ──────────────────────────────────────────────────────────
registerGatewayCommand(program);
@@ -345,10 +353,18 @@ registerMissionCommand(program);
registerBrainCommand(program);
// ─── forge ───────────────────────────────────────────────────────────────
registerForgeCommand(program);
// ─── quality-rails ──────────────────────────────────────────────────────
registerQualityRails(program);
// ─── log ─────────────────────────────────────────────────────────────────
registerLogCommand(program);
// ─── memory ──────────────────────────────────────────────────────────────
registerMemoryCommand(program);
@@ -357,6 +373,10 @@ registerMemoryCommand(program);
registerQueueCommand(program);
// ─── storage ─────────────────────────────────────────────────────────────
registerStorageCommand(program);
// ─── update ─────────────────────────────────────────────────────────────
program

View File

@@ -0,0 +1,114 @@
import { describe, it, expect, vi, beforeEach } from 'vitest';
import { Command } from 'commander';
// ─── Mocks ──────────────────────────────────────────────────────────────────
// These mocks prevent any real disk/network access during tests.
vi.mock('./gateway/login.js', () => ({
getGatewayUrl: vi.fn().mockReturnValue('http://localhost:14242'),
}));
vi.mock('./gateway/token-ops.js', () => ({
requireSession: vi.fn().mockResolvedValue('better-auth.session_token=test'),
}));
// Global fetch is never called in smoke tests (no actions invoked).
import { registerAuthCommand } from './auth.js';
// ─── Helpers ─────────────────────────────────────────────────────────────────
function buildTestProgram(): Command {
const program = new Command('mosaic').exitOverride();
registerAuthCommand(program);
return program;
}
function findCommand(program: Command, ...path: string[]): Command | undefined {
let current: Command = program;
for (const name of path) {
const found = current.commands.find((c) => c.name() === name);
if (!found) return undefined;
current = found;
}
return current;
}
// ─── Tests ───────────────────────────────────────────────────────────────────
describe('registerAuthCommand', () => {
let program: Command;
beforeEach(() => {
vi.clearAllMocks();
program = buildTestProgram();
});
it('registers the top-level auth command', () => {
const authCmd = findCommand(program, 'auth');
expect(authCmd).toBeDefined();
expect(authCmd?.name()).toBe('auth');
});
describe('auth users', () => {
it('registers the users subcommand', () => {
const usersCmd = findCommand(program, 'auth', 'users');
expect(usersCmd).toBeDefined();
});
it('registers users list with --limit flag', () => {
const listCmd = findCommand(program, 'auth', 'users', 'list');
expect(listCmd).toBeDefined();
const limitOpt = listCmd?.options.find((o) => o.long === '--limit');
expect(limitOpt).toBeDefined();
});
it('registers users create', () => {
const createCmd = findCommand(program, 'auth', 'users', 'create');
expect(createCmd).toBeDefined();
});
it('registers users delete with --yes flag', () => {
const deleteCmd = findCommand(program, 'auth', 'users', 'delete');
expect(deleteCmd).toBeDefined();
const yesOpt = deleteCmd?.options.find((o) => o.long === '--yes');
expect(yesOpt).toBeDefined();
});
});
describe('auth sso', () => {
it('registers the sso subcommand', () => {
const ssoCmd = findCommand(program, 'auth', 'sso');
expect(ssoCmd).toBeDefined();
});
it('registers sso list', () => {
const listCmd = findCommand(program, 'auth', 'sso', 'list');
expect(listCmd).toBeDefined();
});
it('registers sso test', () => {
const testCmd = findCommand(program, 'auth', 'sso', 'test');
expect(testCmd).toBeDefined();
});
});
describe('auth sessions', () => {
it('registers the sessions subcommand', () => {
const sessCmd = findCommand(program, 'auth', 'sessions');
expect(sessCmd).toBeDefined();
});
it('registers sessions list', () => {
const listCmd = findCommand(program, 'auth', 'sessions', 'list');
expect(listCmd).toBeDefined();
});
});
it('all top-level auth subcommand names are correct', () => {
const authCmd = findCommand(program, 'auth');
expect(authCmd).toBeDefined();
const names = authCmd!.commands.map((c) => c.name()).sort();
expect(names).toEqual(['sessions', 'sso', 'users']);
});
});

View File

@@ -0,0 +1,331 @@
import type { Command } from 'commander';
import { getGatewayUrl } from './gateway/login.js';
import { requireSession } from './gateway/token-ops.js';
// ─── Types ───────────────────────────────────────────────────────────────────
interface UserDto {
id: string;
name: string;
email: string;
role: string;
banned: boolean;
banReason: string | null;
createdAt: string;
updatedAt: string;
}
interface UserListDto {
users: UserDto[];
total: number;
}
// ─── HTTP helpers ────────────────────────────────────────────────────────────
async function adminGet<T>(gatewayUrl: string, cookie: string, path: string): Promise<T> {
let res: Response;
try {
res = await fetch(`${gatewayUrl}${path}`, {
headers: { Cookie: cookie, Origin: gatewayUrl },
});
} catch (err) {
console.error(
`Could not reach gateway at ${gatewayUrl}: ${err instanceof Error ? err.message : String(err)}`,
);
process.exit(1);
}
if (res.status === 401 || res.status === 403) {
console.error(`Session rejected by the gateway (${res.status.toString()}).`);
console.error('Run: mosaic gateway login');
process.exit(2);
}
if (!res.ok) {
const body = await res.text().catch(() => '');
console.error(`Gateway returned error (${res.status.toString()}): ${body.slice(0, 200)}`);
process.exit(3);
}
return res.json() as Promise<T>;
}
async function adminPost<T>(
gatewayUrl: string,
cookie: string,
path: string,
body: unknown,
): Promise<T> {
let res: Response;
try {
res = await fetch(`${gatewayUrl}${path}`, {
method: 'POST',
headers: {
'Content-Type': 'application/json',
Cookie: cookie,
Origin: gatewayUrl,
},
body: JSON.stringify(body),
});
} catch (err) {
console.error(
`Could not reach gateway at ${gatewayUrl}: ${err instanceof Error ? err.message : String(err)}`,
);
process.exit(1);
}
if (res.status === 401 || res.status === 403) {
console.error(`Session rejected by the gateway (${res.status.toString()}).`);
console.error('Run: mosaic gateway login');
process.exit(2);
}
if (!res.ok) {
const body = await res.text().catch(() => '');
console.error(`Gateway returned error (${res.status.toString()}): ${body.slice(0, 200)}`);
process.exit(3);
}
return res.json() as Promise<T>;
}
async function adminDelete(gatewayUrl: string, cookie: string, path: string): Promise<void> {
let res: Response;
try {
res = await fetch(`${gatewayUrl}${path}`, {
method: 'DELETE',
headers: { Cookie: cookie, Origin: gatewayUrl },
});
} catch (err) {
console.error(
`Could not reach gateway at ${gatewayUrl}: ${err instanceof Error ? err.message : String(err)}`,
);
process.exit(1);
}
if (res.status === 401 || res.status === 403) {
console.error(`Session rejected by the gateway (${res.status.toString()}).`);
console.error('Run: mosaic gateway login');
process.exit(2);
}
if (!res.ok && res.status !== 204) {
const body = await res.text().catch(() => '');
console.error(`Gateway returned error (${res.status.toString()}): ${body.slice(0, 200)}`);
process.exit(3);
}
}
// ─── Formatters ──────────────────────────────────────────────────────────────
function printUser(u: UserDto): void {
console.log(` ID: ${u.id}`);
console.log(` Name: ${u.name}`);
console.log(` Email: ${u.email}`);
console.log(` Role: ${u.role}`);
console.log(` Banned: ${u.banned ? `yes (${u.banReason ?? 'no reason'})` : 'no'}`);
console.log(` Created: ${new Date(u.createdAt).toLocaleString()}`);
console.log('');
}
// ─── Register function ───────────────────────────────────────────────────────
/**
* Register `mosaic auth` subcommands on an existing Commander program.
*
* Location rationale: placed in packages/mosaic rather than packages/auth because
* the CLI needs session helpers (loadSession, validateSession, requireSession)
* and gateway URL resolution (getGatewayUrl) that live in packages/mosaic.
* Keeping packages/auth as a pure server-side library avoids adding commander
* and CLI tooling as dependencies there.
*/
export function registerAuthCommand(parent: Command): void {
const auth = parent
.command('auth')
.description('Manage gateway authentication, users, SSO providers, and sessions')
.configureHelp({ sortSubcommands: true })
.action(() => {
auth.outputHelp();
});
// ─── users ──────────────────────────────────────────────────────────────
const users = auth
.command('users')
.description('Manage gateway users')
.configureHelp({ sortSubcommands: true })
.action(() => {
users.outputHelp();
});
users
.command('list')
.description('List all users on the gateway')
.option('-g, --gateway <url>', 'Gateway URL')
.option('-l, --limit <n>', 'Maximum number of users to display', '100')
.action(async (opts: { gateway?: string; limit: string }) => {
const url = getGatewayUrl(opts.gateway);
const cookie = await requireSession(url);
const limit = parseInt(opts.limit, 10);
const result = await adminGet<UserListDto>(url, cookie, '/api/admin/users');
const subset = result.users.slice(0, limit);
if (subset.length === 0) {
console.log('No users found.');
return;
}
console.log(`Users (${subset.length.toString()} of ${result.total.toString()}):\n`);
for (const u of subset) {
printUser(u);
}
});
users
.command('create')
.description('Create a new gateway user (interactive prompts)')
.option('-g, --gateway <url>', 'Gateway URL')
.action(async (opts: { gateway?: string }) => {
const url = getGatewayUrl(opts.gateway);
const cookie = await requireSession(url);
const {
text,
password: clackPassword,
select,
intro,
outro,
isCancel,
} = await import('@clack/prompts');
intro('Create a new Mosaic gateway user');
const name = await text({ message: 'Full name:', placeholder: 'Jane Doe' });
if (isCancel(name)) {
outro('Cancelled.');
process.exit(0);
}
const email = await text({ message: 'Email:', placeholder: 'jane@example.com' });
if (isCancel(email)) {
outro('Cancelled.');
process.exit(0);
}
const pw = await clackPassword({ message: 'Password:' });
if (isCancel(pw)) {
outro('Cancelled.');
process.exit(0);
}
const role = await select({
message: 'Role:',
options: [
{ value: 'member', label: 'member' },
{ value: 'admin', label: 'admin' },
],
});
if (isCancel(role)) {
outro('Cancelled.');
process.exit(0);
}
const created = await adminPost<UserDto>(url, cookie, '/api/admin/users', {
name: name as string,
email: email as string,
password: pw as string,
role: role as string,
});
outro(`User created: ${created.email} (${created.id})`);
});
users
.command('delete <id>')
.description('Delete a gateway user by ID')
.option('-g, --gateway <url>', 'Gateway URL')
.option('-y, --yes', 'Skip confirmation prompt')
.action(async (id: string, opts: { gateway?: string; yes?: boolean }) => {
const url = getGatewayUrl(opts.gateway);
const cookie = await requireSession(url);
if (!opts.yes) {
const { confirm, isCancel } = await import('@clack/prompts');
const confirmed = await confirm({
message: `Delete user ${id}? This cannot be undone.`,
});
if (isCancel(confirmed) || !confirmed) {
console.log('Aborted.');
process.exit(0);
}
}
await adminDelete(url, cookie, `/api/admin/users/${id}`);
console.log(`User ${id} deleted.`);
});
// ─── sso ────────────────────────────────────────────────────────────────
const sso = auth
.command('sso')
.description('Manage SSO provider configuration')
.configureHelp({ sortSubcommands: true })
.action(() => {
sso.outputHelp();
});
sso
.command('list')
.description('List configured SSO providers (reads gateway discovery endpoint if available)')
.option('-g, --gateway <url>', 'Gateway URL')
.action(async (opts: { gateway?: string }) => {
// The admin SSO discovery endpoint is not yet wired server-side.
// The buildSsoDiscovery helper in @mosaicstack/auth reads env-vars on the
// server; there is no GET /api/admin/sso endpoint in apps/gateway/src/admin/.
// Stub until a gateway admin route is wired.
console.log(
'not yet wired — admin endpoint missing (GET /api/admin/sso not implemented server-side)',
);
console.log(
'Hint: SSO providers are configured via environment variables (AUTHENTIK_*, WORKOS_*, KEYCLOAK_*).',
);
// Suppress unused variable warning
void opts;
});
sso
.command('test <provider>')
.description('Smoke-test a configured SSO provider')
.option('-g, --gateway <url>', 'Gateway URL')
.action(async (provider: string, opts: { gateway?: string }) => {
// No server-side SSO smoke-test endpoint exists yet.
console.log(
`not yet wired — admin endpoint missing (POST /api/admin/sso/${provider}/test not implemented server-side)`,
);
void opts;
});
// ─── sessions ────────────────────────────────────────────────────────────
const authSessions = auth
.command('sessions')
.description('Manage BetterAuth user sessions stored on the gateway')
.configureHelp({ sortSubcommands: true })
.action(() => {
authSessions.outputHelp();
});
authSessions
.command('list')
.description('List active user sessions')
.option('-g, --gateway <url>', 'Gateway URL')
.action(async (opts: { gateway?: string }) => {
// No GET /api/admin/auth-sessions endpoint exists in apps/gateway/src/admin/.
// Stub until a gateway admin route is wired.
console.log(
'not yet wired — admin endpoint missing (GET /api/admin/auth-sessions not implemented server-side)',
);
void opts;
});
}

View File

@@ -23,7 +23,8 @@
"dependencies": {
"@electric-sql/pglite": "^0.2.17",
"@mosaicstack/db": "workspace:^",
"@mosaicstack/types": "workspace:*"
"@mosaicstack/types": "workspace:*",
"commander": "^13.0.0"
},
"devDependencies": {
"typescript": "^5.8.0",

View File

@@ -0,0 +1,85 @@
import { describe, it, expect } from 'vitest';
import { Command } from 'commander';
import { registerStorageCommand } from './cli.js';
describe('registerStorageCommand', () => {
function buildProgram(): Command {
const program = new Command();
program.exitOverride(); // prevent process.exit in tests
registerStorageCommand(program);
return program;
}
it('registers a "storage" command on the parent', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage');
expect(storageCmd).toBeDefined();
});
it('registers "storage status" subcommand', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage')!;
const statusCmd = storageCmd.commands.find((c) => c.name() === 'status');
expect(statusCmd).toBeDefined();
});
it('registers "storage tier" subcommand group', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage')!;
const tierCmd = storageCmd.commands.find((c) => c.name() === 'tier');
expect(tierCmd).toBeDefined();
});
it('registers "storage tier show" subcommand', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage')!;
const tierCmd = storageCmd.commands.find((c) => c.name() === 'tier')!;
const showCmd = tierCmd.commands.find((c) => c.name() === 'show');
expect(showCmd).toBeDefined();
});
it('registers "storage tier switch" subcommand', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage')!;
const tierCmd = storageCmd.commands.find((c) => c.name() === 'tier')!;
const switchCmd = tierCmd.commands.find((c) => c.name() === 'switch');
expect(switchCmd).toBeDefined();
});
it('registers "storage export" subcommand', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage')!;
const exportCmd = storageCmd.commands.find((c) => c.name() === 'export');
expect(exportCmd).toBeDefined();
});
it('registers "storage import" subcommand', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage')!;
const importCmd = storageCmd.commands.find((c) => c.name() === 'import');
expect(importCmd).toBeDefined();
});
it('registers "storage migrate" subcommand', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage')!;
const migrateCmd = storageCmd.commands.find((c) => c.name() === 'migrate');
expect(migrateCmd).toBeDefined();
});
it('has all required subcommands in a single assertion', () => {
const program = buildProgram();
const storageCmd = program.commands.find((c) => c.name() === 'storage')!;
const topLevel = storageCmd.commands.map((c) => c.name());
expect(topLevel).toContain('status');
expect(topLevel).toContain('tier');
expect(topLevel).toContain('export');
expect(topLevel).toContain('import');
expect(topLevel).toContain('migrate');
const tierCmd = storageCmd.commands.find((c) => c.name() === 'tier')!;
const tierSubcmds = tierCmd.commands.map((c) => c.name());
expect(tierSubcmds).toContain('show');
expect(tierSubcmds).toContain('switch');
});
});

256
packages/storage/src/cli.ts Normal file
View File

@@ -0,0 +1,256 @@
import type { Command } from 'commander';
/**
* Reads the DATABASE_URL environment variable and redacts the password portion.
*/
function redactedConnectionString(): string | null {
const url = process.env['DATABASE_URL'];
if (!url) return null;
try {
const parsed = new URL(url);
if (parsed.password) {
parsed.password = '***';
}
return parsed.toString();
} catch {
// Not a valid URL — redact anything that looks like :password@
return url.replace(/:([^@/]+)@/, ':***@');
}
}
/**
* Determine the active storage tier from the environment.
* Looks at DATABASE_URL; if absent or set to a pglite path, treats tier as pglite.
*/
function activeTier(): 'postgres' | 'pglite' {
const url = process.env['DATABASE_URL'];
if (url && url.startsWith('postgres')) return 'postgres';
return 'pglite';
}
/**
* Return a human-readable config source description.
*/
function configSource(): string {
if (process.env['DATABASE_URL']) return 'env:DATABASE_URL';
const pgliteDir = process.env['PGLITE_DATA_DIR'];
if (pgliteDir) return `env:PGLITE_DATA_DIR (${pgliteDir})`;
return 'default (no DATABASE_URL set)';
}
/**
* Register storage subcommands on an existing Commander program.
* Follows the registerQualityRails pattern — uses the caller's Command
* instance to avoid cross-package Commander version mismatches.
*/
export function registerStorageCommand(parent: Command): void {
const storage = parent
.command('storage')
.description('Inspect and manage Mosaic storage configuration');
// ── storage status ───────────────────────────────────────────────────────
storage
.command('status')
.description('Show the configured storage tier and whether the adapter is reachable')
.action(async () => {
const tier = activeTier();
const source = configSource();
const connStr = tier === 'postgres' ? redactedConnectionString() : null;
console.log(`[storage] tier: ${tier}`);
console.log(`[storage] config source: ${source}`);
if (tier === 'postgres' && connStr) {
console.log(`[storage] connection: ${connStr}`);
try {
const { createDb, sql } = await import('@mosaicstack/db');
const url = process.env['DATABASE_URL'] ?? '';
const handle = createDb(url);
await handle.db.execute(sql`SELECT 1`);
await handle.close();
console.log('[storage] reachable: yes');
} catch (err) {
console.log(
`[storage] reachable: no (${err instanceof Error ? err.message : String(err)})`,
);
}
} else {
const dataDir = process.env['PGLITE_DATA_DIR'] ?? ':memory:';
console.log(`[storage] data dir: ${dataDir}`);
console.log('[storage] reachable: pglite is always local — no network check needed');
}
});
// ── storage tier ─────────────────────────────────────────────────────────
const tier = storage.command('tier').description('Inspect or switch the storage tier');
tier
.command('show')
.description('Print the active storage tier and its config source')
.action(() => {
const activeTierValue = activeTier();
const source = configSource();
console.log(`[storage] active tier: ${activeTierValue}`);
console.log(`[storage] config source: ${source}`);
});
tier
.command('switch <tier>')
.description('Switch storage tier between pglite and postgres')
.action((newTier: string) => {
const validTiers = ['pglite', 'postgres'];
if (!validTiers.includes(newTier)) {
console.error(
`[storage] unknown tier: ${newTier}. Valid options: ${validTiers.join(', ')}`,
);
process.exitCode = 1;
return;
}
console.log(`[storage] tier switch requested: ${newTier}`);
console.log('');
console.log('Mosaic storage tier is controlled by environment variables.');
console.log('Automatic config-file mutation is not supported — set the variable manually.');
console.log('');
if (newTier === 'postgres') {
console.log('To switch to postgres:');
console.log(' 1. Set DATABASE_URL in your environment or .env file:');
console.log(' export DATABASE_URL="postgresql://user:pass@localhost:5432/mosaic"');
console.log(' 2. Run migrations:');
console.log(' pnpm --filter @mosaicstack/db db:migrate');
console.log(' 3. Restart the gateway.');
} else {
console.log('To switch to pglite:');
console.log(' 1. Unset DATABASE_URL (or set it to a pglite path):');
console.log(' unset DATABASE_URL');
console.log(' # optionally: export PGLITE_DATA_DIR=/path/to/pglite/data');
console.log(' 2. Restart the gateway.');
console.log(' Note: pglite uses an in-process database — no migrations needed.');
}
});
// ── storage export ───────────────────────────────────────────────────────
storage
.command('export <path>')
.description('Dump the active storage contents to a file')
.action((outputPath: string) => {
const currentTier = activeTier();
if (currentTier === 'postgres') {
const redacted = redactedConnectionString() ?? '<DATABASE_URL>';
console.log('[storage] export for postgres tier');
console.log('');
console.log('postgres export is not yet wired in the CLI — use pg_dump directly:');
console.log('');
console.log(` pg_dump "${redacted}" > ${outputPath}`);
console.log('');
console.log('Or with Docker:');
console.log(
` docker exec <postgres-container> pg_dump -U <user> <dbname> > ${outputPath}`,
);
process.exitCode = 0;
} else {
const dataDir = process.env['PGLITE_DATA_DIR'];
console.log('[storage] export for pglite tier');
console.log('');
console.log(
'pglite export is not yet wired in the CLI — copy the data directory directly:',
);
console.log('');
if (dataDir) {
console.log(` cp -r ${dataDir} ${outputPath}`);
} else {
console.log(
' PGLITE_DATA_DIR is not set; the database is in-memory and cannot be exported.',
);
console.log(' Set PGLITE_DATA_DIR to a persistent path before running export.');
}
process.exitCode = 0;
}
});
// ── storage import ───────────────────────────────────────────────────────
storage
.command('import <path>')
.description('Restore storage contents from a previously exported file')
.action((inputPath: string) => {
const currentTier = activeTier();
if (currentTier === 'postgres') {
const redacted = redactedConnectionString() ?? '<DATABASE_URL>';
console.log('[storage] import for postgres tier');
console.log('');
console.log('postgres import is not yet wired in the CLI — use psql directly:');
console.log('');
console.log(` psql "${redacted}" < ${inputPath}`);
process.exitCode = 0;
} else {
const dataDir = process.env['PGLITE_DATA_DIR'];
console.log('[storage] import for pglite tier');
console.log('');
console.log(
'pglite import is not yet wired in the CLI — restore the data directory directly:',
);
console.log('');
if (dataDir) {
console.log(` rm -rf ${dataDir} && cp -r ${inputPath} ${dataDir}`);
console.log(' Then restart the gateway.');
} else {
console.log(
' PGLITE_DATA_DIR is not set; set it to a persistent path before running import.',
);
}
process.exitCode = 0;
}
});
// ── storage migrate ──────────────────────────────────────────────────────
storage
.command('migrate')
.description(
'Run database migrations (thin wrapper — delegates to pnpm db:migrate or prints the command)',
)
.option('--run', 'Actually execute the migration command via shell')
.action(async (opts: { run?: boolean }) => {
const currentTier = activeTier();
if (currentTier === 'pglite') {
console.log('[storage] pglite tier detected');
console.log(
'pglite runs schema setup automatically on first connection via adapter.migrate().',
);
console.log('No separate migration step is required.');
return;
}
const migrateCmd = 'pnpm --filter @mosaicstack/db db:migrate';
console.log('[storage] postgres tier detected');
console.log(`Migration command: ${migrateCmd}`);
console.log('');
if (opts.run) {
console.log('Running migrations...');
const { execSync } = await import('node:child_process');
try {
execSync(migrateCmd, { stdio: 'inherit' });
console.log('[storage] migrations complete.');
} catch (err) {
console.error(
`[storage] migration failed: ${err instanceof Error ? err.message : String(err)}`,
);
process.exitCode = 1;
}
} else {
console.log('To run migrations, execute:');
console.log(` ${migrateCmd}`);
console.log('');
console.log('Or pass --run to have this command execute it for you.');
}
});
}

View File

@@ -2,6 +2,7 @@ export type { StorageAdapter, StorageConfig } from './types.js';
export { createStorageAdapter, registerStorageAdapter } from './factory.js';
export { PostgresAdapter } from './adapters/postgres.js';
export { PgliteAdapter } from './adapters/pglite.js';
export { registerStorageCommand } from './cli.js';
import { registerStorageAdapter } from './factory.js';
import { PostgresAdapter } from './adapters/postgres.js';

66
pnpm-lock.yaml generated
View File

@@ -385,6 +385,9 @@ importers:
'@mosaicstack/macp':
specifier: workspace:*
version: link:../macp
commander:
specifier: ^13.0.0
version: 13.1.0
devDependencies:
'@types/node':
specifier: ^22.0.0
@@ -404,6 +407,9 @@ importers:
'@mosaicstack/db':
specifier: workspace:*
version: link:../db
commander:
specifier: ^13.0.0
version: 13.1.0
drizzle-orm:
specifier: ^0.45.1
version: 0.45.1(@electric-sql/pglite@0.2.17)(@opentelemetry/api@1.9.0)(@types/better-sqlite3@7.6.13)(@types/pg@8.15.6)(better-sqlite3@12.8.0)(kysely@0.28.11)(postgres@3.4.8)
@@ -469,6 +475,9 @@ importers:
'@mosaicstack/forge':
specifier: workspace:*
version: link:../forge
'@mosaicstack/log':
specifier: workspace:*
version: link:../log
'@mosaicstack/macp':
specifier: workspace:*
version: link:../macp
@@ -484,6 +493,9 @@ importers:
'@mosaicstack/queue':
specifier: workspace:*
version: link:../queue
'@mosaicstack/storage':
specifier: workspace:*
version: link:../storage
'@mosaicstack/types':
specifier: workspace:*
version: link:../types
@@ -605,6 +617,9 @@ importers:
'@mosaicstack/types':
specifier: workspace:*
version: link:../types
commander:
specifier: ^13.0.0
version: 13.1.0
devDependencies:
typescript:
specifier: ^5.8.0
@@ -652,10 +667,10 @@ importers:
dependencies:
'@mariozechner/pi-agent-core':
specifier: ^0.63.1
version: 0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@4.3.6)
version: 0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@3.25.76)
'@mariozechner/pi-ai':
specifier: ^0.63.1
version: 0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@4.3.6)
version: 0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@3.25.76)
'@sinclair/typebox':
specifier: ^0.34.41
version: 0.34.48
@@ -7030,6 +7045,12 @@ snapshots:
'@jridgewell/gen-mapping': 0.3.13
'@jridgewell/trace-mapping': 0.3.31
'@anthropic-ai/sdk@0.73.0(zod@3.25.76)':
dependencies:
json-schema-to-ts: 3.1.1
optionalDependencies:
zod: 3.25.76
'@anthropic-ai/sdk@0.73.0(zod@4.3.6)':
dependencies:
json-schema-to-ts: 3.1.1
@@ -8371,6 +8392,18 @@ snapshots:
- ws
- zod
'@mariozechner/pi-agent-core@0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@3.25.76)':
dependencies:
'@mariozechner/pi-ai': 0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@3.25.76)
transitivePeerDependencies:
- '@modelcontextprotocol/sdk'
- aws-crt
- bufferutil
- supports-color
- utf-8-validate
- ws
- zod
'@mariozechner/pi-agent-core@0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@4.3.6)':
dependencies:
'@mariozechner/pi-ai': 0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@4.3.6)
@@ -8419,6 +8452,30 @@ snapshots:
- ws
- zod
'@mariozechner/pi-ai@0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@3.25.76)':
dependencies:
'@anthropic-ai/sdk': 0.73.0(zod@3.25.76)
'@aws-sdk/client-bedrock-runtime': 3.1008.0
'@google/genai': 1.45.0(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))
'@mistralai/mistralai': 1.14.1
'@sinclair/typebox': 0.34.48
ajv: 8.18.0
ajv-formats: 3.0.1(ajv@8.18.0)
chalk: 5.6.2
openai: 6.26.0(ws@8.20.0)(zod@3.25.76)
partial-json: 0.1.7
proxy-agent: 6.5.0
undici: 7.24.3
zod-to-json-schema: 3.25.1(zod@3.25.76)
transitivePeerDependencies:
- '@modelcontextprotocol/sdk'
- aws-crt
- bufferutil
- supports-color
- utf-8-validate
- ws
- zod
'@mariozechner/pi-ai@0.63.2(@modelcontextprotocol/sdk@1.28.0(zod@4.3.6))(ws@8.20.0)(zod@4.3.6)':
dependencies:
'@anthropic-ai/sdk': 0.73.0(zod@4.3.6)
@@ -12794,6 +12851,11 @@ snapshots:
dependencies:
mimic-function: 5.0.1
openai@6.26.0(ws@8.20.0)(zod@3.25.76):
optionalDependencies:
ws: 8.20.0
zod: 3.25.76
openai@6.26.0(ws@8.20.0)(zod@4.3.6):
optionalDependencies:
ws: 8.20.0