Compare commits
5 Commits
fix/gatewa
...
fix/metapa
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
2abcfee211 | ||
|
|
799f9e76bf | ||
|
|
49abf3bb4c | ||
| c0d0fd44b7 | |||
| 30c0fb1308 |
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@mosaic/gateway",
|
"name": "@mosaic/gateway",
|
||||||
"version": "0.0.2",
|
"version": "0.1.0",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
"url": "https://git.mosaicstack.dev/mosaic/mosaic-stack.git",
|
"url": "https://git.mosaicstack.dev/mosaic/mosaic-stack.git",
|
||||||
@@ -28,8 +28,8 @@
|
|||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@anthropic-ai/sdk": "^0.80.0",
|
"@anthropic-ai/sdk": "^0.80.0",
|
||||||
"@fastify/helmet": "^13.0.2",
|
"@fastify/helmet": "^13.0.2",
|
||||||
"@mariozechner/pi-ai": "~0.57.1",
|
"@mariozechner/pi-ai": "^0.65.0",
|
||||||
"@mariozechner/pi-coding-agent": "~0.57.1",
|
"@mariozechner/pi-coding-agent": "^0.65.0",
|
||||||
"@modelcontextprotocol/sdk": "^1.27.1",
|
"@modelcontextprotocol/sdk": "^1.27.1",
|
||||||
"@mosaic/auth": "workspace:^",
|
"@mosaic/auth": "workspace:^",
|
||||||
"@mosaic/brain": "workspace:^",
|
"@mosaic/brain": "workspace:^",
|
||||||
@@ -49,7 +49,7 @@
|
|||||||
"@nestjs/platform-socket.io": "^11.0.0",
|
"@nestjs/platform-socket.io": "^11.0.0",
|
||||||
"@nestjs/throttler": "^6.5.0",
|
"@nestjs/throttler": "^6.5.0",
|
||||||
"@nestjs/websockets": "^11.0.0",
|
"@nestjs/websockets": "^11.0.0",
|
||||||
"@opentelemetry/auto-instrumentations-node": "^0.71.0",
|
"@opentelemetry/auto-instrumentations-node": "^0.72.0",
|
||||||
"@opentelemetry/exporter-metrics-otlp-http": "^0.213.0",
|
"@opentelemetry/exporter-metrics-otlp-http": "^0.213.0",
|
||||||
"@opentelemetry/exporter-trace-otlp-http": "^0.213.0",
|
"@opentelemetry/exporter-trace-otlp-http": "^0.213.0",
|
||||||
"@opentelemetry/resources": "^2.6.0",
|
"@opentelemetry/resources": "^2.6.0",
|
||||||
|
|||||||
@@ -62,7 +62,7 @@ function restoreEnv(saved: Map<EnvKey, string | undefined>): void {
|
|||||||
}
|
}
|
||||||
|
|
||||||
function makeRegistry(): ModelRegistry {
|
function makeRegistry(): ModelRegistry {
|
||||||
return new ModelRegistry(AuthStorage.inMemory());
|
return ModelRegistry.inMemory(AuthStorage.inMemory());
|
||||||
}
|
}
|
||||||
|
|
||||||
// ---------------------------------------------------------------------------
|
// ---------------------------------------------------------------------------
|
||||||
|
|||||||
@@ -67,7 +67,7 @@ export class ProviderService implements OnModuleInit, OnModuleDestroy {
|
|||||||
|
|
||||||
async onModuleInit(): Promise<void> {
|
async onModuleInit(): Promise<void> {
|
||||||
const authStorage = AuthStorage.inMemory();
|
const authStorage = AuthStorage.inMemory();
|
||||||
this.registry = new ModelRegistry(authStorage);
|
this.registry = ModelRegistry.inMemory(authStorage);
|
||||||
|
|
||||||
// Build the default set of adapters that rely on the registry
|
// Build the default set of adapters that rely on the registry
|
||||||
this.adapters = [
|
this.adapters = [
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"tier": "local",
|
"tier": "local",
|
||||||
"storage": { "type": "sqlite", "path": ".mosaic/data.db" },
|
"storage": { "type": "pglite", "dataDir": ".mosaic/storage-pglite" },
|
||||||
"queue": { "type": "local", "dataDir": ".mosaic/queue" },
|
"queue": { "type": "local", "dataDir": ".mosaic/queue" },
|
||||||
"memory": { "type": "keyword" }
|
"memory": { "type": "keyword" }
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -23,10 +23,5 @@
|
|||||||
"turbo": "^2.0.0",
|
"turbo": "^2.0.0",
|
||||||
"typescript": "^5.8.0",
|
"typescript": "^5.8.0",
|
||||||
"vitest": "^2.0.0"
|
"vitest": "^2.0.0"
|
||||||
},
|
|
||||||
"pnpm": {
|
|
||||||
"onlyBuiltDependencies": [
|
|
||||||
"better-sqlite3"
|
|
||||||
]
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -121,7 +121,7 @@ async function doInstall(rl: ReturnType<typeof createInterface>, opts: InstallOp
|
|||||||
tier === 'local'
|
tier === 'local'
|
||||||
? {
|
? {
|
||||||
tier: 'local',
|
tier: 'local',
|
||||||
storage: { type: 'sqlite', path: join(GATEWAY_HOME, 'data.db') },
|
storage: { type: 'pglite', dataDir: join(GATEWAY_HOME, 'storage-pglite') },
|
||||||
queue: { type: 'local', dataDir: join(GATEWAY_HOME, 'queue') },
|
queue: { type: 'local', dataDir: join(GATEWAY_HOME, 'queue') },
|
||||||
memory: { type: 'keyword' },
|
memory: { type: 'keyword' },
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -27,6 +27,7 @@
|
|||||||
"@mosaic/storage": "workspace:^"
|
"@mosaic/storage": "workspace:^"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
|
"eslint": "^9.0.0",
|
||||||
"typescript": "^5.8.0",
|
"typescript": "^5.8.0",
|
||||||
"vitest": "^2.0.0"
|
"vitest": "^2.0.0"
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -26,7 +26,7 @@ export interface MosaicConfig {
|
|||||||
|
|
||||||
export const DEFAULT_LOCAL_CONFIG: MosaicConfig = {
|
export const DEFAULT_LOCAL_CONFIG: MosaicConfig = {
|
||||||
tier: 'local',
|
tier: 'local',
|
||||||
storage: { type: 'sqlite', path: '.mosaic/data.db' },
|
storage: { type: 'pglite', dataDir: '.mosaic/storage-pglite' },
|
||||||
queue: { type: 'local', dataDir: '.mosaic/queue' },
|
queue: { type: 'local', dataDir: '.mosaic/queue' },
|
||||||
memory: { type: 'keyword' },
|
memory: { type: 'keyword' },
|
||||||
};
|
};
|
||||||
@@ -43,7 +43,7 @@ export const DEFAULT_TEAM_CONFIG: MosaicConfig = {
|
|||||||
/* ------------------------------------------------------------------ */
|
/* ------------------------------------------------------------------ */
|
||||||
|
|
||||||
const VALID_TIERS = new Set<string>(['local', 'team']);
|
const VALID_TIERS = new Set<string>(['local', 'team']);
|
||||||
const VALID_STORAGE_TYPES = new Set<string>(['postgres', 'sqlite', 'files']);
|
const VALID_STORAGE_TYPES = new Set<string>(['postgres', 'pglite', 'files']);
|
||||||
const VALID_QUEUE_TYPES = new Set<string>(['bullmq', 'local']);
|
const VALID_QUEUE_TYPES = new Set<string>(['bullmq', 'local']);
|
||||||
const VALID_MEMORY_TYPES = new Set<string>(['pgvector', 'sqlite-vec', 'keyword']);
|
const VALID_MEMORY_TYPES = new Set<string>(['pgvector', 'sqlite-vec', 'keyword']);
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@mosaic/mosaic",
|
"name": "@mosaic/mosaic",
|
||||||
"version": "0.0.15",
|
"version": "0.0.17",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
"url": "https://git.mosaicstack.dev/mosaic/mosaic-stack.git",
|
"url": "https://git.mosaicstack.dev/mosaic/mosaic-stack.git",
|
||||||
@@ -20,12 +20,14 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"scripts": {
|
"scripts": {
|
||||||
|
"postinstall": "echo '\nMosaic Stack installed. Run `mosaic-wizard` to bootstrap your config, or `mosaic --help` for CLI commands.\n'",
|
||||||
"build": "tsc",
|
"build": "tsc",
|
||||||
"lint": "eslint src",
|
"lint": "eslint src",
|
||||||
"typecheck": "tsc --noEmit",
|
"typecheck": "tsc --noEmit",
|
||||||
"test": "vitest run --passWithNoTests"
|
"test": "vitest run --passWithNoTests"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
|
"@mosaic/cli": "workspace:*",
|
||||||
"@mosaic/forge": "workspace:*",
|
"@mosaic/forge": "workspace:*",
|
||||||
"@mosaic/macp": "workspace:*",
|
"@mosaic/macp": "workspace:*",
|
||||||
"@mosaic/prdy": "workspace:*",
|
"@mosaic/prdy": "workspace:*",
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@mosaic/storage",
|
"name": "@mosaic/storage",
|
||||||
"version": "0.0.2",
|
"version": "0.0.3",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
"url": "https://git.mosaicstack.dev/mosaic/mosaic-stack.git",
|
"url": "https://git.mosaicstack.dev/mosaic/mosaic-stack.git",
|
||||||
@@ -21,12 +21,11 @@
|
|||||||
"test": "vitest run --passWithNoTests"
|
"test": "vitest run --passWithNoTests"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
|
"@electric-sql/pglite": "^0.2.17",
|
||||||
"@mosaic/db": "workspace:^",
|
"@mosaic/db": "workspace:^",
|
||||||
"@mosaic/types": "workspace:*",
|
"@mosaic/types": "workspace:*"
|
||||||
"better-sqlite3": "^12.8.0"
|
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@types/better-sqlite3": "^7.6.13",
|
|
||||||
"typescript": "^5.8.0",
|
"typescript": "^5.8.0",
|
||||||
"vitest": "^2.0.0"
|
"vitest": "^2.0.0"
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -1,11 +1,12 @@
|
|||||||
import { describe, it, expect, beforeEach, afterEach } from 'vitest';
|
import { describe, it, expect, beforeEach, afterEach } from 'vitest';
|
||||||
import { SqliteAdapter } from './sqlite.js';
|
import { PgliteAdapter } from './pglite.js';
|
||||||
|
|
||||||
describe('SqliteAdapter', () => {
|
describe('PgliteAdapter', () => {
|
||||||
let adapter: SqliteAdapter;
|
let adapter: PgliteAdapter;
|
||||||
|
|
||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
adapter = new SqliteAdapter({ type: 'sqlite', path: ':memory:' });
|
// In-memory PGlite instance — no dataDir = memory mode
|
||||||
|
adapter = new PgliteAdapter({ type: 'pglite' });
|
||||||
await adapter.migrate();
|
await adapter.migrate();
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -80,7 +81,7 @@ describe('SqliteAdapter', () => {
|
|||||||
|
|
||||||
it('supports limit and offset', async () => {
|
it('supports limit and offset', async () => {
|
||||||
for (let i = 0; i < 5; i++) {
|
for (let i = 0; i < 5; i++) {
|
||||||
await adapter.create('users', { name: `User${i}`, idx: i });
|
await adapter.create('users', { name: `User${i.toString()}`, idx: i });
|
||||||
}
|
}
|
||||||
|
|
||||||
const page = await adapter.find('users', undefined, {
|
const page = await adapter.find('users', undefined, {
|
||||||
290
packages/storage/src/adapters/pglite.ts
Normal file
290
packages/storage/src/adapters/pglite.ts
Normal file
@@ -0,0 +1,290 @@
|
|||||||
|
import { PGlite } from '@electric-sql/pglite';
|
||||||
|
import { randomUUID } from 'node:crypto';
|
||||||
|
import type { StorageAdapter, StorageConfig } from '../types.js';
|
||||||
|
|
||||||
|
/* eslint-disable @typescript-eslint/no-explicit-any */
|
||||||
|
|
||||||
|
const COLLECTIONS = [
|
||||||
|
'users',
|
||||||
|
'sessions',
|
||||||
|
'accounts',
|
||||||
|
'projects',
|
||||||
|
'missions',
|
||||||
|
'tasks',
|
||||||
|
'agents',
|
||||||
|
'conversations',
|
||||||
|
'messages',
|
||||||
|
'preferences',
|
||||||
|
'insights',
|
||||||
|
'skills',
|
||||||
|
'events',
|
||||||
|
'routing_rules',
|
||||||
|
'provider_credentials',
|
||||||
|
'agent_logs',
|
||||||
|
'teams',
|
||||||
|
'team_members',
|
||||||
|
'mission_tasks',
|
||||||
|
'tickets',
|
||||||
|
'summarization_jobs',
|
||||||
|
'appreciations',
|
||||||
|
'verifications',
|
||||||
|
] as const;
|
||||||
|
|
||||||
|
function buildFilterClause(filter?: Record<string, unknown>): {
|
||||||
|
clause: string;
|
||||||
|
params: unknown[];
|
||||||
|
} {
|
||||||
|
if (!filter || Object.keys(filter).length === 0) return { clause: '', params: [] };
|
||||||
|
const conditions: string[] = [];
|
||||||
|
const params: unknown[] = [];
|
||||||
|
let paramIdx = 1;
|
||||||
|
for (const [key, value] of Object.entries(filter)) {
|
||||||
|
if (key === 'id') {
|
||||||
|
conditions.push(`id = $${paramIdx.toString()}`);
|
||||||
|
params.push(value);
|
||||||
|
paramIdx++;
|
||||||
|
} else {
|
||||||
|
conditions.push(`data->>'${key}' = $${paramIdx.toString()}`);
|
||||||
|
params.push(typeof value === 'object' ? JSON.stringify(value) : value);
|
||||||
|
paramIdx++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return { clause: ` WHERE ${conditions.join(' AND ')}`, params };
|
||||||
|
}
|
||||||
|
|
||||||
|
type PgClient = PGlite | { query: PGlite['query'] };
|
||||||
|
|
||||||
|
async function pgCreate<T extends Record<string, unknown>>(
|
||||||
|
pg: PgClient,
|
||||||
|
collection: string,
|
||||||
|
data: T,
|
||||||
|
): Promise<T & { id: string }> {
|
||||||
|
const id = (data as any).id ?? randomUUID();
|
||||||
|
const rest = Object.fromEntries(Object.entries(data).filter(([k]) => k !== 'id'));
|
||||||
|
await pg.query(`INSERT INTO ${collection} (id, data) VALUES ($1, $2::jsonb)`, [
|
||||||
|
id,
|
||||||
|
JSON.stringify(rest),
|
||||||
|
]);
|
||||||
|
return { ...data, id } as T & { id: string };
|
||||||
|
}
|
||||||
|
|
||||||
|
async function pgRead<T extends Record<string, unknown>>(
|
||||||
|
pg: PgClient,
|
||||||
|
collection: string,
|
||||||
|
id: string,
|
||||||
|
): Promise<T | null> {
|
||||||
|
const result = await pg.query<{ id: string; data: Record<string, unknown> }>(
|
||||||
|
`SELECT id, data FROM ${collection} WHERE id = $1`,
|
||||||
|
[id],
|
||||||
|
);
|
||||||
|
const row = result.rows[0];
|
||||||
|
if (!row) return null;
|
||||||
|
return { id: row.id, ...(row.data as object) } as unknown as T;
|
||||||
|
}
|
||||||
|
|
||||||
|
async function pgUpdate(
|
||||||
|
pg: PgClient,
|
||||||
|
collection: string,
|
||||||
|
id: string,
|
||||||
|
data: Record<string, unknown>,
|
||||||
|
): Promise<boolean> {
|
||||||
|
const existing = await pg.query<{ data: Record<string, unknown> }>(
|
||||||
|
`SELECT data FROM ${collection} WHERE id = $1`,
|
||||||
|
[id],
|
||||||
|
);
|
||||||
|
const row = existing.rows[0];
|
||||||
|
if (!row) return false;
|
||||||
|
const merged = { ...(row.data as object), ...data };
|
||||||
|
const result = await pg.query(
|
||||||
|
`UPDATE ${collection} SET data = $1::jsonb, updated_at = now() WHERE id = $2`,
|
||||||
|
[JSON.stringify(merged), id],
|
||||||
|
);
|
||||||
|
return (result.affectedRows ?? 0) > 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
async function pgDelete(pg: PgClient, collection: string, id: string): Promise<boolean> {
|
||||||
|
const result = await pg.query(`DELETE FROM ${collection} WHERE id = $1`, [id]);
|
||||||
|
return (result.affectedRows ?? 0) > 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
async function pgFind<T extends Record<string, unknown>>(
|
||||||
|
pg: PgClient,
|
||||||
|
collection: string,
|
||||||
|
filter?: Record<string, unknown>,
|
||||||
|
opts?: { limit?: number; offset?: number; orderBy?: string; order?: 'asc' | 'desc' },
|
||||||
|
): Promise<T[]> {
|
||||||
|
const { clause, params } = buildFilterClause(filter);
|
||||||
|
let paramIdx = params.length + 1;
|
||||||
|
let query = `SELECT id, data FROM ${collection}${clause}`;
|
||||||
|
if (opts?.orderBy) {
|
||||||
|
const dir = opts.order === 'desc' ? 'DESC' : 'ASC';
|
||||||
|
const col =
|
||||||
|
opts.orderBy === 'id'
|
||||||
|
? 'id'
|
||||||
|
: opts.orderBy === 'created_at' || opts.orderBy === 'updated_at'
|
||||||
|
? opts.orderBy
|
||||||
|
: `data->>'${opts.orderBy}'`;
|
||||||
|
query += ` ORDER BY ${col} ${dir}`;
|
||||||
|
}
|
||||||
|
if (opts?.limit !== undefined) {
|
||||||
|
query += ` LIMIT $${paramIdx.toString()}`;
|
||||||
|
params.push(opts.limit);
|
||||||
|
paramIdx++;
|
||||||
|
}
|
||||||
|
if (opts?.offset !== undefined) {
|
||||||
|
query += ` OFFSET $${paramIdx.toString()}`;
|
||||||
|
params.push(opts.offset);
|
||||||
|
paramIdx++;
|
||||||
|
}
|
||||||
|
const result = await pg.query<{ id: string; data: Record<string, unknown> }>(query, params);
|
||||||
|
return result.rows.map((row) => ({ id: row.id, ...(row.data as object) }) as unknown as T);
|
||||||
|
}
|
||||||
|
|
||||||
|
async function pgCount(
|
||||||
|
pg: PgClient,
|
||||||
|
collection: string,
|
||||||
|
filter?: Record<string, unknown>,
|
||||||
|
): Promise<number> {
|
||||||
|
const { clause, params } = buildFilterClause(filter);
|
||||||
|
const result = await pg.query<{ count: string }>(
|
||||||
|
`SELECT COUNT(*) as count FROM ${collection}${clause}`,
|
||||||
|
params,
|
||||||
|
);
|
||||||
|
return parseInt(result.rows[0]?.count ?? '0', 10);
|
||||||
|
}
|
||||||
|
|
||||||
|
export class PgliteAdapter implements StorageAdapter {
|
||||||
|
readonly name = 'pglite';
|
||||||
|
private pg: PGlite;
|
||||||
|
|
||||||
|
constructor(config: Extract<StorageConfig, { type: 'pglite' }>) {
|
||||||
|
this.pg = new PGlite(config.dataDir);
|
||||||
|
}
|
||||||
|
|
||||||
|
async create<T extends Record<string, unknown>>(
|
||||||
|
collection: string,
|
||||||
|
data: T,
|
||||||
|
): Promise<T & { id: string }> {
|
||||||
|
return pgCreate(this.pg, collection, data);
|
||||||
|
}
|
||||||
|
|
||||||
|
async read<T extends Record<string, unknown>>(collection: string, id: string): Promise<T | null> {
|
||||||
|
return pgRead(this.pg, collection, id);
|
||||||
|
}
|
||||||
|
|
||||||
|
async update(collection: string, id: string, data: Record<string, unknown>): Promise<boolean> {
|
||||||
|
return pgUpdate(this.pg, collection, id, data);
|
||||||
|
}
|
||||||
|
|
||||||
|
async delete(collection: string, id: string): Promise<boolean> {
|
||||||
|
return pgDelete(this.pg, collection, id);
|
||||||
|
}
|
||||||
|
|
||||||
|
async find<T extends Record<string, unknown>>(
|
||||||
|
collection: string,
|
||||||
|
filter?: Record<string, unknown>,
|
||||||
|
opts?: { limit?: number; offset?: number; orderBy?: string; order?: 'asc' | 'desc' },
|
||||||
|
): Promise<T[]> {
|
||||||
|
return pgFind(this.pg, collection, filter, opts);
|
||||||
|
}
|
||||||
|
|
||||||
|
async findOne<T extends Record<string, unknown>>(
|
||||||
|
collection: string,
|
||||||
|
filter: Record<string, unknown>,
|
||||||
|
): Promise<T | null> {
|
||||||
|
const results = await this.find<T>(collection, filter, { limit: 1 });
|
||||||
|
return results[0] ?? null;
|
||||||
|
}
|
||||||
|
|
||||||
|
async count(collection: string, filter?: Record<string, unknown>): Promise<number> {
|
||||||
|
return pgCount(this.pg, collection, filter);
|
||||||
|
}
|
||||||
|
|
||||||
|
async transaction<T>(fn: (tx: StorageAdapter) => Promise<T>): Promise<T> {
|
||||||
|
return this.pg.transaction(async (tx) => {
|
||||||
|
const txAdapter = new PgliteTxAdapter(tx as unknown as PgClient);
|
||||||
|
return fn(txAdapter);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async migrate(): Promise<void> {
|
||||||
|
for (const name of COLLECTIONS) {
|
||||||
|
await this.pg.query(`
|
||||||
|
CREATE TABLE IF NOT EXISTS ${name} (
|
||||||
|
id TEXT PRIMARY KEY,
|
||||||
|
data JSONB NOT NULL DEFAULT '{}'::jsonb,
|
||||||
|
created_at TIMESTAMPTZ NOT NULL DEFAULT now(),
|
||||||
|
updated_at TIMESTAMPTZ NOT NULL DEFAULT now()
|
||||||
|
)
|
||||||
|
`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async close(): Promise<void> {
|
||||||
|
await this.pg.close();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Transaction wrapper that delegates to the PGlite transaction connection.
|
||||||
|
*/
|
||||||
|
class PgliteTxAdapter implements StorageAdapter {
|
||||||
|
readonly name = 'pglite';
|
||||||
|
private pg: PgClient;
|
||||||
|
|
||||||
|
constructor(pg: PgClient) {
|
||||||
|
this.pg = pg;
|
||||||
|
}
|
||||||
|
|
||||||
|
async create<T extends Record<string, unknown>>(
|
||||||
|
collection: string,
|
||||||
|
data: T,
|
||||||
|
): Promise<T & { id: string }> {
|
||||||
|
return pgCreate(this.pg, collection, data);
|
||||||
|
}
|
||||||
|
|
||||||
|
async read<T extends Record<string, unknown>>(collection: string, id: string): Promise<T | null> {
|
||||||
|
return pgRead(this.pg, collection, id);
|
||||||
|
}
|
||||||
|
|
||||||
|
async update(collection: string, id: string, data: Record<string, unknown>): Promise<boolean> {
|
||||||
|
return pgUpdate(this.pg, collection, id, data);
|
||||||
|
}
|
||||||
|
|
||||||
|
async delete(collection: string, id: string): Promise<boolean> {
|
||||||
|
return pgDelete(this.pg, collection, id);
|
||||||
|
}
|
||||||
|
|
||||||
|
async find<T extends Record<string, unknown>>(
|
||||||
|
collection: string,
|
||||||
|
filter?: Record<string, unknown>,
|
||||||
|
opts?: { limit?: number; offset?: number; orderBy?: string; order?: 'asc' | 'desc' },
|
||||||
|
): Promise<T[]> {
|
||||||
|
return pgFind(this.pg, collection, filter, opts);
|
||||||
|
}
|
||||||
|
|
||||||
|
async findOne<T extends Record<string, unknown>>(
|
||||||
|
collection: string,
|
||||||
|
filter: Record<string, unknown>,
|
||||||
|
): Promise<T | null> {
|
||||||
|
const results = await this.find<T>(collection, filter, { limit: 1 });
|
||||||
|
return results[0] ?? null;
|
||||||
|
}
|
||||||
|
|
||||||
|
async count(collection: string, filter?: Record<string, unknown>): Promise<number> {
|
||||||
|
return pgCount(this.pg, collection, filter);
|
||||||
|
}
|
||||||
|
|
||||||
|
async transaction<T>(fn: (tx: StorageAdapter) => Promise<T>): Promise<T> {
|
||||||
|
// Already inside a transaction — run directly
|
||||||
|
return fn(this);
|
||||||
|
}
|
||||||
|
|
||||||
|
async migrate(): Promise<void> {
|
||||||
|
// No-op inside transaction
|
||||||
|
}
|
||||||
|
|
||||||
|
async close(): Promise<void> {
|
||||||
|
// No-op inside transaction
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,283 +0,0 @@
|
|||||||
import Database from 'better-sqlite3';
|
|
||||||
import { randomUUID } from 'node:crypto';
|
|
||||||
import type { StorageAdapter, StorageConfig } from '../types.js';
|
|
||||||
|
|
||||||
/* eslint-disable @typescript-eslint/no-explicit-any */
|
|
||||||
|
|
||||||
const COLLECTIONS = [
|
|
||||||
'users',
|
|
||||||
'sessions',
|
|
||||||
'accounts',
|
|
||||||
'projects',
|
|
||||||
'missions',
|
|
||||||
'tasks',
|
|
||||||
'agents',
|
|
||||||
'conversations',
|
|
||||||
'messages',
|
|
||||||
'preferences',
|
|
||||||
'insights',
|
|
||||||
'skills',
|
|
||||||
'events',
|
|
||||||
'routing_rules',
|
|
||||||
'provider_credentials',
|
|
||||||
'agent_logs',
|
|
||||||
'teams',
|
|
||||||
'team_members',
|
|
||||||
'mission_tasks',
|
|
||||||
'tickets',
|
|
||||||
'summarization_jobs',
|
|
||||||
'appreciations',
|
|
||||||
'verifications',
|
|
||||||
] as const;
|
|
||||||
|
|
||||||
function buildFilterClause(filter?: Record<string, unknown>): {
|
|
||||||
clause: string;
|
|
||||||
params: unknown[];
|
|
||||||
} {
|
|
||||||
if (!filter || Object.keys(filter).length === 0) return { clause: '', params: [] };
|
|
||||||
const conditions: string[] = [];
|
|
||||||
const params: unknown[] = [];
|
|
||||||
for (const [key, value] of Object.entries(filter)) {
|
|
||||||
if (key === 'id') {
|
|
||||||
conditions.push('id = ?');
|
|
||||||
params.push(value);
|
|
||||||
} else {
|
|
||||||
conditions.push(`json_extract(data_json, '$.${key}') = ?`);
|
|
||||||
params.push(typeof value === 'object' ? JSON.stringify(value) : value);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return { clause: ` WHERE ${conditions.join(' AND ')}`, params };
|
|
||||||
}
|
|
||||||
|
|
||||||
export class SqliteAdapter implements StorageAdapter {
|
|
||||||
readonly name = 'sqlite';
|
|
||||||
private db: Database.Database;
|
|
||||||
|
|
||||||
constructor(config: Extract<StorageConfig, { type: 'sqlite' }>) {
|
|
||||||
this.db = new Database(config.path);
|
|
||||||
this.db.pragma('journal_mode = WAL');
|
|
||||||
this.db.pragma('foreign_keys = ON');
|
|
||||||
}
|
|
||||||
|
|
||||||
async create<T extends Record<string, unknown>>(
|
|
||||||
collection: string,
|
|
||||||
data: T,
|
|
||||||
): Promise<T & { id: string }> {
|
|
||||||
const id = (data as any).id ?? randomUUID();
|
|
||||||
const now = new Date().toISOString();
|
|
||||||
const rest = Object.fromEntries(Object.entries(data).filter(([k]) => k !== 'id'));
|
|
||||||
this.db
|
|
||||||
.prepare(
|
|
||||||
`INSERT INTO ${collection} (id, data_json, created_at, updated_at) VALUES (?, ?, ?, ?)`,
|
|
||||||
)
|
|
||||||
.run(id, JSON.stringify(rest), now, now);
|
|
||||||
return { ...data, id } as T & { id: string };
|
|
||||||
}
|
|
||||||
|
|
||||||
async read<T extends Record<string, unknown>>(collection: string, id: string): Promise<T | null> {
|
|
||||||
const row = this.db.prepare(`SELECT * FROM ${collection} WHERE id = ?`).get(id) as any;
|
|
||||||
if (!row) return null;
|
|
||||||
return { id: row.id, ...JSON.parse(row.data_json as string) } as T;
|
|
||||||
}
|
|
||||||
|
|
||||||
async update(collection: string, id: string, data: Record<string, unknown>): Promise<boolean> {
|
|
||||||
const existing = this.db
|
|
||||||
.prepare(`SELECT data_json FROM ${collection} WHERE id = ?`)
|
|
||||||
.get(id) as any;
|
|
||||||
if (!existing) return false;
|
|
||||||
const merged = { ...JSON.parse(existing.data_json as string), ...data };
|
|
||||||
const now = new Date().toISOString();
|
|
||||||
const result = this.db
|
|
||||||
.prepare(`UPDATE ${collection} SET data_json = ?, updated_at = ? WHERE id = ?`)
|
|
||||||
.run(JSON.stringify(merged), now, id);
|
|
||||||
return result.changes > 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
async delete(collection: string, id: string): Promise<boolean> {
|
|
||||||
const result = this.db.prepare(`DELETE FROM ${collection} WHERE id = ?`).run(id);
|
|
||||||
return result.changes > 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
async find<T extends Record<string, unknown>>(
|
|
||||||
collection: string,
|
|
||||||
filter?: Record<string, unknown>,
|
|
||||||
opts?: { limit?: number; offset?: number; orderBy?: string; order?: 'asc' | 'desc' },
|
|
||||||
): Promise<T[]> {
|
|
||||||
const { clause, params } = buildFilterClause(filter);
|
|
||||||
let query = `SELECT * FROM ${collection}${clause}`;
|
|
||||||
if (opts?.orderBy) {
|
|
||||||
const dir = opts.order === 'desc' ? 'DESC' : 'ASC';
|
|
||||||
const col =
|
|
||||||
opts.orderBy === 'id' || opts.orderBy === 'created_at' || opts.orderBy === 'updated_at'
|
|
||||||
? opts.orderBy
|
|
||||||
: `json_extract(data_json, '$.${opts.orderBy}')`;
|
|
||||||
query += ` ORDER BY ${col} ${dir}`;
|
|
||||||
}
|
|
||||||
if (opts?.limit) {
|
|
||||||
query += ` LIMIT ?`;
|
|
||||||
params.push(opts.limit);
|
|
||||||
}
|
|
||||||
if (opts?.offset) {
|
|
||||||
query += ` OFFSET ?`;
|
|
||||||
params.push(opts.offset);
|
|
||||||
}
|
|
||||||
const rows = this.db.prepare(query).all(...params) as any[];
|
|
||||||
return rows.map((row) => ({ id: row.id, ...JSON.parse(row.data_json as string) }) as T);
|
|
||||||
}
|
|
||||||
|
|
||||||
async findOne<T extends Record<string, unknown>>(
|
|
||||||
collection: string,
|
|
||||||
filter: Record<string, unknown>,
|
|
||||||
): Promise<T | null> {
|
|
||||||
const results = await this.find<T>(collection, filter, { limit: 1 });
|
|
||||||
return results[0] ?? null;
|
|
||||||
}
|
|
||||||
|
|
||||||
async count(collection: string, filter?: Record<string, unknown>): Promise<number> {
|
|
||||||
const { clause, params } = buildFilterClause(filter);
|
|
||||||
const row = this.db
|
|
||||||
.prepare(`SELECT COUNT(*) as count FROM ${collection}${clause}`)
|
|
||||||
.get(...params) as any;
|
|
||||||
return row?.count ?? 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
async transaction<T>(fn: (tx: StorageAdapter) => Promise<T>): Promise<T> {
|
|
||||||
const txAdapter = new SqliteTxAdapter(this.db);
|
|
||||||
this.db.exec('BEGIN');
|
|
||||||
try {
|
|
||||||
const result = await fn(txAdapter);
|
|
||||||
this.db.exec('COMMIT');
|
|
||||||
return result;
|
|
||||||
} catch (err) {
|
|
||||||
this.db.exec('ROLLBACK');
|
|
||||||
throw err;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async migrate(): Promise<void> {
|
|
||||||
const createTable = (name: string) =>
|
|
||||||
this.db.exec(`
|
|
||||||
CREATE TABLE IF NOT EXISTS ${name} (
|
|
||||||
id TEXT PRIMARY KEY,
|
|
||||||
data_json TEXT NOT NULL DEFAULT '{}',
|
|
||||||
created_at TEXT NOT NULL DEFAULT (datetime('now')),
|
|
||||||
updated_at TEXT NOT NULL DEFAULT (datetime('now'))
|
|
||||||
)
|
|
||||||
`);
|
|
||||||
for (const collection of COLLECTIONS) {
|
|
||||||
createTable(collection);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async close(): Promise<void> {
|
|
||||||
this.db.close();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Transaction wrapper that uses the same db handle — better-sqlite3 transactions
|
|
||||||
* are connection-level, so all statements on the same Database instance within
|
|
||||||
* a db.transaction() callback participate in the transaction.
|
|
||||||
*/
|
|
||||||
class SqliteTxAdapter implements StorageAdapter {
|
|
||||||
readonly name = 'sqlite';
|
|
||||||
private db: Database.Database;
|
|
||||||
|
|
||||||
constructor(db: Database.Database) {
|
|
||||||
this.db = db;
|
|
||||||
}
|
|
||||||
|
|
||||||
async create<T extends Record<string, unknown>>(
|
|
||||||
collection: string,
|
|
||||||
data: T,
|
|
||||||
): Promise<T & { id: string }> {
|
|
||||||
const id = (data as any).id ?? randomUUID();
|
|
||||||
const now = new Date().toISOString();
|
|
||||||
const rest = Object.fromEntries(Object.entries(data).filter(([k]) => k !== 'id'));
|
|
||||||
this.db
|
|
||||||
.prepare(
|
|
||||||
`INSERT INTO ${collection} (id, data_json, created_at, updated_at) VALUES (?, ?, ?, ?)`,
|
|
||||||
)
|
|
||||||
.run(id, JSON.stringify(rest), now, now);
|
|
||||||
return { ...data, id } as T & { id: string };
|
|
||||||
}
|
|
||||||
|
|
||||||
async read<T extends Record<string, unknown>>(collection: string, id: string): Promise<T | null> {
|
|
||||||
const row = this.db.prepare(`SELECT * FROM ${collection} WHERE id = ?`).get(id) as any;
|
|
||||||
if (!row) return null;
|
|
||||||
return { id: row.id, ...JSON.parse(row.data_json as string) } as T;
|
|
||||||
}
|
|
||||||
|
|
||||||
async update(collection: string, id: string, data: Record<string, unknown>): Promise<boolean> {
|
|
||||||
const existing = this.db
|
|
||||||
.prepare(`SELECT data_json FROM ${collection} WHERE id = ?`)
|
|
||||||
.get(id) as any;
|
|
||||||
if (!existing) return false;
|
|
||||||
const merged = { ...JSON.parse(existing.data_json as string), ...data };
|
|
||||||
const now = new Date().toISOString();
|
|
||||||
const result = this.db
|
|
||||||
.prepare(`UPDATE ${collection} SET data_json = ?, updated_at = ? WHERE id = ?`)
|
|
||||||
.run(JSON.stringify(merged), now, id);
|
|
||||||
return result.changes > 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
async delete(collection: string, id: string): Promise<boolean> {
|
|
||||||
const result = this.db.prepare(`DELETE FROM ${collection} WHERE id = ?`).run(id);
|
|
||||||
return result.changes > 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
async find<T extends Record<string, unknown>>(
|
|
||||||
collection: string,
|
|
||||||
filter?: Record<string, unknown>,
|
|
||||||
opts?: { limit?: number; offset?: number; orderBy?: string; order?: 'asc' | 'desc' },
|
|
||||||
): Promise<T[]> {
|
|
||||||
const { clause, params } = buildFilterClause(filter);
|
|
||||||
let query = `SELECT * FROM ${collection}${clause}`;
|
|
||||||
if (opts?.orderBy) {
|
|
||||||
const dir = opts.order === 'desc' ? 'DESC' : 'ASC';
|
|
||||||
const col =
|
|
||||||
opts.orderBy === 'id' || opts.orderBy === 'created_at' || opts.orderBy === 'updated_at'
|
|
||||||
? opts.orderBy
|
|
||||||
: `json_extract(data_json, '$.${opts.orderBy}')`;
|
|
||||||
query += ` ORDER BY ${col} ${dir}`;
|
|
||||||
}
|
|
||||||
if (opts?.limit) {
|
|
||||||
query += ` LIMIT ?`;
|
|
||||||
params.push(opts.limit);
|
|
||||||
}
|
|
||||||
if (opts?.offset) {
|
|
||||||
query += ` OFFSET ?`;
|
|
||||||
params.push(opts.offset);
|
|
||||||
}
|
|
||||||
const rows = this.db.prepare(query).all(...params) as any[];
|
|
||||||
return rows.map((row) => ({ id: row.id, ...JSON.parse(row.data_json as string) }) as T);
|
|
||||||
}
|
|
||||||
|
|
||||||
async findOne<T extends Record<string, unknown>>(
|
|
||||||
collection: string,
|
|
||||||
filter: Record<string, unknown>,
|
|
||||||
): Promise<T | null> {
|
|
||||||
const results = await this.find<T>(collection, filter, { limit: 1 });
|
|
||||||
return results[0] ?? null;
|
|
||||||
}
|
|
||||||
|
|
||||||
async count(collection: string, filter?: Record<string, unknown>): Promise<number> {
|
|
||||||
const { clause, params } = buildFilterClause(filter);
|
|
||||||
const row = this.db
|
|
||||||
.prepare(`SELECT COUNT(*) as count FROM ${collection}${clause}`)
|
|
||||||
.get(...params) as any;
|
|
||||||
return row?.count ?? 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
async transaction<T>(fn: (tx: StorageAdapter) => Promise<T>): Promise<T> {
|
|
||||||
return fn(this);
|
|
||||||
}
|
|
||||||
|
|
||||||
async migrate(): Promise<void> {
|
|
||||||
// No-op inside transaction
|
|
||||||
}
|
|
||||||
|
|
||||||
async close(): Promise<void> {
|
|
||||||
// No-op inside transaction
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,17 +1,17 @@
|
|||||||
export type { StorageAdapter, StorageConfig } from './types.js';
|
export type { StorageAdapter, StorageConfig } from './types.js';
|
||||||
export { createStorageAdapter, registerStorageAdapter } from './factory.js';
|
export { createStorageAdapter, registerStorageAdapter } from './factory.js';
|
||||||
export { PostgresAdapter } from './adapters/postgres.js';
|
export { PostgresAdapter } from './adapters/postgres.js';
|
||||||
export { SqliteAdapter } from './adapters/sqlite.js';
|
export { PgliteAdapter } from './adapters/pglite.js';
|
||||||
|
|
||||||
import { registerStorageAdapter } from './factory.js';
|
import { registerStorageAdapter } from './factory.js';
|
||||||
import { PostgresAdapter } from './adapters/postgres.js';
|
import { PostgresAdapter } from './adapters/postgres.js';
|
||||||
import { SqliteAdapter } from './adapters/sqlite.js';
|
import { PgliteAdapter } from './adapters/pglite.js';
|
||||||
import type { StorageConfig } from './types.js';
|
import type { StorageConfig } from './types.js';
|
||||||
|
|
||||||
registerStorageAdapter('postgres', (config: StorageConfig) => {
|
registerStorageAdapter('postgres', (config: StorageConfig) => {
|
||||||
return new PostgresAdapter(config as Extract<StorageConfig, { type: 'postgres' }>);
|
return new PostgresAdapter(config as Extract<StorageConfig, { type: 'postgres' }>);
|
||||||
});
|
});
|
||||||
|
|
||||||
registerStorageAdapter('sqlite', (config: StorageConfig) => {
|
registerStorageAdapter('pglite', (config: StorageConfig) => {
|
||||||
return new SqliteAdapter(config as Extract<StorageConfig, { type: 'sqlite' }>);
|
return new PgliteAdapter(config as Extract<StorageConfig, { type: 'pglite' }>);
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -39,5 +39,5 @@ export interface StorageAdapter {
|
|||||||
|
|
||||||
export type StorageConfig =
|
export type StorageConfig =
|
||||||
| { type: 'postgres'; url: string }
|
| { type: 'postgres'; url: string }
|
||||||
| { type: 'sqlite'; path: string }
|
| { type: 'pglite'; dataDir?: string }
|
||||||
| { type: 'files'; dataDir: string; format?: 'json' | 'md' };
|
| { type: 'files'; dataDir: string; format?: 'json' | 'md' };
|
||||||
|
|||||||
1712
pnpm-lock.yaml
generated
1712
pnpm-lock.yaml
generated
File diff suppressed because it is too large
Load Diff
Reference in New Issue
Block a user