Compare commits
1 Commits
main
...
fix/base-i
| Author | SHA1 | Date | |
|---|---|---|---|
| 07d44a4c85 |
2
.npmrc
2
.npmrc
@@ -1,3 +1 @@
|
|||||||
@mosaicstack:registry=https://git.mosaicstack.dev/api/packages/mosaic/npm/
|
@mosaicstack:registry=https://git.mosaicstack.dev/api/packages/mosaic/npm/
|
||||||
supportedArchitectures[libc][]=glibc
|
|
||||||
supportedArchitectures[cpu][]=x64
|
|
||||||
|
|||||||
@@ -29,10 +29,9 @@ when:
|
|||||||
- ".trivyignore"
|
- ".trivyignore"
|
||||||
|
|
||||||
variables:
|
variables:
|
||||||
- &node_image "node:24-slim"
|
- &node_image "node:24-alpine"
|
||||||
- &install_deps |
|
- &install_deps |
|
||||||
corepack enable
|
corepack enable
|
||||||
apt-get update && apt-get install -y --no-install-recommends python3 make g++
|
|
||||||
pnpm config set store-dir /root/.local/share/pnpm/store
|
pnpm config set store-dir /root/.local/share/pnpm/store
|
||||||
pnpm install --frozen-lockfile
|
pnpm install --frozen-lockfile
|
||||||
- &use_deps |
|
- &use_deps |
|
||||||
@@ -170,7 +169,7 @@ steps:
|
|||||||
elif [ "$CI_COMMIT_BRANCH" = "main" ]; then
|
elif [ "$CI_COMMIT_BRANCH" = "main" ]; then
|
||||||
DESTINATIONS="--destination git.mosaicstack.dev/mosaic/stack-api:latest"
|
DESTINATIONS="--destination git.mosaicstack.dev/mosaic/stack-api:latest"
|
||||||
fi
|
fi
|
||||||
/kaniko/executor --context . --dockerfile apps/api/Dockerfile --snapshot-mode=redo --cache=true --cache-repo git.mosaicstack.dev/mosaic/stack-api/cache $DESTINATIONS
|
/kaniko/executor --context . --dockerfile apps/api/Dockerfile --snapshot-mode=redo $DESTINATIONS
|
||||||
when:
|
when:
|
||||||
- branch: [main]
|
- branch: [main]
|
||||||
event: [push, manual, tag]
|
event: [push, manual, tag]
|
||||||
@@ -195,7 +194,7 @@ steps:
|
|||||||
elif [ "$CI_COMMIT_BRANCH" = "main" ]; then
|
elif [ "$CI_COMMIT_BRANCH" = "main" ]; then
|
||||||
DESTINATIONS="--destination git.mosaicstack.dev/mosaic/stack-orchestrator:latest"
|
DESTINATIONS="--destination git.mosaicstack.dev/mosaic/stack-orchestrator:latest"
|
||||||
fi
|
fi
|
||||||
/kaniko/executor --context . --dockerfile apps/orchestrator/Dockerfile --snapshot-mode=redo --cache=true --cache-repo git.mosaicstack.dev/mosaic/stack-orchestrator/cache $DESTINATIONS
|
/kaniko/executor --context . --dockerfile apps/orchestrator/Dockerfile --snapshot-mode=redo $DESTINATIONS
|
||||||
when:
|
when:
|
||||||
- branch: [main]
|
- branch: [main]
|
||||||
event: [push, manual, tag]
|
event: [push, manual, tag]
|
||||||
@@ -220,7 +219,7 @@ steps:
|
|||||||
elif [ "$CI_COMMIT_BRANCH" = "main" ]; then
|
elif [ "$CI_COMMIT_BRANCH" = "main" ]; then
|
||||||
DESTINATIONS="--destination git.mosaicstack.dev/mosaic/stack-web:latest"
|
DESTINATIONS="--destination git.mosaicstack.dev/mosaic/stack-web:latest"
|
||||||
fi
|
fi
|
||||||
/kaniko/executor --context . --dockerfile apps/web/Dockerfile --snapshot-mode=redo --cache=true --cache-repo git.mosaicstack.dev/mosaic/stack-web/cache --build-arg NEXT_PUBLIC_API_URL=https://api.mosaicstack.dev $DESTINATIONS
|
/kaniko/executor --context . --dockerfile apps/web/Dockerfile --snapshot-mode=redo --build-arg NEXT_PUBLIC_API_URL=https://api.mosaicstack.dev $DESTINATIONS
|
||||||
when:
|
when:
|
||||||
- branch: [main]
|
- branch: [main]
|
||||||
event: [push, manual, tag]
|
event: [push, manual, tag]
|
||||||
@@ -337,46 +336,3 @@ steps:
|
|||||||
- security-trivy-api
|
- security-trivy-api
|
||||||
- security-trivy-orchestrator
|
- security-trivy-orchestrator
|
||||||
- security-trivy-web
|
- security-trivy-web
|
||||||
|
|
||||||
# ─── Deploy to Docker Swarm via Portainer API (main only) ─────────────────────
|
|
||||||
|
|
||||||
deploy-swarm:
|
|
||||||
image: alpine:3
|
|
||||||
environment:
|
|
||||||
PORTAINER_URL:
|
|
||||||
from_secret: portainer_url
|
|
||||||
PORTAINER_API_KEY:
|
|
||||||
from_secret: portainer_api_key
|
|
||||||
PORTAINER_STACK_ID: "121"
|
|
||||||
commands:
|
|
||||||
- apk add --no-cache curl
|
|
||||||
- |
|
|
||||||
set -e
|
|
||||||
echo "🚀 Deploying to Docker Swarm via Portainer API..."
|
|
||||||
|
|
||||||
# Use Portainer API to update the stack (forces pull of new images)
|
|
||||||
RESPONSE=$(curl -s -w "\n%{http_code}" -X POST \
|
|
||||||
-H "X-API-Key: $PORTAINER_API_KEY" \
|
|
||||||
-H "Content-Type: application/json" \
|
|
||||||
"$PORTAINER_URL/api/stacks/$PORTAINER_STACK_ID/git/redeploy")
|
|
||||||
|
|
||||||
HTTP_CODE=$(echo "$RESPONSE" | tail -1)
|
|
||||||
BODY=$(echo "$RESPONSE" | head -n -1)
|
|
||||||
|
|
||||||
if [ "$HTTP_CODE" = "200" ] || [ "$HTTP_CODE" = "202" ]; then
|
|
||||||
echo "✅ Stack update triggered successfully"
|
|
||||||
else
|
|
||||||
echo "❌ Stack update failed (HTTP $HTTP_CODE)"
|
|
||||||
echo "$BODY"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Wait for services to converge
|
|
||||||
echo "⏳ Waiting for services to converge..."
|
|
||||||
sleep 30
|
|
||||||
echo "✅ Deploy complete"
|
|
||||||
when:
|
|
||||||
- branch: [main]
|
|
||||||
event: [push, manual, tag]
|
|
||||||
depends_on:
|
|
||||||
- link-packages
|
|
||||||
|
|||||||
@@ -30,9 +30,6 @@ COPY packages/ui/package.json ./packages/ui/
|
|||||||
COPY packages/config/package.json ./packages/config/
|
COPY packages/config/package.json ./packages/config/
|
||||||
COPY apps/api/package.json ./apps/api/
|
COPY apps/api/package.json ./apps/api/
|
||||||
|
|
||||||
# Copy npm configuration for native binary architecture hints
|
|
||||||
COPY .npmrc ./
|
|
||||||
|
|
||||||
# Install dependencies (no cache mount — Kaniko builds are ephemeral in CI)
|
# Install dependencies (no cache mount — Kaniko builds are ephemeral in CI)
|
||||||
# Then explicitly rebuild node-pty from source since pnpm may skip postinstall
|
# Then explicitly rebuild node-pty from source since pnpm may skip postinstall
|
||||||
# scripts or fail to find prebuilt binaries for this Node.js version
|
# scripts or fail to find prebuilt binaries for this Node.js version
|
||||||
|
|||||||
@@ -1,13 +0,0 @@
|
|||||||
-- MS21: Add admin, local auth, and invitation fields to users table
|
|
||||||
-- These columns were added to schema.prisma but never captured in a migration.
|
|
||||||
|
|
||||||
ALTER TABLE "users"
|
|
||||||
ADD COLUMN IF NOT EXISTS "deactivated_at" TIMESTAMPTZ,
|
|
||||||
ADD COLUMN IF NOT EXISTS "is_local_auth" BOOLEAN NOT NULL DEFAULT false,
|
|
||||||
ADD COLUMN IF NOT EXISTS "password_hash" TEXT,
|
|
||||||
ADD COLUMN IF NOT EXISTS "invited_by" UUID,
|
|
||||||
ADD COLUMN IF NOT EXISTS "invitation_token" TEXT,
|
|
||||||
ADD COLUMN IF NOT EXISTS "invited_at" TIMESTAMPTZ;
|
|
||||||
|
|
||||||
-- CreateIndex
|
|
||||||
CREATE UNIQUE INDEX IF NOT EXISTS "users_invitation_token_key" ON "users"("invitation_token");
|
|
||||||
@@ -1,79 +1,31 @@
|
|||||||
import { Body, Controller, HttpException, Logger, Post, Req, Res, UseGuards } from "@nestjs/common";
|
import {
|
||||||
|
Body,
|
||||||
|
Controller,
|
||||||
|
HttpException,
|
||||||
|
Logger,
|
||||||
|
Post,
|
||||||
|
Req,
|
||||||
|
Res,
|
||||||
|
UnauthorizedException,
|
||||||
|
UseGuards,
|
||||||
|
} from "@nestjs/common";
|
||||||
import type { Response } from "express";
|
import type { Response } from "express";
|
||||||
import { AuthGuard } from "../auth/guards/auth.guard";
|
import { AuthGuard } from "../auth/guards/auth.guard";
|
||||||
import { SkipCsrf } from "../common/decorators/skip-csrf.decorator";
|
|
||||||
import type { MaybeAuthenticatedRequest } from "../auth/types/better-auth-request.interface";
|
import type { MaybeAuthenticatedRequest } from "../auth/types/better-auth-request.interface";
|
||||||
import { ChatStreamDto } from "./chat-proxy.dto";
|
import { ChatStreamDto } from "./chat-proxy.dto";
|
||||||
import { ChatProxyService } from "./chat-proxy.service";
|
import { ChatProxyService } from "./chat-proxy.service";
|
||||||
|
|
||||||
@Controller("chat")
|
@Controller("chat")
|
||||||
|
@UseGuards(AuthGuard)
|
||||||
export class ChatProxyController {
|
export class ChatProxyController {
|
||||||
private readonly logger = new Logger(ChatProxyController.name);
|
private readonly logger = new Logger(ChatProxyController.name);
|
||||||
|
|
||||||
constructor(private readonly chatProxyService: ChatProxyService) {}
|
constructor(private readonly chatProxyService: ChatProxyService) {}
|
||||||
|
|
||||||
// POST /api/chat/guest
|
|
||||||
// Guest chat endpoint - no authentication required
|
|
||||||
// Uses a shared LLM configuration for unauthenticated users
|
|
||||||
@SkipCsrf()
|
|
||||||
@Post("guest")
|
|
||||||
async guestChat(
|
|
||||||
@Body() body: ChatStreamDto,
|
|
||||||
@Req() req: MaybeAuthenticatedRequest,
|
|
||||||
@Res() res: Response
|
|
||||||
): Promise<void> {
|
|
||||||
const abortController = new AbortController();
|
|
||||||
req.once("close", () => {
|
|
||||||
abortController.abort();
|
|
||||||
});
|
|
||||||
|
|
||||||
res.setHeader("Content-Type", "text/event-stream");
|
|
||||||
res.setHeader("Cache-Control", "no-cache");
|
|
||||||
res.setHeader("Connection", "keep-alive");
|
|
||||||
res.setHeader("X-Accel-Buffering", "no");
|
|
||||||
|
|
||||||
try {
|
|
||||||
const upstreamResponse = await this.chatProxyService.proxyGuestChat(
|
|
||||||
body.messages,
|
|
||||||
abortController.signal
|
|
||||||
);
|
|
||||||
|
|
||||||
const upstreamContentType = upstreamResponse.headers.get("content-type");
|
|
||||||
if (upstreamContentType) {
|
|
||||||
res.setHeader("Content-Type", upstreamContentType);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!upstreamResponse.body) {
|
|
||||||
throw new Error("LLM response did not include a stream body");
|
|
||||||
}
|
|
||||||
|
|
||||||
for await (const chunk of upstreamResponse.body as unknown as AsyncIterable<Uint8Array>) {
|
|
||||||
if (res.writableEnded || res.destroyed) {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
res.write(Buffer.from(chunk));
|
|
||||||
}
|
|
||||||
} catch (error: unknown) {
|
|
||||||
this.logStreamError(error);
|
|
||||||
|
|
||||||
if (!res.writableEnded && !res.destroyed) {
|
|
||||||
res.write("event: error\n");
|
|
||||||
res.write(`data: ${JSON.stringify({ error: this.toSafeClientMessage(error) })}\n\n`);
|
|
||||||
}
|
|
||||||
} finally {
|
|
||||||
if (!res.writableEnded && !res.destroyed) {
|
|
||||||
res.end();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// POST /api/chat/stream
|
// POST /api/chat/stream
|
||||||
// Request: { messages: Array<{role, content}> }
|
// Request: { messages: Array<{role, content}> }
|
||||||
// Response: SSE stream of chat completion events
|
// Response: SSE stream of chat completion events
|
||||||
// Requires authentication - uses user's personal OpenClaw container
|
|
||||||
@Post("stream")
|
@Post("stream")
|
||||||
@UseGuards(AuthGuard)
|
|
||||||
async streamChat(
|
async streamChat(
|
||||||
@Body() body: ChatStreamDto,
|
@Body() body: ChatStreamDto,
|
||||||
@Req() req: MaybeAuthenticatedRequest,
|
@Req() req: MaybeAuthenticatedRequest,
|
||||||
@@ -81,8 +33,7 @@ export class ChatProxyController {
|
|||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const userId = req.user?.id;
|
const userId = req.user?.id;
|
||||||
if (!userId) {
|
if (!userId) {
|
||||||
this.logger.warn("streamChat called without user ID after AuthGuard");
|
throw new UnauthorizedException("No authenticated user found on request");
|
||||||
throw new HttpException("Authentication required", 401);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const abortController = new AbortController();
|
const abortController = new AbortController();
|
||||||
|
|||||||
@@ -1,5 +1,4 @@
|
|||||||
import { Module } from "@nestjs/common";
|
import { Module } from "@nestjs/common";
|
||||||
import { ConfigModule } from "@nestjs/config";
|
|
||||||
import { AuthModule } from "../auth/auth.module";
|
import { AuthModule } from "../auth/auth.module";
|
||||||
import { AgentConfigModule } from "../agent-config/agent-config.module";
|
import { AgentConfigModule } from "../agent-config/agent-config.module";
|
||||||
import { ContainerLifecycleModule } from "../container-lifecycle/container-lifecycle.module";
|
import { ContainerLifecycleModule } from "../container-lifecycle/container-lifecycle.module";
|
||||||
@@ -8,7 +7,7 @@ import { ChatProxyController } from "./chat-proxy.controller";
|
|||||||
import { ChatProxyService } from "./chat-proxy.service";
|
import { ChatProxyService } from "./chat-proxy.service";
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [AuthModule, PrismaModule, ContainerLifecycleModule, AgentConfigModule, ConfigModule],
|
imports: [AuthModule, PrismaModule, ContainerLifecycleModule, AgentConfigModule],
|
||||||
controllers: [ChatProxyController],
|
controllers: [ChatProxyController],
|
||||||
providers: [ChatProxyService],
|
providers: [ChatProxyService],
|
||||||
exports: [ChatProxyService],
|
exports: [ChatProxyService],
|
||||||
|
|||||||
@@ -4,14 +4,11 @@ import {
|
|||||||
Logger,
|
Logger,
|
||||||
ServiceUnavailableException,
|
ServiceUnavailableException,
|
||||||
} from "@nestjs/common";
|
} from "@nestjs/common";
|
||||||
import { ConfigService } from "@nestjs/config";
|
|
||||||
import { ContainerLifecycleService } from "../container-lifecycle/container-lifecycle.service";
|
import { ContainerLifecycleService } from "../container-lifecycle/container-lifecycle.service";
|
||||||
import { PrismaService } from "../prisma/prisma.service";
|
import { PrismaService } from "../prisma/prisma.service";
|
||||||
import type { ChatMessage } from "./chat-proxy.dto";
|
import type { ChatMessage } from "./chat-proxy.dto";
|
||||||
|
|
||||||
const DEFAULT_OPENCLAW_MODEL = "openclaw:default";
|
const DEFAULT_OPENCLAW_MODEL = "openclaw:default";
|
||||||
const DEFAULT_GUEST_LLM_URL = "http://10.1.1.42:11434/v1";
|
|
||||||
const DEFAULT_GUEST_LLM_MODEL = "llama3.2";
|
|
||||||
|
|
||||||
interface ContainerConnection {
|
interface ContainerConnection {
|
||||||
url: string;
|
url: string;
|
||||||
@@ -24,8 +21,7 @@ export class ChatProxyService {
|
|||||||
|
|
||||||
constructor(
|
constructor(
|
||||||
private readonly prisma: PrismaService,
|
private readonly prisma: PrismaService,
|
||||||
private readonly containerLifecycle: ContainerLifecycleService,
|
private readonly containerLifecycle: ContainerLifecycleService
|
||||||
private readonly config: ConfigService
|
|
||||||
) {}
|
) {}
|
||||||
|
|
||||||
// Get the user's OpenClaw container URL and mark it active.
|
// Get the user's OpenClaw container URL and mark it active.
|
||||||
@@ -83,65 +79,6 @@ export class ChatProxyService {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Proxy guest chat request to configured LLM endpoint.
|
|
||||||
* Uses environment variables for configuration:
|
|
||||||
* - GUEST_LLM_URL: OpenAI-compatible endpoint URL
|
|
||||||
* - GUEST_LLM_API_KEY: API key (optional, for cloud providers)
|
|
||||||
* - GUEST_LLM_MODEL: Model name to use
|
|
||||||
*/
|
|
||||||
async proxyGuestChat(messages: ChatMessage[], signal?: AbortSignal): Promise<Response> {
|
|
||||||
const llmUrl = this.config.get<string>("GUEST_LLM_URL") ?? DEFAULT_GUEST_LLM_URL;
|
|
||||||
const llmApiKey = this.config.get<string>("GUEST_LLM_API_KEY");
|
|
||||||
const llmModel = this.config.get<string>("GUEST_LLM_MODEL") ?? DEFAULT_GUEST_LLM_MODEL;
|
|
||||||
|
|
||||||
const headers: Record<string, string> = {
|
|
||||||
"Content-Type": "application/json",
|
|
||||||
};
|
|
||||||
|
|
||||||
if (llmApiKey) {
|
|
||||||
headers.Authorization = `Bearer ${llmApiKey}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
const requestInit: RequestInit = {
|
|
||||||
method: "POST",
|
|
||||||
headers,
|
|
||||||
body: JSON.stringify({
|
|
||||||
messages,
|
|
||||||
model: llmModel,
|
|
||||||
stream: true,
|
|
||||||
}),
|
|
||||||
};
|
|
||||||
|
|
||||||
if (signal) {
|
|
||||||
requestInit.signal = signal;
|
|
||||||
}
|
|
||||||
|
|
||||||
try {
|
|
||||||
this.logger.debug(`Guest chat proxying to ${llmUrl} with model ${llmModel}`);
|
|
||||||
const response = await fetch(`${llmUrl}/chat/completions`, requestInit);
|
|
||||||
|
|
||||||
if (!response.ok) {
|
|
||||||
const detail = await this.readResponseText(response);
|
|
||||||
const status = `${String(response.status)} ${response.statusText}`.trim();
|
|
||||||
this.logger.warn(
|
|
||||||
detail ? `Guest LLM returned ${status}: ${detail}` : `Guest LLM returned ${status}`
|
|
||||||
);
|
|
||||||
throw new BadGatewayException(`Guest LLM returned ${status}`);
|
|
||||||
}
|
|
||||||
|
|
||||||
return response;
|
|
||||||
} catch (error: unknown) {
|
|
||||||
if (error instanceof BadGatewayException) {
|
|
||||||
throw error;
|
|
||||||
}
|
|
||||||
|
|
||||||
const message = error instanceof Error ? error.message : String(error);
|
|
||||||
this.logger.warn(`Failed to proxy guest chat request: ${message}`);
|
|
||||||
throw new ServiceUnavailableException("Failed to proxy guest chat to LLM");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private async getContainerConnection(userId: string): Promise<ContainerConnection> {
|
private async getContainerConnection(userId: string): Promise<ContainerConnection> {
|
||||||
const connection = await this.containerLifecycle.ensureRunning(userId);
|
const connection = await this.containerLifecycle.ensureRunning(userId);
|
||||||
await this.containerLifecycle.touch(userId);
|
await this.containerLifecycle.touch(userId);
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { Controller, Get, Query, Res, UseGuards } from "@nestjs/common";
|
import { Controller, Get, Res, UseGuards } from "@nestjs/common";
|
||||||
import { AgentStatus } from "@prisma/client";
|
import { AgentStatus } from "@prisma/client";
|
||||||
import type { Response } from "express";
|
import type { Response } from "express";
|
||||||
import { AuthGuard } from "../auth/guards/auth.guard";
|
import { AuthGuard } from "../auth/guards/auth.guard";
|
||||||
@@ -6,7 +6,6 @@ import { PrismaService } from "../prisma/prisma.service";
|
|||||||
|
|
||||||
const AGENT_POLL_INTERVAL_MS = 5_000;
|
const AGENT_POLL_INTERVAL_MS = 5_000;
|
||||||
const SSE_HEARTBEAT_MS = 15_000;
|
const SSE_HEARTBEAT_MS = 15_000;
|
||||||
const DEFAULT_EVENTS_LIMIT = 25;
|
|
||||||
|
|
||||||
interface OrchestratorAgentDto {
|
interface OrchestratorAgentDto {
|
||||||
id: string;
|
id: string;
|
||||||
@@ -16,26 +15,6 @@ interface OrchestratorAgentDto {
|
|||||||
createdAt: Date;
|
createdAt: Date;
|
||||||
}
|
}
|
||||||
|
|
||||||
interface OrchestratorEventDto {
|
|
||||||
type: string;
|
|
||||||
timestamp: string;
|
|
||||||
agentId?: string;
|
|
||||||
taskId?: string;
|
|
||||||
data?: Record<string, unknown>;
|
|
||||||
}
|
|
||||||
|
|
||||||
interface OrchestratorHealthDto {
|
|
||||||
status: "healthy" | "degraded" | "unhealthy";
|
|
||||||
database: "connected" | "disconnected";
|
|
||||||
agents: {
|
|
||||||
total: number;
|
|
||||||
working: number;
|
|
||||||
idle: number;
|
|
||||||
errored: number;
|
|
||||||
};
|
|
||||||
timestamp: string;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Controller("orchestrator")
|
@Controller("orchestrator")
|
||||||
@UseGuards(AuthGuard)
|
@UseGuards(AuthGuard)
|
||||||
export class OrchestratorController {
|
export class OrchestratorController {
|
||||||
@@ -46,81 +25,6 @@ export class OrchestratorController {
|
|||||||
return this.fetchActiveAgents();
|
return this.fetchActiveAgents();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Get("events/recent")
|
|
||||||
async getRecentEvents(
|
|
||||||
@Query("limit") limit?: string
|
|
||||||
): Promise<{ events: OrchestratorEventDto[] }> {
|
|
||||||
const eventsLimit = limit ? parseInt(limit, 10) : DEFAULT_EVENTS_LIMIT;
|
|
||||||
const safeLimit = Math.min(Math.max(eventsLimit, 1), 100);
|
|
||||||
|
|
||||||
// Fetch recent agent activity to derive events
|
|
||||||
const agents = await this.prisma.agent.findMany({
|
|
||||||
where: {
|
|
||||||
status: {
|
|
||||||
not: AgentStatus.TERMINATED,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
orderBy: {
|
|
||||||
createdAt: "desc",
|
|
||||||
},
|
|
||||||
take: safeLimit,
|
|
||||||
});
|
|
||||||
|
|
||||||
// Derive events from agent status changes
|
|
||||||
const events: OrchestratorEventDto[] = agents.map((agent) => ({
|
|
||||||
type: `agent:${agent.status.toLowerCase()}`,
|
|
||||||
timestamp: agent.createdAt.toISOString(),
|
|
||||||
agentId: agent.id,
|
|
||||||
data: {
|
|
||||||
name: agent.name,
|
|
||||||
role: agent.role,
|
|
||||||
model: agent.model,
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
return { events };
|
|
||||||
}
|
|
||||||
|
|
||||||
@Get("health")
|
|
||||||
async getHealth(): Promise<OrchestratorHealthDto> {
|
|
||||||
let databaseConnected = false;
|
|
||||||
let agents: OrchestratorAgentDto[] = [];
|
|
||||||
|
|
||||||
try {
|
|
||||||
// Check database connectivity
|
|
||||||
await this.prisma.$queryRaw`SELECT 1`;
|
|
||||||
databaseConnected = true;
|
|
||||||
|
|
||||||
// Get agent counts
|
|
||||||
agents = await this.fetchActiveAgents();
|
|
||||||
} catch {
|
|
||||||
databaseConnected = false;
|
|
||||||
}
|
|
||||||
|
|
||||||
const working = agents.filter((a) => a.status === AgentStatus.WORKING).length;
|
|
||||||
const idle = agents.filter((a) => a.status === AgentStatus.IDLE).length;
|
|
||||||
const errored = agents.filter((a) => a.status === AgentStatus.ERROR).length;
|
|
||||||
|
|
||||||
let status: OrchestratorHealthDto["status"] = "healthy";
|
|
||||||
if (!databaseConnected) {
|
|
||||||
status = "unhealthy";
|
|
||||||
} else if (errored > 0) {
|
|
||||||
status = "degraded";
|
|
||||||
}
|
|
||||||
|
|
||||||
return {
|
|
||||||
status,
|
|
||||||
database: databaseConnected ? "connected" : "disconnected",
|
|
||||||
agents: {
|
|
||||||
total: agents.length,
|
|
||||||
working,
|
|
||||||
idle,
|
|
||||||
errored,
|
|
||||||
},
|
|
||||||
timestamp: new Date().toISOString(),
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
@Get("events")
|
@Get("events")
|
||||||
async streamEvents(@Res() res: Response): Promise<void> {
|
async streamEvents(@Res() res: Response): Promise<void> {
|
||||||
res.setHeader("Content-Type", "text/event-stream");
|
res.setHeader("Content-Type", "text/event-stream");
|
||||||
|
|||||||
@@ -601,21 +601,9 @@ class TestCoordinatorIntegration:
|
|||||||
coordinator = Coordinator(queue_manager=queue_manager, poll_interval=0.02)
|
coordinator = Coordinator(queue_manager=queue_manager, poll_interval=0.02)
|
||||||
|
|
||||||
task = asyncio.create_task(coordinator.start())
|
task = asyncio.create_task(coordinator.start())
|
||||||
|
await asyncio.sleep(0.5) # Allow time for processing
|
||||||
# Poll for completion with timeout instead of fixed sleep
|
|
||||||
deadline = asyncio.get_event_loop().time() + 5.0 # 5 second timeout
|
|
||||||
while asyncio.get_event_loop().time() < deadline:
|
|
||||||
all_completed = True
|
|
||||||
for i in range(157, 162):
|
|
||||||
item = queue_manager.get_item(i)
|
|
||||||
if item is None or item.status != QueueItemStatus.COMPLETED:
|
|
||||||
all_completed = False
|
|
||||||
break
|
|
||||||
if all_completed:
|
|
||||||
break
|
|
||||||
await asyncio.sleep(0.05)
|
|
||||||
|
|
||||||
await coordinator.stop()
|
await coordinator.stop()
|
||||||
|
|
||||||
task.cancel()
|
task.cancel()
|
||||||
try:
|
try:
|
||||||
await task
|
await task
|
||||||
|
|||||||
@@ -22,9 +22,6 @@ COPY packages/shared/package.json ./packages/shared/
|
|||||||
COPY packages/config/package.json ./packages/config/
|
COPY packages/config/package.json ./packages/config/
|
||||||
COPY apps/orchestrator/package.json ./apps/orchestrator/
|
COPY apps/orchestrator/package.json ./apps/orchestrator/
|
||||||
|
|
||||||
# Copy npm configuration for native binary architecture hints
|
|
||||||
COPY .npmrc ./
|
|
||||||
|
|
||||||
# Install ALL dependencies (not just production)
|
# Install ALL dependencies (not just production)
|
||||||
# No cache mount — Kaniko builds are ephemeral in CI
|
# No cache mount — Kaniko builds are ephemeral in CI
|
||||||
RUN pnpm install --frozen-lockfile
|
RUN pnpm install --frozen-lockfile
|
||||||
|
|||||||
@@ -24,9 +24,6 @@ COPY packages/ui/package.json ./packages/ui/
|
|||||||
COPY packages/config/package.json ./packages/config/
|
COPY packages/config/package.json ./packages/config/
|
||||||
COPY apps/web/package.json ./apps/web/
|
COPY apps/web/package.json ./apps/web/
|
||||||
|
|
||||||
# Copy npm configuration for native binary architecture hints
|
|
||||||
COPY .npmrc ./
|
|
||||||
|
|
||||||
# Install dependencies (no cache mount — Kaniko builds are ephemeral in CI)
|
# Install dependencies (no cache mount — Kaniko builds are ephemeral in CI)
|
||||||
RUN pnpm install --frozen-lockfile
|
RUN pnpm install --frozen-lockfile
|
||||||
|
|
||||||
@@ -41,9 +38,6 @@ COPY packages/ui/package.json ./packages/ui/
|
|||||||
COPY packages/config/package.json ./packages/config/
|
COPY packages/config/package.json ./packages/config/
|
||||||
COPY apps/web/package.json ./apps/web/
|
COPY apps/web/package.json ./apps/web/
|
||||||
|
|
||||||
# Copy npm configuration for native binary architecture hints
|
|
||||||
COPY .npmrc ./
|
|
||||||
|
|
||||||
# Install production dependencies only
|
# Install production dependencies only
|
||||||
RUN pnpm install --frozen-lockfile --prod
|
RUN pnpm install --frozen-lockfile --prod
|
||||||
|
|
||||||
|
|||||||
@@ -352,7 +352,7 @@ export const Chat = forwardRef<ChatRef, ChatProps>(function Chat(
|
|||||||
<div className="mx-auto max-w-4xl px-4 py-4 lg:px-8">
|
<div className="mx-auto max-w-4xl px-4 py-4 lg:px-8">
|
||||||
<ChatInput
|
<ChatInput
|
||||||
onSend={handleSendMessage}
|
onSend={handleSendMessage}
|
||||||
disabled={isChatLoading}
|
disabled={isChatLoading || !user}
|
||||||
inputRef={inputRef}
|
inputRef={inputRef}
|
||||||
isStreaming={isStreaming}
|
isStreaming={isStreaming}
|
||||||
onStopStreaming={abortStream}
|
onStopStreaming={abortStream}
|
||||||
|
|||||||
@@ -16,21 +16,6 @@ interface Agent {
|
|||||||
error?: string;
|
error?: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
function isWorking(status: string): boolean {
|
|
||||||
const s = status.toLowerCase();
|
|
||||||
return s === "running" || s === "working";
|
|
||||||
}
|
|
||||||
|
|
||||||
function isIdle(status: string): boolean {
|
|
||||||
const s = status.toLowerCase();
|
|
||||||
return s === "idle" || s === "spawning" || s === "waiting" || s === "queued";
|
|
||||||
}
|
|
||||||
|
|
||||||
function isErrored(status: string): boolean {
|
|
||||||
const s = status.toLowerCase();
|
|
||||||
return s === "failed" || s === "error";
|
|
||||||
}
|
|
||||||
|
|
||||||
export function AgentStatusWidget({ id: _id, config: _config }: WidgetProps): React.JSX.Element {
|
export function AgentStatusWidget({ id: _id, config: _config }: WidgetProps): React.JSX.Element {
|
||||||
const [agents, setAgents] = useState<Agent[]>([]);
|
const [agents, setAgents] = useState<Agent[]>([]);
|
||||||
const [isLoading, setIsLoading] = useState(true);
|
const [isLoading, setIsLoading] = useState(true);
|
||||||
@@ -89,20 +74,25 @@ export function AgentStatusWidget({ id: _id, config: _config }: WidgetProps): Re
|
|||||||
}, [fetchAgents]);
|
}, [fetchAgents]);
|
||||||
|
|
||||||
const getStatusIcon = (status: string): React.JSX.Element => {
|
const getStatusIcon = (status: string): React.JSX.Element => {
|
||||||
if (isWorking(status)) {
|
const statusLower = status.toLowerCase();
|
||||||
|
switch (statusLower) {
|
||||||
|
case "running":
|
||||||
|
case "working":
|
||||||
return <Activity className="w-4 h-4 text-blue-500 animate-pulse" />;
|
return <Activity className="w-4 h-4 text-blue-500 animate-pulse" />;
|
||||||
}
|
case "spawning":
|
||||||
if (isIdle(status)) {
|
case "queued":
|
||||||
return <Clock className="w-4 h-4 text-yellow-500" />;
|
return <Clock className="w-4 h-4 text-yellow-500" />;
|
||||||
}
|
case "completed":
|
||||||
if (isErrored(status)) {
|
return <CheckCircle className="w-4 h-4 text-green-500" />;
|
||||||
|
case "failed":
|
||||||
|
case "error":
|
||||||
return <AlertCircle className="w-4 h-4 text-red-500" />;
|
return <AlertCircle className="w-4 h-4 text-red-500" />;
|
||||||
}
|
case "terminated":
|
||||||
const s = status.toLowerCase();
|
case "killed":
|
||||||
if (s === "completed" || s === "terminated" || s === "killed") {
|
|
||||||
return <CheckCircle className="w-4 h-4 text-gray-500" />;
|
return <CheckCircle className="w-4 h-4 text-gray-500" />;
|
||||||
}
|
default:
|
||||||
return <Clock className="w-4 h-4 text-gray-400" />;
|
return <Clock className="w-4 h-4 text-gray-400" />;
|
||||||
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
const getStatusText = (status: string): string => {
|
const getStatusText = (status: string): string => {
|
||||||
@@ -131,9 +121,9 @@ export function AgentStatusWidget({ id: _id, config: _config }: WidgetProps): Re
|
|||||||
|
|
||||||
const stats = {
|
const stats = {
|
||||||
total: agents.length,
|
total: agents.length,
|
||||||
working: agents.filter((a) => isWorking(a.status)).length,
|
working: agents.filter((a) => a.status.toLowerCase() === "running").length,
|
||||||
idle: agents.filter((a) => isIdle(a.status)).length,
|
idle: agents.filter((a) => a.status.toLowerCase() === "spawning").length,
|
||||||
error: agents.filter((a) => isErrored(a.status)).length,
|
error: agents.filter((a) => a.status.toLowerCase() === "failed").length,
|
||||||
};
|
};
|
||||||
|
|
||||||
if (isLoading) {
|
if (isLoading) {
|
||||||
@@ -186,9 +176,9 @@ export function AgentStatusWidget({ id: _id, config: _config }: WidgetProps): Re
|
|||||||
<div
|
<div
|
||||||
key={agent.agentId}
|
key={agent.agentId}
|
||||||
className={`p-3 rounded-lg border ${
|
className={`p-3 rounded-lg border ${
|
||||||
isErrored(agent.status)
|
agent.status.toLowerCase() === "failed"
|
||||||
? "bg-red-50 border-red-200"
|
? "bg-red-50 border-red-200"
|
||||||
: isWorking(agent.status)
|
: agent.status.toLowerCase() === "running"
|
||||||
? "bg-blue-50 border-blue-200"
|
? "bg-blue-50 border-blue-200"
|
||||||
: "bg-gray-50 border-gray-200"
|
: "bg-gray-50 border-gray-200"
|
||||||
}`}
|
}`}
|
||||||
|
|||||||
@@ -4,43 +4,61 @@
|
|||||||
|
|
||||||
import { useState, useEffect } from "react";
|
import { useState, useEffect } from "react";
|
||||||
import { Calendar as CalendarIcon, Clock, MapPin } from "lucide-react";
|
import { Calendar as CalendarIcon, Clock, MapPin } from "lucide-react";
|
||||||
import type { WidgetProps, Event } from "@mosaic/shared";
|
import type { WidgetProps } from "@mosaic/shared";
|
||||||
import { fetchEvents } from "@/lib/api/events";
|
|
||||||
|
interface Event {
|
||||||
|
id: string;
|
||||||
|
title: string;
|
||||||
|
startTime: string;
|
||||||
|
endTime?: string;
|
||||||
|
location?: string;
|
||||||
|
allDay: boolean;
|
||||||
|
}
|
||||||
|
|
||||||
export function CalendarWidget({ id: _id, config: _config }: WidgetProps): React.JSX.Element {
|
export function CalendarWidget({ id: _id, config: _config }: WidgetProps): React.JSX.Element {
|
||||||
const [events, setEvents] = useState<Event[]>([]);
|
const [events, setEvents] = useState<Event[]>([]);
|
||||||
const [isLoading, setIsLoading] = useState(true);
|
const [isLoading, setIsLoading] = useState(true);
|
||||||
|
|
||||||
|
// Mock data for now - will fetch from API later
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
let isMounted = true;
|
|
||||||
|
|
||||||
const loadEvents = async (): Promise<void> => {
|
|
||||||
setIsLoading(true);
|
setIsLoading(true);
|
||||||
try {
|
const now = new Date();
|
||||||
const data = await fetchEvents();
|
const today = new Date(now.getFullYear(), now.getMonth(), now.getDate());
|
||||||
if (isMounted) {
|
const tomorrow = new Date(today);
|
||||||
setEvents(data);
|
tomorrow.setDate(tomorrow.getDate() + 1);
|
||||||
}
|
|
||||||
} catch {
|
setTimeout(() => {
|
||||||
if (isMounted) {
|
setEvents([
|
||||||
setEvents([]);
|
{
|
||||||
}
|
id: "1",
|
||||||
} finally {
|
title: "Team Standup",
|
||||||
if (isMounted) {
|
startTime: new Date(today.setHours(9, 0, 0, 0)).toISOString(),
|
||||||
|
endTime: new Date(today.setHours(9, 30, 0, 0)).toISOString(),
|
||||||
|
location: "Zoom",
|
||||||
|
allDay: false,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: "2",
|
||||||
|
title: "Project Review",
|
||||||
|
startTime: new Date(today.setHours(14, 0, 0, 0)).toISOString(),
|
||||||
|
endTime: new Date(today.setHours(15, 0, 0, 0)).toISOString(),
|
||||||
|
location: "Conference Room A",
|
||||||
|
allDay: false,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: "3",
|
||||||
|
title: "Sprint Planning",
|
||||||
|
startTime: new Date(tomorrow.setHours(10, 0, 0, 0)).toISOString(),
|
||||||
|
endTime: new Date(tomorrow.setHours(12, 0, 0, 0)).toISOString(),
|
||||||
|
allDay: false,
|
||||||
|
},
|
||||||
|
]);
|
||||||
setIsLoading(false);
|
setIsLoading(false);
|
||||||
}
|
}, 500);
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
void loadEvents();
|
|
||||||
|
|
||||||
return (): void => {
|
|
||||||
isMounted = false;
|
|
||||||
};
|
|
||||||
}, []);
|
}, []);
|
||||||
|
|
||||||
const formatTime = (dateValue: Date | string): string => {
|
const formatTime = (dateString: string): string => {
|
||||||
const date = new Date(dateValue);
|
const date = new Date(dateString);
|
||||||
return date.toLocaleTimeString("en-US", {
|
return date.toLocaleTimeString("en-US", {
|
||||||
hour: "numeric",
|
hour: "numeric",
|
||||||
minute: "2-digit",
|
minute: "2-digit",
|
||||||
@@ -48,8 +66,8 @@ export function CalendarWidget({ id: _id, config: _config }: WidgetProps): React
|
|||||||
});
|
});
|
||||||
};
|
};
|
||||||
|
|
||||||
const formatDay = (dateValue: Date | string): string => {
|
const formatDay = (dateString: string): string => {
|
||||||
const date = new Date(dateValue);
|
const date = new Date(dateString);
|
||||||
const today = new Date();
|
const today = new Date();
|
||||||
const tomorrow = new Date(today);
|
const tomorrow = new Date(today);
|
||||||
tomorrow.setDate(tomorrow.getDate() + 1);
|
tomorrow.setDate(tomorrow.getDate() + 1);
|
||||||
|
|||||||
@@ -4,56 +4,68 @@
|
|||||||
|
|
||||||
import { useState, useEffect } from "react";
|
import { useState, useEffect } from "react";
|
||||||
import { CheckCircle, Circle, Clock, AlertCircle } from "lucide-react";
|
import { CheckCircle, Circle, Clock, AlertCircle } from "lucide-react";
|
||||||
import { TaskPriority, TaskStatus, type WidgetProps, type Task } from "@mosaic/shared";
|
import type { WidgetProps } from "@mosaic/shared";
|
||||||
import { fetchTasks } from "@/lib/api/tasks";
|
|
||||||
|
|
||||||
export function TasksWidget({ id: _id, config: _config }: WidgetProps): React.JSX.Element {
|
interface Task {
|
||||||
|
id: string;
|
||||||
|
title: string;
|
||||||
|
status: string;
|
||||||
|
priority: string;
|
||||||
|
dueDate?: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
// eslint-disable-next-line no-empty-pattern
|
||||||
|
export function TasksWidget({}: WidgetProps): React.JSX.Element {
|
||||||
const [tasks, setTasks] = useState<Task[]>([]);
|
const [tasks, setTasks] = useState<Task[]>([]);
|
||||||
const [isLoading, setIsLoading] = useState(true);
|
const [isLoading, setIsLoading] = useState(true);
|
||||||
|
|
||||||
|
// Mock data for now - will fetch from API later
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
let isMounted = true;
|
|
||||||
|
|
||||||
const loadTasks = async (): Promise<void> => {
|
|
||||||
setIsLoading(true);
|
setIsLoading(true);
|
||||||
try {
|
// Simulate API call
|
||||||
const data = await fetchTasks();
|
setTimeout(() => {
|
||||||
if (isMounted) {
|
setTasks([
|
||||||
setTasks(data);
|
{
|
||||||
}
|
id: "1",
|
||||||
} catch {
|
title: "Complete project documentation",
|
||||||
if (isMounted) {
|
status: "IN_PROGRESS",
|
||||||
setTasks([]);
|
priority: "HIGH",
|
||||||
}
|
dueDate: "2024-02-01",
|
||||||
} finally {
|
},
|
||||||
if (isMounted) {
|
{
|
||||||
|
id: "2",
|
||||||
|
title: "Review pull requests",
|
||||||
|
status: "NOT_STARTED",
|
||||||
|
priority: "MEDIUM",
|
||||||
|
dueDate: "2024-02-02",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: "3",
|
||||||
|
title: "Update dependencies",
|
||||||
|
status: "COMPLETED",
|
||||||
|
priority: "LOW",
|
||||||
|
dueDate: "2024-01-30",
|
||||||
|
},
|
||||||
|
]);
|
||||||
setIsLoading(false);
|
setIsLoading(false);
|
||||||
}
|
}, 500);
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
void loadTasks();
|
|
||||||
|
|
||||||
return (): void => {
|
|
||||||
isMounted = false;
|
|
||||||
};
|
|
||||||
}, []);
|
}, []);
|
||||||
|
|
||||||
const getPriorityIcon = (priority: TaskPriority): React.JSX.Element => {
|
const getPriorityIcon = (priority: string): React.JSX.Element => {
|
||||||
switch (priority) {
|
switch (priority) {
|
||||||
case TaskPriority.HIGH:
|
case "HIGH":
|
||||||
return <AlertCircle className="w-4 h-4 text-red-500" />;
|
return <AlertCircle className="w-4 h-4 text-red-500" />;
|
||||||
case TaskPriority.MEDIUM:
|
case "MEDIUM":
|
||||||
return <Clock className="w-4 h-4 text-yellow-500" />;
|
return <Clock className="w-4 h-4 text-yellow-500" />;
|
||||||
case TaskPriority.LOW:
|
case "LOW":
|
||||||
return <Circle className="w-4 h-4 text-gray-400" />;
|
return <Circle className="w-4 h-4 text-gray-400" />;
|
||||||
default:
|
default:
|
||||||
return <Circle className="w-4 h-4 text-gray-400" />;
|
return <Circle className="w-4 h-4 text-gray-400" />;
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
const getStatusIcon = (status: TaskStatus): React.JSX.Element => {
|
const getStatusIcon = (status: string): React.JSX.Element => {
|
||||||
return status === TaskStatus.COMPLETED ? (
|
return status === "COMPLETED" ? (
|
||||||
<CheckCircle className="w-4 h-4 text-green-500" />
|
<CheckCircle className="w-4 h-4 text-green-500" />
|
||||||
) : (
|
) : (
|
||||||
<Circle className="w-4 h-4 text-gray-400" />
|
<Circle className="w-4 h-4 text-gray-400" />
|
||||||
@@ -62,8 +74,8 @@ export function TasksWidget({ id: _id, config: _config }: WidgetProps): React.JS
|
|||||||
|
|
||||||
const stats = {
|
const stats = {
|
||||||
total: tasks.length,
|
total: tasks.length,
|
||||||
inProgress: tasks.filter((t) => t.status === TaskStatus.IN_PROGRESS).length,
|
inProgress: tasks.filter((t) => t.status === "IN_PROGRESS").length,
|
||||||
completed: tasks.filter((t) => t.status === TaskStatus.COMPLETED).length,
|
completed: tasks.filter((t) => t.status === "COMPLETED").length,
|
||||||
};
|
};
|
||||||
|
|
||||||
if (isLoading) {
|
if (isLoading) {
|
||||||
|
|||||||
@@ -1,58 +1,16 @@
|
|||||||
import { describe, it, expect, beforeEach, afterEach, vi } from "vitest";
|
import { describe, it, expect, beforeEach, afterEach, vi } from "vitest";
|
||||||
import { render, screen, waitFor } from "@testing-library/react";
|
import { act, render, screen } from "@testing-library/react";
|
||||||
import type { Event } from "@mosaic/shared";
|
|
||||||
import { CalendarWidget } from "../CalendarWidget";
|
import { CalendarWidget } from "../CalendarWidget";
|
||||||
import { fetchEvents } from "@/lib/api/events";
|
|
||||||
|
|
||||||
vi.mock("@/lib/api/events", () => ({
|
|
||||||
fetchEvents: vi.fn(),
|
|
||||||
}));
|
|
||||||
|
|
||||||
const mockEvents: Event[] = [
|
|
||||||
{
|
|
||||||
id: "event-1",
|
|
||||||
title: "API Planning",
|
|
||||||
description: null,
|
|
||||||
startTime: new Date("2026-02-01T09:00:00Z"),
|
|
||||||
endTime: new Date("2026-02-01T09:30:00Z"),
|
|
||||||
allDay: false,
|
|
||||||
location: "Zoom",
|
|
||||||
recurrence: null,
|
|
||||||
creatorId: "user-1",
|
|
||||||
workspaceId: "workspace-1",
|
|
||||||
projectId: null,
|
|
||||||
metadata: {},
|
|
||||||
createdAt: new Date("2026-01-30T09:00:00Z"),
|
|
||||||
updatedAt: new Date("2026-01-30T09:00:00Z"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
id: "event-2",
|
|
||||||
title: "API Review",
|
|
||||||
description: null,
|
|
||||||
startTime: new Date("2026-02-02T10:00:00Z"),
|
|
||||||
endTime: new Date("2026-02-02T11:00:00Z"),
|
|
||||||
allDay: false,
|
|
||||||
location: "Room 1",
|
|
||||||
recurrence: null,
|
|
||||||
creatorId: "user-1",
|
|
||||||
workspaceId: "workspace-1",
|
|
||||||
projectId: null,
|
|
||||||
metadata: {},
|
|
||||||
createdAt: new Date("2026-01-30T09:00:00Z"),
|
|
||||||
updatedAt: new Date("2026-01-30T09:00:00Z"),
|
|
||||||
},
|
|
||||||
];
|
|
||||||
|
|
||||||
async function finishWidgetLoad(): Promise<void> {
|
async function finishWidgetLoad(): Promise<void> {
|
||||||
await waitFor(() => {
|
await act(async () => {
|
||||||
expect(screen.queryByText("Loading events...")).not.toBeInTheDocument();
|
await vi.advanceTimersByTimeAsync(500);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
describe("CalendarWidget", (): void => {
|
describe("CalendarWidget", (): void => {
|
||||||
beforeEach((): void => {
|
beforeEach((): void => {
|
||||||
vi.clearAllMocks();
|
vi.useFakeTimers();
|
||||||
vi.mocked(fetchEvents).mockResolvedValue(mockEvents);
|
|
||||||
vi.setSystemTime(new Date("2026-02-01T08:00:00Z"));
|
vi.setSystemTime(new Date("2026-02-01T08:00:00Z"));
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -66,15 +24,15 @@ describe("CalendarWidget", (): void => {
|
|||||||
expect(screen.getByText("Loading events...")).toBeInTheDocument();
|
expect(screen.getByText("Loading events...")).toBeInTheDocument();
|
||||||
});
|
});
|
||||||
|
|
||||||
it("fetches and renders upcoming events after loading", async (): Promise<void> => {
|
it("renders upcoming events after loading", async (): Promise<void> => {
|
||||||
render(<CalendarWidget id="calendar-1" />);
|
render(<CalendarWidget id="calendar-1" />);
|
||||||
|
|
||||||
await finishWidgetLoad();
|
await finishWidgetLoad();
|
||||||
|
|
||||||
expect(fetchEvents).toHaveBeenCalledTimes(1);
|
|
||||||
expect(screen.getByText("Upcoming Events")).toBeInTheDocument();
|
expect(screen.getByText("Upcoming Events")).toBeInTheDocument();
|
||||||
expect(screen.getByText("API Planning")).toBeInTheDocument();
|
expect(screen.getByText("Team Standup")).toBeInTheDocument();
|
||||||
expect(screen.getByText("API Review")).toBeInTheDocument();
|
expect(screen.getByText("Project Review")).toBeInTheDocument();
|
||||||
|
expect(screen.getByText("Sprint Planning")).toBeInTheDocument();
|
||||||
});
|
});
|
||||||
|
|
||||||
it("shows relative day labels", async (): Promise<void> => {
|
it("shows relative day labels", async (): Promise<void> => {
|
||||||
@@ -92,15 +50,6 @@ describe("CalendarWidget", (): void => {
|
|||||||
await finishWidgetLoad();
|
await finishWidgetLoad();
|
||||||
|
|
||||||
expect(screen.getByText("Zoom")).toBeInTheDocument();
|
expect(screen.getByText("Zoom")).toBeInTheDocument();
|
||||||
expect(screen.getByText("Room 1")).toBeInTheDocument();
|
expect(screen.getByText("Conference Room A")).toBeInTheDocument();
|
||||||
});
|
|
||||||
|
|
||||||
it("shows empty state when no events are returned", async (): Promise<void> => {
|
|
||||||
vi.mocked(fetchEvents).mockResolvedValueOnce([]);
|
|
||||||
|
|
||||||
render(<CalendarWidget id="calendar-1" />);
|
|
||||||
await finishWidgetLoad();
|
|
||||||
|
|
||||||
expect(screen.getByText("No upcoming events")).toBeInTheDocument();
|
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -1,80 +1,20 @@
|
|||||||
import { describe, it, expect, beforeEach, vi } from "vitest";
|
import { describe, it, expect, beforeEach, afterEach, vi } from "vitest";
|
||||||
import { render, screen, waitFor } from "@testing-library/react";
|
import { act, render, screen } from "@testing-library/react";
|
||||||
import { TaskStatus, TaskPriority, type Task } from "@mosaic/shared";
|
|
||||||
import { TasksWidget } from "../TasksWidget";
|
import { TasksWidget } from "../TasksWidget";
|
||||||
import { fetchTasks } from "@/lib/api/tasks";
|
|
||||||
|
|
||||||
vi.mock("@/lib/api/tasks", () => ({
|
|
||||||
fetchTasks: vi.fn(),
|
|
||||||
}));
|
|
||||||
|
|
||||||
const mockTasks: Task[] = [
|
|
||||||
{
|
|
||||||
id: "task-1",
|
|
||||||
title: "API task one",
|
|
||||||
description: null,
|
|
||||||
status: TaskStatus.IN_PROGRESS,
|
|
||||||
priority: TaskPriority.HIGH,
|
|
||||||
dueDate: new Date("2026-02-03T09:00:00Z"),
|
|
||||||
creatorId: "user-1",
|
|
||||||
assigneeId: "user-1",
|
|
||||||
workspaceId: "workspace-1",
|
|
||||||
projectId: null,
|
|
||||||
parentId: null,
|
|
||||||
sortOrder: 0,
|
|
||||||
metadata: {},
|
|
||||||
completedAt: null,
|
|
||||||
createdAt: new Date("2026-02-01T09:00:00Z"),
|
|
||||||
updatedAt: new Date("2026-02-01T09:00:00Z"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
id: "task-2",
|
|
||||||
title: "API task two",
|
|
||||||
description: null,
|
|
||||||
status: TaskStatus.NOT_STARTED,
|
|
||||||
priority: TaskPriority.MEDIUM,
|
|
||||||
dueDate: new Date("2026-02-04T09:00:00Z"),
|
|
||||||
creatorId: "user-1",
|
|
||||||
assigneeId: "user-1",
|
|
||||||
workspaceId: "workspace-1",
|
|
||||||
projectId: null,
|
|
||||||
parentId: null,
|
|
||||||
sortOrder: 1,
|
|
||||||
metadata: {},
|
|
||||||
completedAt: null,
|
|
||||||
createdAt: new Date("2026-02-01T09:00:00Z"),
|
|
||||||
updatedAt: new Date("2026-02-01T09:00:00Z"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
id: "task-3",
|
|
||||||
title: "API task three",
|
|
||||||
description: null,
|
|
||||||
status: TaskStatus.COMPLETED,
|
|
||||||
priority: TaskPriority.LOW,
|
|
||||||
dueDate: new Date("2026-02-05T09:00:00Z"),
|
|
||||||
creatorId: "user-1",
|
|
||||||
assigneeId: "user-1",
|
|
||||||
workspaceId: "workspace-1",
|
|
||||||
projectId: null,
|
|
||||||
parentId: null,
|
|
||||||
sortOrder: 2,
|
|
||||||
metadata: {},
|
|
||||||
completedAt: new Date("2026-02-02T09:00:00Z"),
|
|
||||||
createdAt: new Date("2026-02-01T09:00:00Z"),
|
|
||||||
updatedAt: new Date("2026-02-02T09:00:00Z"),
|
|
||||||
},
|
|
||||||
];
|
|
||||||
|
|
||||||
async function finishWidgetLoad(): Promise<void> {
|
async function finishWidgetLoad(): Promise<void> {
|
||||||
await waitFor(() => {
|
await act(async () => {
|
||||||
expect(screen.queryByText("Loading tasks...")).not.toBeInTheDocument();
|
await vi.advanceTimersByTimeAsync(500);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
describe("TasksWidget", (): void => {
|
describe("TasksWidget", (): void => {
|
||||||
beforeEach((): void => {
|
beforeEach((): void => {
|
||||||
vi.clearAllMocks();
|
vi.useFakeTimers();
|
||||||
vi.mocked(fetchTasks).mockResolvedValue(mockTasks);
|
});
|
||||||
|
|
||||||
|
afterEach((): void => {
|
||||||
|
vi.useRealTimers();
|
||||||
});
|
});
|
||||||
|
|
||||||
it("renders loading state initially", (): void => {
|
it("renders loading state initially", (): void => {
|
||||||
@@ -83,26 +23,25 @@ describe("TasksWidget", (): void => {
|
|||||||
expect(screen.getByText("Loading tasks...")).toBeInTheDocument();
|
expect(screen.getByText("Loading tasks...")).toBeInTheDocument();
|
||||||
});
|
});
|
||||||
|
|
||||||
it("fetches tasks and renders summary stats", async (): Promise<void> => {
|
it("renders default summary stats", async (): Promise<void> => {
|
||||||
render(<TasksWidget id="tasks-1" />);
|
render(<TasksWidget id="tasks-1" />);
|
||||||
|
|
||||||
await finishWidgetLoad();
|
await finishWidgetLoad();
|
||||||
|
|
||||||
expect(fetchTasks).toHaveBeenCalledTimes(1);
|
|
||||||
expect(screen.getByText("Total")).toBeInTheDocument();
|
expect(screen.getByText("Total")).toBeInTheDocument();
|
||||||
expect(screen.getByText("In Progress")).toBeInTheDocument();
|
expect(screen.getByText("In Progress")).toBeInTheDocument();
|
||||||
expect(screen.getByText("Done")).toBeInTheDocument();
|
expect(screen.getByText("Done")).toBeInTheDocument();
|
||||||
expect(screen.getByText("3")).toBeInTheDocument();
|
expect(screen.getByText("3")).toBeInTheDocument();
|
||||||
});
|
});
|
||||||
|
|
||||||
it("renders task rows from API response", async (): Promise<void> => {
|
it("renders default task rows", async (): Promise<void> => {
|
||||||
render(<TasksWidget id="tasks-1" />);
|
render(<TasksWidget id="tasks-1" />);
|
||||||
|
|
||||||
await finishWidgetLoad();
|
await finishWidgetLoad();
|
||||||
|
|
||||||
expect(screen.getByText("API task one")).toBeInTheDocument();
|
expect(screen.getByText("Complete project documentation")).toBeInTheDocument();
|
||||||
expect(screen.getByText("API task two")).toBeInTheDocument();
|
expect(screen.getByText("Review pull requests")).toBeInTheDocument();
|
||||||
expect(screen.getByText("API task three")).toBeInTheDocument();
|
expect(screen.getByText("Update dependencies")).toBeInTheDocument();
|
||||||
});
|
});
|
||||||
|
|
||||||
it("shows due date labels for each task", async (): Promise<void> => {
|
it("shows due date labels for each task", async (): Promise<void> => {
|
||||||
@@ -112,13 +51,4 @@ describe("TasksWidget", (): void => {
|
|||||||
|
|
||||||
expect(screen.getAllByText(/Due:/).length).toBe(3);
|
expect(screen.getAllByText(/Due:/).length).toBe(3);
|
||||||
});
|
});
|
||||||
|
|
||||||
it("shows empty state when API returns no tasks", async (): Promise<void> => {
|
|
||||||
vi.mocked(fetchTasks).mockResolvedValueOnce([]);
|
|
||||||
|
|
||||||
render(<TasksWidget id="tasks-1" />);
|
|
||||||
await finishWidgetLoad();
|
|
||||||
|
|
||||||
expect(screen.getByText("No tasks yet")).toBeInTheDocument();
|
|
||||||
});
|
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -7,7 +7,6 @@ import { useState, useCallback, useRef } from "react";
|
|||||||
import {
|
import {
|
||||||
sendChatMessage,
|
sendChatMessage,
|
||||||
streamChatMessage,
|
streamChatMessage,
|
||||||
streamGuestChat,
|
|
||||||
type ChatMessage as ApiChatMessage,
|
type ChatMessage as ApiChatMessage,
|
||||||
} from "@/lib/api/chat";
|
} from "@/lib/api/chat";
|
||||||
import { createConversation, updateConversation, getIdea, type Idea } from "@/lib/api/ideas";
|
import { createConversation, updateConversation, getIdea, type Idea } from "@/lib/api/ideas";
|
||||||
@@ -279,68 +278,6 @@ export function useChat(options: UseChatOptions = {}): UseChatReturn {
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Streaming failed - check if auth error, try guest mode
|
|
||||||
const isAuthError =
|
|
||||||
err instanceof Error &&
|
|
||||||
(err.message.includes("403") ||
|
|
||||||
err.message.includes("401") ||
|
|
||||||
err.message.includes("auth") ||
|
|
||||||
err.message.includes("Forbidden"));
|
|
||||||
|
|
||||||
if (isAuthError) {
|
|
||||||
console.warn("Auth failed, trying guest chat mode");
|
|
||||||
|
|
||||||
// Try guest chat streaming
|
|
||||||
try {
|
|
||||||
await new Promise<void>((guestResolve, guestReject) => {
|
|
||||||
let hasReceivedData = false;
|
|
||||||
|
|
||||||
streamGuestChat(
|
|
||||||
request,
|
|
||||||
(chunk: string) => {
|
|
||||||
if (!hasReceivedData) {
|
|
||||||
hasReceivedData = true;
|
|
||||||
setIsLoading(false);
|
|
||||||
setIsStreaming(true);
|
|
||||||
setMessages((prev) => {
|
|
||||||
const updated = [...prev, { ...placeholderMessage }];
|
|
||||||
messagesRef.current = updated;
|
|
||||||
return updated;
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
setMessages((prev) => {
|
|
||||||
const updated = prev.map((msg) =>
|
|
||||||
msg.id === assistantMessageId ? { ...msg, content: msg.content + chunk } : msg
|
|
||||||
);
|
|
||||||
messagesRef.current = updated;
|
|
||||||
return updated;
|
|
||||||
});
|
|
||||||
},
|
|
||||||
() => {
|
|
||||||
streamingSucceeded = true;
|
|
||||||
setIsStreaming(false);
|
|
||||||
guestResolve();
|
|
||||||
},
|
|
||||||
(guestErr: Error) => {
|
|
||||||
guestReject(guestErr);
|
|
||||||
},
|
|
||||||
controller.signal
|
|
||||||
);
|
|
||||||
});
|
|
||||||
} catch (guestErr: unknown) {
|
|
||||||
// Guest also failed
|
|
||||||
setMessages((prev) => {
|
|
||||||
const withoutPlaceholder = prev.filter((m) => m.id !== assistantMessageId);
|
|
||||||
messagesRef.current = withoutPlaceholder;
|
|
||||||
return withoutPlaceholder;
|
|
||||||
});
|
|
||||||
const errorMsg = guestErr instanceof Error ? guestErr.message : "Chat unavailable";
|
|
||||||
setError(`Unable to connect to chat: ${errorMsg}`);
|
|
||||||
setIsLoading(false);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// Streaming failed — fall back to non-streaming
|
// Streaming failed — fall back to non-streaming
|
||||||
console.warn("Streaming failed, falling back to non-streaming", {
|
console.warn("Streaming failed, falling back to non-streaming", {
|
||||||
error: err instanceof Error ? err : new Error(String(err)),
|
error: err instanceof Error ? err : new Error(String(err)),
|
||||||
@@ -405,7 +342,6 @@ export function useChat(options: UseChatOptions = {}): UseChatReturn {
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
setIsLoading(false);
|
setIsLoading(false);
|
||||||
|
|
||||||
|
|||||||
@@ -92,141 +92,6 @@ async function ensureCsrfTokenForStream(): Promise<string> {
|
|||||||
return fetchCsrfToken();
|
return fetchCsrfToken();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Stream a guest chat message (no authentication required).
|
|
||||||
* Uses /api/chat/guest endpoint with shared LLM configuration.
|
|
||||||
*
|
|
||||||
* @param request - Chat request
|
|
||||||
* @param onChunk - Called with each token string as it arrives
|
|
||||||
* @param onComplete - Called when the stream finishes successfully
|
|
||||||
* @param onError - Called if the stream encounters an error
|
|
||||||
* @param signal - Optional AbortSignal for cancellation
|
|
||||||
*/
|
|
||||||
export function streamGuestChat(
|
|
||||||
request: ChatRequest,
|
|
||||||
onChunk: (chunk: string) => void,
|
|
||||||
onComplete: () => void,
|
|
||||||
onError: (error: Error) => void,
|
|
||||||
signal?: AbortSignal
|
|
||||||
): void {
|
|
||||||
void (async (): Promise<void> => {
|
|
||||||
try {
|
|
||||||
const response = await fetch(`${API_BASE_URL}/api/chat/guest`, {
|
|
||||||
method: "POST",
|
|
||||||
headers: {
|
|
||||||
"Content-Type": "application/json",
|
|
||||||
},
|
|
||||||
credentials: "include",
|
|
||||||
body: JSON.stringify({ messages: request.messages, stream: true }),
|
|
||||||
signal: signal ?? null,
|
|
||||||
});
|
|
||||||
|
|
||||||
if (!response.ok) {
|
|
||||||
const errorText = await response.text().catch(() => response.statusText);
|
|
||||||
throw new Error(`Guest chat failed: ${errorText}`);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!response.body) {
|
|
||||||
throw new Error("Response body is not readable");
|
|
||||||
}
|
|
||||||
|
|
||||||
const reader = response.body.getReader();
|
|
||||||
const decoder = new TextDecoder("utf-8");
|
|
||||||
let buffer = "";
|
|
||||||
|
|
||||||
let readerDone = false;
|
|
||||||
while (!readerDone) {
|
|
||||||
const { done, value } = await reader.read();
|
|
||||||
readerDone = done;
|
|
||||||
if (done) {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
buffer += decoder.decode(value, { stream: true });
|
|
||||||
|
|
||||||
// SSE messages are separated by double newlines
|
|
||||||
const parts = buffer.split("\n\n");
|
|
||||||
buffer = parts.pop() ?? "";
|
|
||||||
|
|
||||||
for (const part of parts) {
|
|
||||||
const trimmed = part.trim();
|
|
||||||
if (!trimmed) continue;
|
|
||||||
|
|
||||||
// Handle event: error format
|
|
||||||
const eventMatch = /^event:\s*(\S+)\n/i.exec(trimmed);
|
|
||||||
const dataMatch = /^data:\s*(.+)$/im.exec(trimmed);
|
|
||||||
|
|
||||||
if (eventMatch?.[1] === "error" && dataMatch?.[1]) {
|
|
||||||
try {
|
|
||||||
const errorData = JSON.parse(dataMatch[1].trim()) as {
|
|
||||||
error?: string;
|
|
||||||
};
|
|
||||||
throw new Error(errorData.error ?? "Stream error occurred");
|
|
||||||
} catch (parseErr) {
|
|
||||||
if (parseErr instanceof SyntaxError) {
|
|
||||||
throw new Error("Stream error occurred");
|
|
||||||
}
|
|
||||||
throw parseErr;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Standard SSE format: data: {...}
|
|
||||||
for (const line of trimmed.split("\n")) {
|
|
||||||
if (!line.startsWith("data: ")) continue;
|
|
||||||
|
|
||||||
const data = line.slice("data: ".length).trim();
|
|
||||||
|
|
||||||
if (data === "[DONE]") {
|
|
||||||
onComplete();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
try {
|
|
||||||
const parsed: unknown = JSON.parse(data);
|
|
||||||
|
|
||||||
// Handle OpenAI format
|
|
||||||
const openAiChunk = parsed as OpenAiSseChunk;
|
|
||||||
if (openAiChunk.choices?.[0]?.delta?.content) {
|
|
||||||
onChunk(openAiChunk.choices[0].delta.content);
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Handle simple token format
|
|
||||||
const simpleChunk = parsed as SimpleTokenChunk;
|
|
||||||
if (simpleChunk.token) {
|
|
||||||
onChunk(simpleChunk.token);
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (simpleChunk.done === true) {
|
|
||||||
onComplete();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
const error = openAiChunk.error ?? simpleChunk.error;
|
|
||||||
if (error) {
|
|
||||||
throw new Error(error);
|
|
||||||
}
|
|
||||||
} catch (parseErr) {
|
|
||||||
if (parseErr instanceof SyntaxError) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
throw parseErr;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
onComplete();
|
|
||||||
} catch (err: unknown) {
|
|
||||||
if (err instanceof DOMException && err.name === "AbortError") {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
onError(err instanceof Error ? err : new Error(String(err)));
|
|
||||||
}
|
|
||||||
})();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Stream a chat message from the LLM using SSE over fetch.
|
* Stream a chat message from the LLM using SSE over fetch.
|
||||||
*
|
*
|
||||||
|
|||||||
@@ -1,53 +0,0 @@
|
|||||||
import { describe, it, expect, beforeEach, afterEach, vi } from "vitest";
|
|
||||||
import { fetchUsageSummary } from "./telemetry";
|
|
||||||
|
|
||||||
vi.mock("./client", () => ({
|
|
||||||
apiGet: vi.fn(),
|
|
||||||
}));
|
|
||||||
|
|
||||||
const { apiGet } = await import("./client");
|
|
||||||
|
|
||||||
describe("Telemetry API Client", (): void => {
|
|
||||||
beforeEach((): void => {
|
|
||||||
vi.clearAllMocks();
|
|
||||||
vi.useFakeTimers();
|
|
||||||
vi.setSystemTime(new Date("2026-03-02T12:00:00Z"));
|
|
||||||
});
|
|
||||||
|
|
||||||
afterEach((): void => {
|
|
||||||
vi.useRealTimers();
|
|
||||||
});
|
|
||||||
|
|
||||||
it("fetches usage summary from llm usage analytics endpoint", async (): Promise<void> => {
|
|
||||||
vi.mocked(apiGet).mockResolvedValueOnce({
|
|
||||||
data: {
|
|
||||||
totalCalls: 47,
|
|
||||||
totalPromptTokens: 120000,
|
|
||||||
totalCompletionTokens: 125800,
|
|
||||||
totalTokens: 245800,
|
|
||||||
totalCostCents: 342,
|
|
||||||
averageDurationMs: 3200,
|
|
||||||
byProvider: [],
|
|
||||||
byModel: [],
|
|
||||||
byTaskType: [],
|
|
||||||
},
|
|
||||||
});
|
|
||||||
|
|
||||||
const result = await fetchUsageSummary("30d");
|
|
||||||
|
|
||||||
const calledEndpoint = vi.mocked(apiGet).mock.calls[0]?.[0];
|
|
||||||
expect(calledEndpoint).toMatch(/^\/api\/llm-usage\/analytics\?/);
|
|
||||||
|
|
||||||
const queryString = calledEndpoint?.split("?")[1] ?? "";
|
|
||||||
const params = new URLSearchParams(queryString);
|
|
||||||
expect(params.get("startDate")).toBeTruthy();
|
|
||||||
expect(params.get("endDate")).toBeTruthy();
|
|
||||||
|
|
||||||
expect(result).toEqual({
|
|
||||||
totalTokens: 245800,
|
|
||||||
totalCost: 3.42,
|
|
||||||
taskCount: 47,
|
|
||||||
avgQualityGatePassRate: 0,
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
@@ -1,6 +1,10 @@
|
|||||||
/**
|
/**
|
||||||
* Telemetry API Client
|
* Telemetry API Client
|
||||||
* Handles telemetry data fetching for the usage dashboard.
|
* Handles telemetry data fetching for the usage dashboard.
|
||||||
|
*
|
||||||
|
* NOTE: Currently returns mock/placeholder data since the telemetry API
|
||||||
|
* aggregation endpoints don't exist yet. The important thing is the UI structure.
|
||||||
|
* When the backend endpoints are ready, replace mock calls with real apiGet() calls.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { apiGet, type ApiResponse } from "./client";
|
import { apiGet, type ApiResponse } from "./client";
|
||||||
@@ -56,84 +60,65 @@ export interface EstimateResponse {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
interface ProviderUsageAnalyticsItem {
|
// ─── Mock Data Generators ────────────────────────────────────────────
|
||||||
provider: string;
|
|
||||||
calls: number;
|
|
||||||
promptTokens: number;
|
|
||||||
completionTokens: number;
|
|
||||||
totalTokens: number;
|
|
||||||
costCents: number;
|
|
||||||
averageDurationMs: number;
|
|
||||||
}
|
|
||||||
|
|
||||||
interface ModelUsageAnalyticsItem {
|
function generateDateRange(range: TimeRange): string[] {
|
||||||
model: string;
|
const days = range === "7d" ? 7 : range === "30d" ? 30 : 90;
|
||||||
calls: number;
|
const dates: string[] = [];
|
||||||
promptTokens: number;
|
const now = new Date();
|
||||||
completionTokens: number;
|
|
||||||
totalTokens: number;
|
|
||||||
costCents: number;
|
|
||||||
averageDurationMs: number;
|
|
||||||
}
|
|
||||||
|
|
||||||
interface TaskTypeUsageAnalyticsItem {
|
for (let i = days - 1; i >= 0; i--) {
|
||||||
taskType: string;
|
const d = new Date(now);
|
||||||
calls: number;
|
d.setDate(d.getDate() - i);
|
||||||
promptTokens: number;
|
dates.push(d.toISOString().split("T")[0] ?? "");
|
||||||
completionTokens: number;
|
|
||||||
totalTokens: number;
|
|
||||||
costCents: number;
|
|
||||||
averageDurationMs: number;
|
|
||||||
}
|
|
||||||
|
|
||||||
interface UsageAnalyticsResponse {
|
|
||||||
totalCalls: number;
|
|
||||||
totalPromptTokens: number;
|
|
||||||
totalCompletionTokens: number;
|
|
||||||
totalTokens: number;
|
|
||||||
totalCostCents: number;
|
|
||||||
averageDurationMs: number;
|
|
||||||
byProvider: ProviderUsageAnalyticsItem[];
|
|
||||||
byModel: ModelUsageAnalyticsItem[];
|
|
||||||
byTaskType: TaskTypeUsageAnalyticsItem[];
|
|
||||||
}
|
|
||||||
|
|
||||||
const TASK_OUTCOME_COLORS = ["#6EBF8B", "#F5C862", "#94A3B8", "#C4A5DE", "#7AA2F7"];
|
|
||||||
const DAYS_BY_RANGE: Record<TimeRange, number> = {
|
|
||||||
"7d": 7,
|
|
||||||
"30d": 30,
|
|
||||||
"90d": 90,
|
|
||||||
};
|
|
||||||
const analyticsRequestCache = new Map<TimeRange, Promise<UsageAnalyticsResponse>>();
|
|
||||||
|
|
||||||
function buildAnalyticsEndpoint(timeRange: TimeRange): string {
|
|
||||||
const endDate = new Date();
|
|
||||||
const startDate = new Date(endDate);
|
|
||||||
startDate.setDate(startDate.getDate() - (DAYS_BY_RANGE[timeRange] - 1));
|
|
||||||
startDate.setHours(0, 0, 0, 0);
|
|
||||||
|
|
||||||
const query = new URLSearchParams({
|
|
||||||
startDate: startDate.toISOString(),
|
|
||||||
endDate: endDate.toISOString(),
|
|
||||||
}).toString();
|
|
||||||
|
|
||||||
return `/api/llm-usage/analytics?${query}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
async function fetchUsageAnalytics(timeRange: TimeRange): Promise<UsageAnalyticsResponse> {
|
|
||||||
const cachedRequest = analyticsRequestCache.get(timeRange);
|
|
||||||
if (cachedRequest) {
|
|
||||||
return cachedRequest;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const request = apiGet<ApiResponse<UsageAnalyticsResponse>>(buildAnalyticsEndpoint(timeRange))
|
return dates;
|
||||||
.then((response) => response.data)
|
}
|
||||||
.finally(() => {
|
|
||||||
analyticsRequestCache.delete(timeRange);
|
|
||||||
});
|
|
||||||
|
|
||||||
analyticsRequestCache.set(timeRange, request);
|
function generateMockTokenUsage(range: TimeRange): TokenUsagePoint[] {
|
||||||
return request;
|
const dates = generateDateRange(range);
|
||||||
|
|
||||||
|
return dates.map((date) => {
|
||||||
|
const baseInput = 8000 + Math.floor(Math.random() * 12000);
|
||||||
|
const baseOutput = 3000 + Math.floor(Math.random() * 7000);
|
||||||
|
return {
|
||||||
|
date,
|
||||||
|
inputTokens: baseInput,
|
||||||
|
outputTokens: baseOutput,
|
||||||
|
totalTokens: baseInput + baseOutput,
|
||||||
|
};
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
function generateMockSummary(range: TimeRange): UsageSummary {
|
||||||
|
const multiplier = range === "7d" ? 1 : range === "30d" ? 4 : 12;
|
||||||
|
return {
|
||||||
|
totalTokens: 245_800 * multiplier,
|
||||||
|
totalCost: 3.42 * multiplier,
|
||||||
|
taskCount: 47 * multiplier,
|
||||||
|
avgQualityGatePassRate: 0.87,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
function generateMockCostBreakdown(): CostBreakdownItem[] {
|
||||||
|
return [
|
||||||
|
{ model: "claude-sonnet-4-5", provider: "anthropic", cost: 18.5, taskCount: 124 },
|
||||||
|
{ model: "gpt-4o", provider: "openai", cost: 12.3, taskCount: 89 },
|
||||||
|
{ model: "claude-haiku-3.5", provider: "anthropic", cost: 4.2, taskCount: 156 },
|
||||||
|
{ model: "llama-3.3-70b", provider: "ollama", cost: 0, taskCount: 67 },
|
||||||
|
{ model: "gemini-2.0-flash", provider: "google", cost: 2.8, taskCount: 42 },
|
||||||
|
];
|
||||||
|
}
|
||||||
|
|
||||||
|
// PDA-friendly colors: calm, no aggressive reds
|
||||||
|
function generateMockTaskOutcomes(): TaskOutcomeItem[] {
|
||||||
|
return [
|
||||||
|
{ outcome: "Success", count: 312, color: "#6EBF8B" },
|
||||||
|
{ outcome: "Partial", count: 48, color: "#F5C862" },
|
||||||
|
{ outcome: "Timeout", count: 18, color: "#94A3B8" },
|
||||||
|
{ outcome: "Incomplete", count: 22, color: "#C4A5DE" },
|
||||||
|
];
|
||||||
}
|
}
|
||||||
|
|
||||||
// ─── API Functions ───────────────────────────────────────────────────
|
// ─── API Functions ───────────────────────────────────────────────────
|
||||||
@@ -142,54 +127,47 @@ async function fetchUsageAnalytics(timeRange: TimeRange): Promise<UsageAnalytics
|
|||||||
* Fetch usage summary data (total tokens, cost, task count, quality rate)
|
* Fetch usage summary data (total tokens, cost, task count, quality rate)
|
||||||
*/
|
*/
|
||||||
export async function fetchUsageSummary(timeRange: TimeRange): Promise<UsageSummary> {
|
export async function fetchUsageSummary(timeRange: TimeRange): Promise<UsageSummary> {
|
||||||
const analytics = await fetchUsageAnalytics(timeRange);
|
// TODO: Replace with real API call when backend aggregation endpoints are ready
|
||||||
|
// const response = await apiGet<ApiResponse<UsageSummary>>(`/api/telemetry/summary?range=${timeRange}`);
|
||||||
return {
|
// return response.data;
|
||||||
totalTokens: analytics.totalTokens,
|
void apiGet; // suppress unused import warning in the meantime
|
||||||
totalCost: analytics.totalCostCents / 100,
|
await new Promise((resolve) => setTimeout(resolve, 200));
|
||||||
taskCount: analytics.totalCalls,
|
return generateMockSummary(timeRange);
|
||||||
avgQualityGatePassRate: 0,
|
|
||||||
};
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Fetch token usage time series for charts
|
* Fetch token usage time series for charts
|
||||||
*/
|
*/
|
||||||
export function fetchTokenUsage(timeRange: TimeRange): Promise<TokenUsagePoint[]> {
|
export async function fetchTokenUsage(timeRange: TimeRange): Promise<TokenUsagePoint[]> {
|
||||||
void timeRange;
|
// TODO: Replace with real API call
|
||||||
return Promise.resolve([]);
|
// const response = await apiGet<ApiResponse<TokenUsagePoint[]>>(`/api/telemetry/tokens?range=${timeRange}`);
|
||||||
|
// return response.data;
|
||||||
|
await new Promise((resolve) => setTimeout(resolve, 250));
|
||||||
|
return generateMockTokenUsage(timeRange);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Fetch cost breakdown by model
|
* Fetch cost breakdown by model
|
||||||
*/
|
*/
|
||||||
export async function fetchCostBreakdown(timeRange: TimeRange): Promise<CostBreakdownItem[]> {
|
export async function fetchCostBreakdown(timeRange: TimeRange): Promise<CostBreakdownItem[]> {
|
||||||
const analytics = await fetchUsageAnalytics(timeRange);
|
// TODO: Replace with real API call
|
||||||
|
// const response = await apiGet<ApiResponse<CostBreakdownItem[]>>(`/api/telemetry/costs?range=${timeRange}`);
|
||||||
return analytics.byModel
|
// return response.data;
|
||||||
.filter((item) => item.calls > 0)
|
await new Promise((resolve) => setTimeout(resolve, 200));
|
||||||
.sort((a, b) => b.costCents - a.costCents)
|
void timeRange;
|
||||||
.map((item) => ({
|
return generateMockCostBreakdown();
|
||||||
model: item.model,
|
|
||||||
provider: "unknown",
|
|
||||||
cost: item.costCents / 100,
|
|
||||||
taskCount: item.calls,
|
|
||||||
}));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Fetch task outcome distribution
|
* Fetch task outcome distribution
|
||||||
*/
|
*/
|
||||||
export async function fetchTaskOutcomes(timeRange: TimeRange): Promise<TaskOutcomeItem[]> {
|
export async function fetchTaskOutcomes(timeRange: TimeRange): Promise<TaskOutcomeItem[]> {
|
||||||
const analytics = await fetchUsageAnalytics(timeRange);
|
// TODO: Replace with real API call
|
||||||
|
// const response = await apiGet<ApiResponse<TaskOutcomeItem[]>>(`/api/telemetry/outcomes?range=${timeRange}`);
|
||||||
return analytics.byTaskType
|
// return response.data;
|
||||||
.filter((item) => item.calls > 0)
|
await new Promise((resolve) => setTimeout(resolve, 150));
|
||||||
.map((item, index) => ({
|
void timeRange;
|
||||||
outcome: item.taskType,
|
return generateMockTaskOutcomes();
|
||||||
count: item.calls,
|
|
||||||
color: TASK_OUTCOME_COLORS[index % TASK_OUTCOME_COLORS.length] ?? "#94A3B8",
|
|
||||||
}));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|||||||
@@ -9,8 +9,6 @@
|
|||||||
# - OpenBao: Standalone container (see docker-compose.openbao.yml)
|
# - OpenBao: Standalone container (see docker-compose.openbao.yml)
|
||||||
# - Authentik: External OIDC provider
|
# - Authentik: External OIDC provider
|
||||||
# - Ollama: External AI inference
|
# - Ollama: External AI inference
|
||||||
# - PostgreSQL: Provided by the openbrain stack (openbrain_brain-db)
|
|
||||||
# Deploy openbrain stack before this stack.
|
|
||||||
#
|
#
|
||||||
# Usage (Portainer):
|
# Usage (Portainer):
|
||||||
# 1. Stacks -> Add Stack -> Upload or paste
|
# 1. Stacks -> Add Stack -> Upload or paste
|
||||||
@@ -38,75 +36,37 @@
|
|||||||
# Required vars use plain ${VAR} — the app validates at startup.
|
# Required vars use plain ${VAR} — the app validates at startup.
|
||||||
#
|
#
|
||||||
# ==============================================
|
# ==============================================
|
||||||
# DATABASE (openbrain_brain-db — external)
|
|
||||||
# ==============================================
|
|
||||||
#
|
|
||||||
# This stack uses the PostgreSQL instance from the openbrain stack.
|
|
||||||
# The openbrain stack must be deployed first and its brain-internal
|
|
||||||
# overlay network must exist.
|
|
||||||
#
|
|
||||||
# Required env vars for DB access:
|
|
||||||
# BRAIN_DB_ADMIN_USER — openbrain superuser (default: openbrain)
|
|
||||||
# BRAIN_DB_ADMIN_PASSWORD — openbrain superuser password
|
|
||||||
# (must match openbrain stack POSTGRES_PASSWORD)
|
|
||||||
# POSTGRES_USER — mosaic application DB user (created by mosaic-db-init)
|
|
||||||
# POSTGRES_PASSWORD — mosaic application DB password
|
|
||||||
# POSTGRES_DB — mosaic application database name (default: mosaic)
|
|
||||||
#
|
|
||||||
# ==============================================
|
|
||||||
|
|
||||||
services:
|
services:
|
||||||
# ============================================
|
# ============================================
|
||||||
# DATABASE INIT
|
# CORE INFRASTRUCTURE
|
||||||
# ============================================
|
# ============================================
|
||||||
|
|
||||||
# ======================
|
# ======================
|
||||||
# Mosaic Database Init
|
# PostgreSQL Database
|
||||||
# ======================
|
# ======================
|
||||||
# Creates the mosaic application user and database in the shared
|
postgres:
|
||||||
# openbrain PostgreSQL instance (openbrain_brain-db).
|
image: git.mosaicstack.dev/mosaic/stack-postgres:${IMAGE_TAG:-latest}
|
||||||
# Runs once and exits. Idempotent — safe to run on every deploy.
|
|
||||||
mosaic-db-init:
|
|
||||||
image: postgres:17-alpine
|
|
||||||
environment:
|
environment:
|
||||||
PGHOST: openbrain_brain-db
|
POSTGRES_USER: ${POSTGRES_USER}
|
||||||
PGPORT: 5432
|
POSTGRES_PASSWORD: ${POSTGRES_PASSWORD}
|
||||||
PGUSER: ${BRAIN_DB_ADMIN_USER:-openbrain}
|
POSTGRES_DB: ${POSTGRES_DB}
|
||||||
PGPASSWORD: ${BRAIN_DB_ADMIN_PASSWORD}
|
POSTGRES_SHARED_BUFFERS: ${POSTGRES_SHARED_BUFFERS:-256MB}
|
||||||
MOSAIC_USER: ${POSTGRES_USER}
|
POSTGRES_EFFECTIVE_CACHE_SIZE: ${POSTGRES_EFFECTIVE_CACHE_SIZE:-1GB}
|
||||||
MOSAIC_PASSWORD: ${POSTGRES_PASSWORD}
|
POSTGRES_MAX_CONNECTIONS: ${POSTGRES_MAX_CONNECTIONS:-100}
|
||||||
MOSAIC_DB: ${POSTGRES_DB:-mosaic}
|
volumes:
|
||||||
entrypoint: ["sh", "-c"]
|
- postgres_data:/var/lib/postgresql/data
|
||||||
command:
|
healthcheck:
|
||||||
- |
|
test: ["CMD-SHELL", "pg_isready -U ${POSTGRES_USER} -d ${POSTGRES_DB}"]
|
||||||
until pg_isready -h openbrain_brain-db -p 5432 -U $${PGUSER}; do
|
interval: 10s
|
||||||
echo "Waiting for openbrain_brain-db..."
|
timeout: 5s
|
||||||
sleep 2
|
retries: 5
|
||||||
done
|
start_period: 30s
|
||||||
echo "Database ready. Creating mosaic user and database..."
|
|
||||||
|
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -tc "SELECT 1 FROM pg_roles WHERE rolname='$${MOSAIC_USER}'" | grep -q 1 || \
|
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -c "CREATE USER $${MOSAIC_USER} WITH PASSWORD '$${MOSAIC_PASSWORD}';"
|
|
||||||
|
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -tc "SELECT 1 FROM pg_database WHERE datname='$${MOSAIC_DB}'" | grep -q 1 || \
|
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -c "CREATE DATABASE $${MOSAIC_DB} OWNER $${MOSAIC_USER} ENCODING 'UTF8' LC_COLLATE='C' LC_CTYPE='C' TEMPLATE template0;"
|
|
||||||
|
|
||||||
echo "Enabling required extensions in $${MOSAIC_DB}..."
|
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -d $${MOSAIC_DB} -c "CREATE EXTENSION IF NOT EXISTS vector;"
|
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -d $${MOSAIC_DB} -c "CREATE EXTENSION IF NOT EXISTS \"uuid-ossp\";"
|
|
||||||
|
|
||||||
echo "Mosaic database ready: $${MOSAIC_DB}"
|
|
||||||
networks:
|
networks:
|
||||||
- openbrain-brain-internal
|
- internal
|
||||||
deploy:
|
deploy:
|
||||||
restart_policy:
|
restart_policy:
|
||||||
condition: on-failure
|
condition: on-failure
|
||||||
delay: 5s
|
|
||||||
max_attempts: 5
|
|
||||||
|
|
||||||
# ============================================
|
|
||||||
# CORE INFRASTRUCTURE
|
|
||||||
# ============================================
|
|
||||||
|
|
||||||
# ======================
|
# ======================
|
||||||
# Valkey Cache
|
# Valkey Cache
|
||||||
@@ -145,7 +105,7 @@ services:
|
|||||||
NODE_ENV: production
|
NODE_ENV: production
|
||||||
PORT: ${API_PORT:-3001}
|
PORT: ${API_PORT:-3001}
|
||||||
API_HOST: ${API_HOST:-0.0.0.0}
|
API_HOST: ${API_HOST:-0.0.0.0}
|
||||||
DATABASE_URL: postgresql://${POSTGRES_USER}:${POSTGRES_PASSWORD}@openbrain_brain-db:5432/${POSTGRES_DB:-mosaic}
|
DATABASE_URL: postgresql://${POSTGRES_USER}:${POSTGRES_PASSWORD}@postgres:5432/${POSTGRES_DB}
|
||||||
VALKEY_URL: redis://valkey:6379
|
VALKEY_URL: redis://valkey:6379
|
||||||
# Auth (external Authentik)
|
# Auth (external Authentik)
|
||||||
OIDC_ENABLED: ${OIDC_ENABLED:-false}
|
OIDC_ENABLED: ${OIDC_ENABLED:-false}
|
||||||
@@ -203,7 +163,6 @@ services:
|
|||||||
networks:
|
networks:
|
||||||
- internal
|
- internal
|
||||||
- traefik-public
|
- traefik-public
|
||||||
- openbrain-brain-internal
|
|
||||||
deploy:
|
deploy:
|
||||||
restart_policy:
|
restart_policy:
|
||||||
condition: on-failure
|
condition: on-failure
|
||||||
@@ -348,36 +307,36 @@ services:
|
|||||||
# ======================
|
# ======================
|
||||||
# Synapse Database Init
|
# Synapse Database Init
|
||||||
# ======================
|
# ======================
|
||||||
# Creates the 'synapse' database in the shared openbrain PostgreSQL instance.
|
# Creates the 'synapse' database in the shared PostgreSQL instance.
|
||||||
# Runs once and exits. Idempotent — safe to run on every deploy.
|
# Runs once and exits. Idempotent — safe to run on every deploy.
|
||||||
synapse-db-init:
|
synapse-db-init:
|
||||||
image: postgres:17-alpine
|
image: postgres:17-alpine
|
||||||
environment:
|
environment:
|
||||||
PGHOST: openbrain_brain-db
|
PGHOST: postgres
|
||||||
PGPORT: 5432
|
PGPORT: 5432
|
||||||
PGUSER: ${BRAIN_DB_ADMIN_USER:-openbrain}
|
PGUSER: ${POSTGRES_USER}
|
||||||
PGPASSWORD: ${BRAIN_DB_ADMIN_PASSWORD}
|
PGPASSWORD: ${POSTGRES_PASSWORD}
|
||||||
SYNAPSE_DB: ${SYNAPSE_POSTGRES_DB}
|
SYNAPSE_DB: ${SYNAPSE_POSTGRES_DB}
|
||||||
SYNAPSE_USER: ${SYNAPSE_POSTGRES_USER}
|
SYNAPSE_USER: ${SYNAPSE_POSTGRES_USER}
|
||||||
SYNAPSE_PASSWORD: ${SYNAPSE_POSTGRES_PASSWORD}
|
SYNAPSE_PASSWORD: ${SYNAPSE_POSTGRES_PASSWORD}
|
||||||
entrypoint: ["sh", "-c"]
|
entrypoint: ["sh", "-c"]
|
||||||
command:
|
command:
|
||||||
- |
|
- |
|
||||||
until pg_isready -h openbrain_brain-db -p 5432 -U $${PGUSER}; do
|
until pg_isready -h postgres -p 5432 -U $${PGUSER}; do
|
||||||
echo "Waiting for openbrain_brain-db..."
|
echo "Waiting for PostgreSQL..."
|
||||||
sleep 2
|
sleep 2
|
||||||
done
|
done
|
||||||
echo "Database ready. Creating Synapse user and database..."
|
echo "PostgreSQL is ready. Creating Synapse database and user..."
|
||||||
|
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -tc "SELECT 1 FROM pg_roles WHERE rolname='$${SYNAPSE_USER}'" | grep -q 1 || \
|
psql -h postgres -U $${PGUSER} -tc "SELECT 1 FROM pg_roles WHERE rolname='$${SYNAPSE_USER}'" | grep -q 1 || \
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -c "CREATE USER $${SYNAPSE_USER} WITH PASSWORD '$${SYNAPSE_PASSWORD}';"
|
psql -h postgres -U $${PGUSER} -c "CREATE USER $${SYNAPSE_USER} WITH PASSWORD '$${SYNAPSE_PASSWORD}';"
|
||||||
|
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -tc "SELECT 1 FROM pg_database WHERE datname='$${SYNAPSE_DB}'" | grep -q 1 || \
|
psql -h postgres -U $${PGUSER} -tc "SELECT 1 FROM pg_database WHERE datname='$${SYNAPSE_DB}'" | grep -q 1 || \
|
||||||
psql -h openbrain_brain-db -U $${PGUSER} -c "CREATE DATABASE $${SYNAPSE_DB} OWNER $${SYNAPSE_USER} ENCODING 'UTF8' LC_COLLATE='C' LC_CTYPE='C' TEMPLATE template0;"
|
psql -h postgres -U $${PGUSER} -c "CREATE DATABASE $${SYNAPSE_DB} OWNER $${SYNAPSE_USER} ENCODING 'UTF8' LC_COLLATE='C' LC_CTYPE='C' TEMPLATE template0;"
|
||||||
|
|
||||||
echo "Synapse database ready: $${SYNAPSE_DB}"
|
echo "Synapse database ready: $${SYNAPSE_DB}"
|
||||||
networks:
|
networks:
|
||||||
- openbrain-brain-internal
|
- internal
|
||||||
deploy:
|
deploy:
|
||||||
restart_policy:
|
restart_policy:
|
||||||
condition: on-failure
|
condition: on-failure
|
||||||
@@ -492,6 +451,7 @@ services:
|
|||||||
# Volumes
|
# Volumes
|
||||||
# ======================
|
# ======================
|
||||||
volumes:
|
volumes:
|
||||||
|
postgres_data:
|
||||||
valkey_data:
|
valkey_data:
|
||||||
orchestrator_workspace:
|
orchestrator_workspace:
|
||||||
speaches_models:
|
speaches_models:
|
||||||
@@ -504,6 +464,3 @@ networks:
|
|||||||
driver: overlay
|
driver: overlay
|
||||||
traefik-public:
|
traefik-public:
|
||||||
external: true
|
external: true
|
||||||
openbrain-brain-internal:
|
|
||||||
external: true
|
|
||||||
name: openbrain_brain-internal
|
|
||||||
|
|||||||
Reference in New Issue
Block a user