Skip to content

Roadmap

Nova is under active development. This page summarizes the major phases of the project. For the full internal roadmap with implementation details, see the docs/roadmap.md file in the repository.

A self-directed autonomous AI platform. You define a goal. Nova breaks it into subtasks, executes them through a coordinated pipeline of specialized agents with built-in safety rails, evaluates its own progress, re-plans as needed, and completes the goal — with minimal human intervention except when it genuinely needs a decision.

LevelDescriptionStatus
1 — Pipeline autonomyQuartet runs all agents without human input; escalates only on critical flagsImplemented
2 — Async executionTasks run in the background; submit and come backImplemented
3 — Triggered executionTasks start from external events — git push, cron, webhook, SlackPlanned
4 — Self-directedNova breaks goals into subtasks, executes, evaluates, re-plans, loops to completionPlanned

The foundation: containerized microservices communicating over HTTP. Multi-turn agent loop with tool use, streaming responses via SSE and WebSocket, pluggable tool system, and 39+ registered model IDs.

Phase 2 — Auth, Billing & IDE Integration

Section titled “Phase 2 — Auth, Billing & IDE Integration”

API key authentication (SHA-256 hashed, sk-nova-* format), per-key rate limiting via Redis, admin-only endpoints, PostgreSQL storage for keys and usage events, token counting and cost tracking, and OpenAI-compatible endpoints for IDE integration (Continue.dev, Cursor, Aider).

Workspace-scoped file I/O (list_dir, read_file, write_file), shell execution with timeout and denylist, ripgrep code search, git operations, path traversal protection, and Docker workspace volume mounting. Includes sandbox tier design (isolated/nova/workspace/host).

Phase 4 — Quartet Pipeline & Async Queue

Section titled “Phase 4 — Quartet Pipeline & Async Queue”

The 5-stage agent pipeline (Context, Task, Guardrail, Code Review, Decision), Redis BRPOP task queue with heartbeat and stale reaper, 11-state task state machine, human-in-the-loop review, clarification requests, pod and agent configuration (per-agent model, tools, system prompt, run conditions), and subscription-based LLM providers (Claude Max, ChatGPT Plus) for zero-cost operation.

React admin UI with Overview, Chat, Usage, Keys, Models, Tasks, Pods, MCP, Memory Inspector, Agent Endpoints, Settings, Recovery, and Remote Access pages. Built with Vite, Tailwind CSS, TanStack Query, and Recharts.

Operational maturity improvements: fixed MCP tool visibility for agents, test foundation (pytest), streaming token count fixes, reaper race condition fix, structured JSON logging across all services, 3-tier embedding cache, and working memory cleanup.

Three-tier memory architecture (semantic, procedural, episodic) with hybrid retrieval (70% cosine similarity + 30% ts_rank), ACT-R confidence decay, fact upserts, embedding fallback chain, and Memory Inspector dashboard page. Upcoming: auto fact extraction from conversations, “what do I know about X?” knowledge queries, cross-session consolidation, and persistent task history with full reasoning traces.

Reusable prompt templates (skills) shared across agents/pods, and declarative behavior constraints (rules) with soft (LLM-based) and hard (pre-execution regex) enforcement. See Skills & Rules for details.

Phase 6c — SDK, CLI/TUI & Documentation (planned)

Section titled “Phase 6c — SDK, CLI/TUI & Documentation (planned)”

Typed Python SDK (nova-sdk), CLI with Typer + Rich (nova-cli), interactive TUI with Textual, auto-generated TypeScript types from Pydantic contracts, and a comprehensive documentation system.

Phase 7 — Self-Directed Autonomy (planned)

Section titled “Phase 7 — Self-Directed Autonomy (planned)”

Planning Agent that decomposes goals into subtask DAGs, executes them through the pipeline, evaluates results, and re-plans. This is the core of Nova’s autonomous operation.

Phase 8b — MCP Integrations Hub (planned)

Section titled “Phase 8b — MCP Integrations Hub (planned)”

One-click integrations connecting Nova to your self-hosted services and developer tools via MCP servers. Browse, enable, and configure integrations from the dashboard with minimal setup.

Homelab: Home Assistant (device control), n8n (workflow orchestration), Nextcloud (files/calendar), Paperless-ngx (documents), Immich (photos), Gitea (local git), Uptime Kuma (monitoring), Portainer (containers).

Developer: GitHub (repos/PRs/issues), Linear (project tracking), Notion (knowledge base), Slack/Discord (messaging).

System: Filesystem (host file access), Docker (container management), Cloudflare (DNS, tunnels, custom domain deployment), SSH (remote execution), Prometheus/Grafana (metrics).

Knowledge: Brave Search (web search), Playwright (browser automation), external vector DBs, arbitrary SQL databases.

Each integration ships as a Docker Compose profile or connects to an existing service via URL + API key. Dashboard provides an Integrations page with enable/disable toggles, config modals, connection testing, and real-time health status.

Devices & Infrastructure: A dashboard page showing all physical machines connected to Nova — real-time status (online/sleeping/offline), hardware specs, running services, installed models, resource utilization, and Wake-on-LAN controls. Nova uses device awareness for smart inference routing (auto-wake GPU box when needed, fall back to cloud when offline).

Custom Domain Self-Deployment: With the Cloudflare MCP integration, Nova can deploy itself at a user’s custom domain (e.g., nova.mydomain.com). Nova creates the Cloudflare Tunnel, DNS record, and SSL configuration automatically — zero manual DNS setup required.

External event triggers: git webhooks, cron schedules, Slack commands, and custom webhook endpoints that automatically submit tasks to the pipeline.

Browser-based code editor with git workspace management, integrated with Nova’s agent pipeline for AI-assisted development.

Remote Access + Multi-Device Gateway (planned)

Section titled “Remote Access + Multi-Device Gateway (planned)”

Nova as a distributed personal AI network. Each device (mini-PC, desktop, laptop) runs its own Nova gateway with different LLM backends, sharing one memory service. Chat through the same PWA from your phone regardless of which gateway you’re hitting. Per-device routing: cloud-only on low-power devices, local Ollama on GPU boxes, hybrid elsewhere.

Migration from nova.arialabs.ai (docs site) to a split architecture: arialabs.ai becomes the Aria Labs company site with Nova docs at arialabs.ai/nova/docs/, while nova.arialabs.ai becomes a live private Nova instance accessible from any device via Cloudflare Tunnel + Access. When SaaS launches, the personal instance moves to a personal subdomain and nova.arialabs.ai becomes the SaaS endpoint.

Edge deployment capabilities for running Nova agents closer to data sources and event triggers.

Multi-cloud deployment support for distributing Nova across cloud providers.

Multiple local inference backends (Ollama, vLLM, llama.cpp, SGLang) for concurrent serving, GPU upgrade path, and optimized multi-user performance.

User isolation for multi-person deployments: separate chat histories, memory spaces, API keys, preferences, and usage tracking. Includes authentication, role-based access, and per-user data scoping.

Phase 14 — SaaS & Hosted Offering (planned)

Section titled “Phase 14 — SaaS & Hosted Offering (planned)”

Nova Cloud at nova.arialabs.ai — a hosted version where users sign up and use Nova without self-hosting. Three tiers (Free, Pro, Enterprise), Stripe billing, Kubernetes infrastructure, and full data portability between SaaS and self-hosted. Builds on Phase 12 (concurrent inference) and Phase 13 (multi-tenancy).

Nova is open source. Check the full roadmap for detailed implementation plans, or dive into the codebase to start contributing.