Patching servers at scale forces infrastructure teams to balance critical security updates against the risk of system downtime. Operators navigate disparate tools across thousands of nodes, where failure means executing high-risk operations manually under tight maintenance windows — and diagnosing every error themselves.
Nexus reframes patching as a coordination problem solved by AI agents. Dependency mapping, staggered rollouts, and node-by-node verification run autonomously — but the operator stays in the loop at moments of real consequence, with escalation designed to surface in the UI for human review and approval.
The operator's primary surface anchors on a central Topology for live host and cluster state, with the Agent Roster, Queued Tasks, Live Feed, and a Requires Attention lane along the bottom. Every signal needed to read the system sits on one screen.
When the Critic Agent detects an anomaly during the patching process, it deliberately halts. A "Requires Attention" lane lights up at the foot of the dashboard, and a modal forces explicit human authorization before any irreversible step. Friction here isn't a UX bug — it's the product.
Once authorized, the orchestration proceeds visually. Workloads animate across the topology showing VM migration from host-a to host-b, so the operator can see exactly which cluster was memory-constrained, and required an explicit authorization to allow patching to proceed.
I took these workflows and wireframes to Claude Design and asked it to design the interface while keeping the happy path in the workflow intact. After iterative prompting, it produced a complete prototype — login, dashboard, and the visualization of VMs migrating between hosts.
Consumer design optimizes friction away. Mission-critical AI requires intentional friction — escalation modals, explicit handoffs — at the exact moments where speed would be a liability.
Operators need spatial awareness of where agents are working and chronological auditing of what they did. Opaque automation erodes trust the moment something goes wrong.
Specific intents ("Evacuating Host-A") transform AI from opaque to collaborative. Status indicators tell you the system is alive; named intents tell you what it's actually trying to do.