Early Access Preview
Back to blog
engineeringrelease-notesinfrastructure

Rapid Evolution: Containerizing Prometheus, Squashing vLLM Bugs, and Expanding MCP

February 27, 20266 min readAitherium
Share

Building an operating system for AI agents means the infrastructure has to evolve as fast as the cognitive models running on top of it. Today was one of those massive inflection points for AitherOS where multiple foundational layers leveled up simultaneously.

1. The vLLM Context Window Fix

If you run local LLMs, you know the pain of context window overflows. We caught a critical bug where our DynamicTokenScaler, when pushed to max effort on Ultra-tier hardware (like the RTX 5090), was eagerly allocating up to 131,000 tokens for generation.

The problem? The safety clamp meant to protect vLLM from these massive allocations was failing silently due to a missing method (_get_vllm_context_window). This resulted in hard crashes with max_model_len errors, especially during complex multi-step tasks like our 4-step SDXL image generations. We implemented the missing logic, ensuring AitherOS dynamically queries the loaded model's true context size and mathematically clamps the request. Now, you can safely crank the agent effort to 10 without ever crashing the inference server.

2. Integrating Durable Memory (Hypernetworks) and NanoProver

AitherOS isn't just about executing tasks; it's about learning from them. We officially integrated AitherNanoGPT and AitherNanoProver into our core Docker Compose stack.

This allows us to leverage Durable Memory via hypernetworks. Instead of just relying on RAG (Retrieval-Augmented Generation) which scales linearly and bloats context, our agents can now crystallize knowledge into dense, recallable memory structures. We also fixed the Daydream and Slumber pipelines, ensuring our idle agents actually use their downtime to process these memories and evolve.

3. Expanding Perception: ComfyUI & Canvas Consolidation

Visual generation just got a massive workflow upgrade. We successfully bridged the ComfyUI Beginner Bible dictionary directly into AitherOS via the Model Context Protocol (MCP). Now, our agents don't just generate images — they actually understand the node structures required to build complex ComfyUI workflows and can explain them to users.

Alongside this, we consolidated all random HTTP calls to ComfyUI across the codebase into a single, unified canvas.py client, giving us centralized logging, watermarking, and error handling for all visual cognition.

4. Containerizing Prometheus

We completely Dockerized the Vite-based Prometheus web application (prometheus.aitherium.com). It now runs seamlessly behind our Nginx proxy within the primary docker-compose.aitheros.yml stack. In the process, we also fixed our Cloudflare Tunnel configuration that was crash-looping, ensuring the entire AitherOS microservice mesh is securely exposed to the outside world.

The Loop Continues

Every bug squashed and every container deployed gets us closer to the vision: a fully sovereign, self-healing, locally-hosted operating system for agentic workflows. And the best part? The agents themselves helped write the code to fix these issues today.

Enjoyed this post?
Share