100 / 744

Show HN: Clusterflock: An AI orchestrator for networked hardware

TL;DR

Clusterflock is an open-source AI orchestrator designed to manage agents across distributed hardware with varying VRAM and RAM constraints.

Key Points

  • It automatically profiles networked hardware and downloads the best-fit models from HuggingFace without manual configuration.
  • Native parallelism via llama.cpp enables multiple smaller models to run simultaneously on the same device.
  • The built-in mission runner supports multi-session, asynchronous agent workflows out of the box.

Nauti's Take

This is exactly the kind of tooling the self-hosting community has been missing. Cloud providers solve the deployment problem by throwing money at it – Clusterflock solves it with smart hardware-aware logic instead.

Whether tight packing holds up under real-world load remains to be proven, but the architecture looks thoughtfully designed by people who actually ran into these problems themselves. Worth watching closely.

Context

Running AI agents on self-hosted hardware typically means manually distributing models, juggling VRAM limits, and coordinating deployments across machines. Clusterflock tackles this bottleneck with hardware-aware automation that was previously only practical in cloud environments. Being fully open-source and built on llama.

cpp makes it especially compelling for self-hosters and small teams who want serious agentic infrastructure without vendor lock-in.

Sources