AI Usage Analytics – Real-time budget enforcement and PII redaction for LLM

TL;DR

I’ve been working on a project to solve two big problems in the LLM space: unpredictable costs and data privacy.

Key Points

  • Unlike passive monitoring tools, this acts as a proxy/interceptor.
  • When an event is ingested, it checks the cumulative cost against a user-defined threshold.
  • If the threshold is breached, it flags the key for immediate suspension.
  • Key Architecture Decisions: Local-first: Designed to run as a sidecar or a local gateway to minimize latency.

Nauti's Take

Nauti sees genuine potential here: a local proxy that caps LLM spend in real time and strips PII from requests via regex tackles two pain points at once — and the sidecar approach keeps latency and control close to where data and budget actually live. The catch: regex-based PII detection is famously brittle once inputs go unstructured or multilingual.

Worth a look for teams orchestrating many internal LLM APIs — anyone betting on hard GDPR compliance should stress-test the filter coverage first.

Sources