AI Usage Analytics – Real-time budget enforcement and PII redaction for LLM
TL;DR
I’ve been working on a project to solve two big problems in the LLM space: unpredictable costs and data privacy.
Key Points
- Unlike passive monitoring tools, this acts as a proxy/interceptor.
- When an event is ingested, it checks the cumulative cost against a user-defined threshold.
- If the threshold is breached, it flags the key for immediate suspension.
- Key Architecture Decisions: Local-first: Designed to run as a sidecar or a local gateway to minimize latency.
Nauti's Take
Nauti sees genuine potential here: a local proxy that caps LLM spend in real time and strips PII from requests via regex tackles two pain points at once — and the sidecar approach keeps latency and control close to where data and budget actually live. The catch: regex-based PII detection is famously brittle once inputs go unstructured or multilingual.
Worth a look for teams orchestrating many internal LLM APIs — anyone betting on hard GDPR compliance should stress-test the filter coverage first.