Stemma exists because debugging AI in production shouldn't require a data team, a custom dashboard, or expensive monthly bills.
Every developer building with LLMs hits the same wall: something breaks in production, and you have no idea which prompt version caused it, how much it cost, or how slow it was. You start throwing console.log statements around your API calls, building one-off dashboards, and manually scanning costs in your cloud provider's billing page.
Stemma is the tool we wish existed: a lightweight SDK that captures everything that matters — latency, tokens, cost, inputs, outputs — and surfaces it in a clean dashboard. Three lines of code, fire-and-forget, and you have full visibility from day one.
We believe observability tooling should be transparent, simple, and affordable for solo developers and small teams — not just enterprises with six-figure contracts.
The principles behind every product decision we make.
Observability shouldn't require a platform team. If it takes more than 5 minutes to set up, we've failed.
Hard cost caps, no overages, no hidden fees. Your bill is exactly what you signed up for — every month.
The entire codebase is public. Run it yourself, fork it, audit it. You own your data and always will.
Questions, feedback, or just want to talk LLM observability?
hello@stemma.dev