VibeHunt
Back to browse

Calcis

Know what your LLM calls cost before you make them.

Visit

Calcis provides pre‑flight cost estimates for calls to large language model APIs, allowing users to see the exact price of a request before it is sent. It supports models from OpenAI, Anthropic, and Google, tracking more than twenty‑eight models and updating pricing within hours of any change. The service offers a price ledger, a REST API, a CLI, and a GitHub Action, giving engineers a reliable source of truth for LLM pricing that can be integrated into development workflows.

The tool is aimed at engineers and teams that need to manage AI spend and avoid unexpected charges. By using the same tokenizers as each provider, Calcis can calculate per‑call costs based on input and output token counts, presenting the data in a searchable index and live comparison view. It logs price changes, timestamps updates, and verifies the information to maintain accuracy.

Calcis is experimental software, maintained as an open‑source project. It focuses on transparency and predictability of LLM usage costs rather than post‑mortem billing analysis, and it can be accessed via command‑line, API, or CI/CD integration.

Reviews

Sign in to leave a review.

Loading reviews…

Similar apps