jmuncor/tokentap
Intercept LLM API traffic and visualize token usage in a real-time terminal dashboard. Track costs, debug prompts, and monitor context window usage across your AI development sessions.
This tool helps AI application developers monitor and debug their interactions with large language models (LLMs) in real time. It intercepts your LLM API calls, displaying token usage, cost, and context window consumption in a live terminal dashboard. Developers building with LLM CLI tools for providers like Claude, OpenAI Codex, or MiniMax will find this essential for managing their AI-powered applications.
761 stars. Actively maintained with 7 commits in the last 30 days. Available on PyPI.
Use this if you are developing with LLMs via command-line tools and need clear, immediate insights into token usage, context limits, and the actual prompts being sent.
Not ideal if you are a non-developer or if your LLM interactions are solely through web UIs or cloud-managed platforms where token tracking is already integrated.
Stars
761
Forks
36
Language
Python
License
MIT
Category
Last pushed
Feb 02, 2026
Commits (30d)
7
Dependencies
4
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/jmuncor/tokentap"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related tools
AgentOps-AI/tokencost
Easy token price estimates for 400+ LLMs. TokenOps.
Merit-Systems/echo
The User Pays AI SDK
Ruthwik000/tokenfirewall
Scalable LLM cost enforcement middleware for Node.js with budget protection and multi-provider support
adarshxs/TokenTally
Estimate Your LLM's Token Toll Across Various Platforms and Configurations
azat-io/token-limit
🛰 Monitor how many tokens your code and configs consume in AI tools. Set budgets and get alerts...