jmuncor/tokentap

Intercept LLM API traffic and visualize token usage in a real-time terminal dashboard. Track costs, debug prompts, and monitor context window usage across your AI development sessions.

60
/ 100
Established

This tool helps AI application developers monitor and debug their interactions with large language models (LLMs) in real time. It intercepts your LLM API calls, displaying token usage, cost, and context window consumption in a live terminal dashboard. Developers building with LLM CLI tools for providers like Claude, OpenAI Codex, or MiniMax will find this essential for managing their AI-powered applications.

761 stars. Actively maintained with 7 commits in the last 30 days. Available on PyPI.

Use this if you are developing with LLMs via command-line tools and need clear, immediate insights into token usage, context limits, and the actual prompts being sent.

Not ideal if you are a non-developer or if your LLM interactions are solely through web UIs or cloud-managed platforms where token tracking is already integrated.

LLM development prompt engineering API monitoring cost management debugging
Maintenance 17 / 25
Adoption 10 / 25
Maturity 20 / 25
Community 13 / 25

How are scores calculated?

Stars

761

Forks

36

Language

Python

License

MIT

Last pushed

Feb 02, 2026

Commits (30d)

7

Dependencies

4

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/jmuncor/tokentap"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.