snip - LLM token-saving proxy in Go
CLI proxy for Claude Code, Cursor, Copilot, and Gemini that strips noise from LLM context with declarative YAML filters. Reports 60-90% token savings on typical agent traffic.
This entry doesn't have a long-form writeup yet. Follow the source link above for the full context.
Featured in
Claude Code tools, plugins, and integrations
The best tools, MCP servers, and harnesses for getting more out of Claude Code - orchestration, observability, telemetry, and remote control.
Go CLI tools for developers and operators
Single-binary Go CLIs and TUIs across observability, git, shells, and DevOps - the tools that make Go the practical default for distribution-friendly CLIs.
Observability for AI coding agents
Tools that show you what your coding agents are actually doing: token spend, session state, tool calls, and parallel execution.
Tools for OpenAI Codex CLI
The Codex-aware slice of the directory: orchestration, observability, sandboxes, and bridges built specifically for the OpenAI Codex runtime.
Related entries
Codeburn - token cost TUI for Claude Code & Codex
Interactive terminal dashboard that breaks down where your AI coding tokens actually go. Surfaces the chat-vs-tool-use split most users get wrong.
jeeves - TUI for browsing AI agent sessions
Terminal UI to search, preview, read, and resume Claude Code and Codex sessions in a unified view. More framework integrations planned.
coding_agent_session_search - 11-provider session search
Rust TUI and CLI that indexes and searches local coding-agent session history across Codex, Claude Code, Gemini, Cursor, Aider and seven other providers.
haft - engineering decisions engine with evidence decay
Go CLI that records, frames, compares, and decides engineering choices with parity enforcement and decay tracking on the underlying evidence. Integrates with Claude Code, Cursor, Gemini CLI, and Codex.