Show HN: Hydra – Never stop coding when your AI CLI hits a rate limit

Show HN (score: 5)
Found: April 21, 2026
ID: 4267

Description

CLI Tool
Show HN: Hydra – Never stop coding when your AI CLI hits a rate limit I built Hydra because I kept losing my flow when Claude Code hit usage limits mid-task. I would copy context, open another tool, and then re-explain everything. This would be super annoying for me.

Hydra wraps your AI coding CLIs (Claude Code, Codex, OpenCode, Pi, or any terminal-based tool) in a single command. It monitors terminal output for rate limit patterns, and when one provider runs out, you switch to another with one keypress. Your conversation history, git diff, and recent commits are automatically copied to your clipboard so you can paste and keep going.

The fallback chain is configurable. Mine goes Claude Code → OpenCode (free Gemini) → Codex → Pi (free Gemini). The free tiers alone give you ~3000 requests/day, so even after burning through paid limits you can keep working.

Key details: Full PTY passthrough you see the exact same TUI as running the CLI directly hydra switch from another terminal signals ALL running sessions (rate limits are account-wide) Context extraction parses Claude Code's JSONL session files for real conversation history, not just recent output Any CLI that runs in a terminal works as a provider Single Go binary, ~200 lines of core logic

https://github.com/saadnvd1/hydra

Nothing amazing, but wanted to share with others in case it's useful. Feel free to modify it as you see fit.

More from Show

Show HN: Gemini Plugin for Claude Code

Show HN: Gemini Plugin for Claude Code I built a plugin that lets Claude Code delegate work to Gemini CLI.<p>I started this after finding myself reaching for Gemini more often on long context repo work. I have been especially liking Gemini’s codebaseinvestigator for long context.<p>This is inspired by openai&#x2F;codex-plugin-cc.<p>Code Review, adversarial review. Under the hood its Gemini CLI over ACP<p>Would love feedback from people using Claude Code, Gemini CLI, or ACP. I am especially curious whether this feels useful outside my own workflow.<p>Its a great combo with Opus 4.7 + Gemini 3.1 workflows

Show HN: gcx – The Official Grafana Cloud CLI

Show HN: gcx – The Official Grafana Cloud CLI Hi HN,<p>We’re excited to share gcx, a new CLI we’ve been building for Grafana Cloud.<p>With the rise of agentic coding tools like Claude Code and Codex we&#x27;re building faster than ever, but these agents are often blind to what’s actually happening in production.<p>gcx brings the full power of Grafana Cloud observability to your terminal. Query production. Investigate alerts. Let the Assistant root-cause issues. Ship fixes with observability built in. Without leaving your editor. gcx also comes packaged with a skills bundle that allow agents to see and act on your production telemetry. You can ask an agent to root-cause a latency spike, and it can actually fetch the telemetry, analyze the spans, and suggest a fix—all while having the full context of your codebase.<p>Do check it out and give us feedback!<p>Github link: <a href="https:&#x2F;&#x2F;github.com&#x2F;grafana&#x2F;gcx" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;grafana&#x2F;gcx</a>

No other tools from this source yet.