๐ ๏ธ All DevTools
Showing 1–20 of 4172 tools
Last Updated
April 15, 2026 at 04:00 PM
Elevated errors on Claude.ai, API, Claude Code
Hacker News (score: 168)[Other] Elevated errors on Claude.ai, API, Claude Code
MCP as Observability Interface: Connecting AI Agents to Kernel Tracepoints
Hacker News (score: 31)[Monitoring/Observability] MCP as Observability Interface: Connecting AI Agents to Kernel Tracepoints
Donchitos/Claude-Code-Game-Studios
GitHub Trending[Other] Turn Claude Code into a full game dev studio โ 49 AI agents, 72 workflow skills, and a complete coordination system mirroring real studio hierarchy.
vercel-labs/open-agents
GitHub Trending[Other] An open source template for building cloud agents.
Show HN: Xit โ a Git-compatible VCS written in Zig
Show HN (score: 8)[Other] Show HN: Xit โ a Git-compatible VCS written in Zig The marquee feature is patch-based merging, similar to Darcs and Pijul. I think xit is the first version control system (VCS) to have this feature while still being git compatible. See the 100% human-written readme for more.
Want to Write a Compiler? Just Read These Two Papers (2008)
Hacker News (score: 304)[Other] Want to Write a Compiler? Just Read These Two Papers (2008)
Direct Win32 API, Weird-Shaped Windows, and Why They Mostly Disappeared
Hacker News (score: 61)[Other] Direct Win32 API, Weird-Shaped Windows, and Why They Mostly Disappeared
Wacli โ WhatsApp CLI: sync, search, send
Hacker News (score: 128)[CLI Tool] Wacli โ WhatsApp CLI: sync, search, send
Your codebase doesn't care how it got written
Hacker News (score: 13)[Other] Your codebase doesn't care how it got written
Show HN: Plain โ The full-stack Python framework designed for humans and agents
Hacker News (score: 20)[Other] Show HN: Plain โ The full-stack Python framework designed for humans and agents
Turn your best AI prompts into one-click tools in Chrome
Hacker News (score: 35)[Other] Turn your best AI prompts into one-click tools in Chrome
Claude Code Routines
Hacker News (score: 360)[Other] Claude Code Routines
5NF and Database Design
Hacker News (score: 157)[Database] 5NF and Database Design
Show HN: Kelet โ Root Cause Analysis agent for your LLM apps
Hacker News (score: 38)[Monitoring/Observability] Show HN: Kelet โ Root Cause Analysis agent for your LLM apps I've spent the past few years building 50+ AI agents in prod (some reached 1M+ sessions/day), and the hardest part was never building them โ it was figuring out why they fail.<p>AI agents don't crash. They just quietly give wrong answers. You end up scrolling through traces one by one, trying to find a pattern across hundreds of sessions.<p>Kelet automates that investigation. Here's how it works:<p>1. You connect your traces and signals (user feedback, edits, clicks, sentiment, LLM-as-a-judge, etc.) 2. Kelet processes those signals and extracts facts about each session 3. It forms hypotheses about what went wrong in each case 4. It clusters similar hypotheses across sessions and investigates them together 5. It surfaces a root cause with a suggested fix you can review and apply<p>The key insight: individual session failures look random. But when you cluster the hypotheses, failure patterns emerge.<p>The fastest way to integrate is through the Kelet Skill for coding agents โ it scans your codebase, discovers where signals should be collected, and sets everything up for you. There are also Python and TypeScript SDKs if you prefer manual setup.<p>Itโs currently free during beta. No credit card required. Docs: <a href="https://kelet.ai/docs/" rel="nofollow">https://kelet.ai/docs/</a><p>I'd love feedback on the approach, especially from anyone running agents in prod. Does automating the manual error analysis sound right?
Show HN: A memory database that forgets, consolidates, and detects contradiction
Hacker News (score: 20)[Database] Show HN: A memory database that forgets, consolidates, and detects contradiction Vector databases store memories. They don't manage them. After 10k memories, recall quality degrades because there's no consolidation, no forgetting, no conflict resolution. Your AI agent just gets noisier.<p>YantrikDB is a cognitive memory engine โ embed it, run it as a server, or connect via MCP. It thinks about what it stores: consolidation collapses duplicate memories, contradiction detection flags incompatible facts, temporal decay with configurable half-life lets unimportant memories fade like human memory does.<p>Single Rust binary. HTTP + binary wire protocol. 2-voter + 1-witness HA cluster via Docker Compose or Kubernetes. Chaos-tested failover, runtime deadlock detection (parking_lot), per-tenant quotas, Prometheus metrics. Ran a 42-task hardening sprint last week โ 1178 core tests, cargo-fuzz targets, CRDT property tests, 5 ops runbooks.<p>Live on a 3-node Proxmox homelab cluster with multiple tenants. Alpha โ primary user is me, looking for the second one.
Show HN: MลBrowser, a TypeScript-first desktop app framework with typed IPC
Show HN (score: 5)[Other] Show HN: MลBrowser, a TypeScript-first desktop app framework with typed IPC Hi HN,<p>For the last ~15 years I've worked on embedding web browsers into Java and .NET desktop apps (JxBrowser, DotNetBrowser). Over time, I watched many teams move from embedding web views into native apps, to building full desktop apps with frameworks like Electron and Tauri.<p>Both are useful, but in practice I kept running into several problems.<p>With Electron, beyond the larger app footprint, I often ran into:<p><pre><code> - lack of type-safe IPC - no source code protection - weak support for the modern web stack </code></pre> Tauri solves some problems (like app size), but introduces others:<p><pre><code> - different WebViews across platforms โ inconsistent behavior - requires Rust + JS instead of a single stack </code></pre> So we built MลBrowser, a framework for building desktop apps with TypeScript, Node.js, and Chromium.<p>Some of the things we focused on:<p><pre><code> - typed IPC using Protobuf + code generation (RPC-style communication instead of string channels) - consistent rendering and behavior across different platforms - Node.js runtime - built-in packaging, updates, and scaffolding - source code protection - small delta auto-updates </code></pre> The goal is to let web developers ship desktop apps with a web stack they already know and fewer cross-platform surprises.<p>I'd especially love feedback from people who have built production apps with Electron or Tauri.<p>Happy to answer any questions.
Show HN: LangAlpha โ what if Claude Code was built for Wall Street?
Hacker News (score: 64)[Other] Show HN: LangAlpha โ what if Claude Code was built for Wall Street? Some technical context on what we ran into building this.<p>MCP tools don't really work for financial data at scale. One tool call for five years of daily prices dumps tens of thousands of tokens into the context window. And data vendors pack dozens of tools into a single MCP server, schemas alone can eat 50k+ tokens before the agent does anything useful. So we auto-generate typed Python modules from the MCP schemas at workspace init and upload them into the sandbox. The agent just imports them like a normal library. Only a one-line summary per server stays in the prompt. We have around 80 tools across our servers and the prompt cost is the same whether a server has 3 tools or 30. This part isn't finance-specific, it works with any MCP server.<p>The other big thing was making research actually persist across sessions. Most agents treat a single deliverable (a PDF, a spreadsheet) as the end goal. In investing that's day one. You update the model when earnings drop, re-run comps when a competitor reports, keep layering new analysis on old. But try doing that across agent sessions, files don't carry over, you re-paste context every time. So we built everything around workspaces. Each one maps to a persistent sandbox, one per research goal. The agent maintains its own memory file with findings and a file index that gets re-read before every LLM call. Come back a week later, start a new thread, it picks up where it left off.<p>We also wanted the agent to have real domain context the way Claude Code has codebase context. Portfolio, watchlist, risk tolerance, financial data sources, all injected into every call. Existing AI investing platforms have some of that but nothing close to what a proper agent harness can do. We wanted both and couldn't find it, so we built it and open-sourced the whole thing.
Show HN: Kontext CLI โ Credential broker for AI coding agents in Go
Hacker News (score: 13)[CLI Tool] Show HN: Kontext CLI โ Credential broker for AI coding agents in Go We built the Kontext CLI because AI coding agents need access to GitHub, Stripe, databases, and dozens of other services โ and right now most teams handle this by copy-pasting long-lived API keys into .env files, or the actual chat interface, whilst hoping for the best.<p>The problem isn't just secret sprawl. It's that there's no lineage of access. You don't know which developer launched which agent, what it accessed, or whether it should have been allowed to. The moment you hand raw credentials to a process, you've lost the ability to enforce policy, audit access, or rotate without pain. The credential is the authorization, and that's fundamentally broken when autonomous agents are making hundreds of API calls per session.<p>Kontext takes a different approach. You declare what credentials a project needs in a .env.kontext file:<p><pre><code> GITHUB_TOKEN={{kontext:github}} STRIPE_KEY={{kontext:stripe}} LINEAR_TOKEN={{kontext:linear}} </code></pre> Then run `kontext start --agent claude`. The CLI authenticates you via OIDC, and for each placeholder: if the service supports OAuth, it exchanges the placeholder for a short-lived access token via RFC 8693 token exchange; for static API keys, the backend injects the credential directly into the agent's runtime environment. Either way, secrets exist only in memory during the session โ never written to disk on your machine. Every tool call is streamed for audit as the agent runs.<p>The closest analogy is a Security Token Service (STS): you authenticate once, and the backend mints short-lived, scoped credentials on-the-fly โ except unlike a classical STS, we hold the upstream secrets, so nothing long-lived ever reaches the agent. The backend holds your OAuth refresh tokens and API keys; the CLI never sees them. It gets back short-lived access tokens scoped to the session.<p>What the CLI captures for every tool call: what the agent tried to do, what happened, whether it was allowed, and who did it โ attributed to a user, session, and org.<p>Install with one command: `brew install kontext-dev/tap/kontext`<p>The CLI is written in Go (~5ms hook overhead per tool call), uses ConnectRPC for backend communication, and stores auth in the system keyring. Works with Claude Code today, Codex support coming soon.<p>We're working on server-side policy enforcement next โ the infrastructure for allow/deny decisions on every tool call is already wired, we just need to close the loop so tool calls can also be rejected.<p>We'd love feedback on the approach. Especially curious: how are teams handling credential management for AI agents today? Are you just pasting env vars into the agent chat, or have you found something better?<p>GitHub: <a href="https://github.com/kontext-dev/kontext-cli" rel="nofollow">https://github.com/kontext-dev/kontext-cli</a> Site: <a href="https://kontext.security" rel="nofollow">https://kontext.security</a>
jj โ the CLI for Jujutsu
Hacker News (score: 337)[CLI Tool] jj โ the CLI for Jujutsu
Show HN: A CLI that writes its own integration code
Show HN (score: 5)[CLI Tool] Show HN: A CLI that writes its own integration code We run superglue, an OSS agentic integration platform. Last week I talked to a founder of another YC startup. She found a use case for our CLI that we hadn't officially launched yet.<p>Her problem: customers wanted to create Opps in Salesforce from inside the chat in her app. We kept seeing this pattern: teams build agents and their users can perfectly describe what they want: "pull these three objects from Salesforce and push to nCino when X condition is true", but translating that into a generalized hard-coded tool the agent can call is a lot of work and does not scale since the logic is different for every user.<p>What superglue CLI does: you point it at any API, and your agent gets the ability to reason over that API at runtime. No pre-built tools. The agent reads the spec, plans the calls, executes them.<p>The founder using this in production described it like this: she gave the CLI to her agent with an instruction set and told it not to build tools, just run against the API. It handled multi-step Salesforce object creation correctly, including per-user field logic and record type templates.<p>Concretely: instead of writing a createSalesforceOpp tool that handles contact -> account -> Opp creation with all the conditional logic, you write a skill doc and let the agent figure out which endpoints to hit and in what order.<p>The tradeoff is: you're giving the agent more autonomy over what API calls it makes. That requires good instructions and some guardrails. But for long-tail, user-specific connectors, it's a lot more practical than building a tool for every case.<p>Happy to discuss. Curious if others have run into the "pre-defined tool" ceiling with MCP-based connectors and how you've worked around it.<p>Docs: <a href="https://docs.superglue.cloud/getting-started/cli-skills" rel="nofollow">https://docs.superglue.cloud/getting-started/cli-skills</a> Repo: <a href="https://github.com/superglue-ai/superglue" rel="nofollow">https://github.com/superglue-ai/superglue</a>