🛠️ Hacker News Tools
Showing 821–840 of 2486 tools from Hacker News
Last Updated
April 22, 2026 at 04:00 PM
Reproducible and traceable configuration for Conan C and C++ package manager
Hacker News (score: 16)[Package Manager] Reproducible and traceable configuration for Conan C and C++ package manager
Show HN: Trust Protocols for Anthropic/OpenAI/Gemini
Hacker News (score: 23)[API/SDK] Show HN: Trust Protocols for Anthropic/OpenAI/Gemini Much of my work right now involves complex, long-running, multi-agentic teams of agents. I kept running into the same problem: “How do I keep these guys in line?” Rules weren’t cutting it, and we needed a scalable, agentic-native STANDARD I could count on. There wasn’t one. So I built one.<p>Here are two open-source protocols that extend A2A, granting AI agents behavioral contracts and runtime integrity monitoring:<p>- Agent Alignment Protocol (AAP): What an agent can do / has done. - Agent Integrity Protocol (AIP): What an agent is thinking about doing / is allowed to do.<p>The problem: AI agents make autonomous decisions but have no standard way to declare what they're allowed to do, prove they're doing it, or detect when they've drifted. Observability tools tell you what happened. These protocols tell you whether what happened was okay.<p>Here's a concrete example. Say you have an agent who handles customer support tickets. Its Alignment Card declares:<p>{ "permitted": ["read_tickets", "draft_responses", "escalate_to_human"], "forbidden": ["access_payment_data", "issue_refunds", "modify_account_settings"], "escalation_triggers": ["billing_request_over_500"], "values": ["accuracy", "empathy", "privacy"] }<p>The agent gets a ticket: "Can you refund my last three orders?" The agent's reasoning trace shows it considering a call to the payments API. AIP reads that thinking, compares it to the card, and produces an Integrity Checkpoint:<p>{ "verdict": "boundary_violation", "concerns": ["forbidden_action: access_payment_data"], "reasoning": "Agent considered payments API access, which is explicitly forbidden. Should escalate to human.", "confidence": 0.95 }<p>The agent gets nudged back before it acts. Not after. Not in a log you review during a 2:00 AM triage. Between this turn and the next.<p>That's the core idea. AAP defines what agents should do (the contract). AIP watches what they're actually thinking and flags when those diverge (the conscience). Over time, AIP builds a drift profile — if an agent that was cautious starts getting aggressive, the system notices.<p>When multiple agents work together, it gets more interesting. Agents exchange Alignment Cards and verify value compatibility before coordination begins. An agent that values "move fast" and one that values "rollback safety" registers low coherence, and the system surfaces that conflict before work starts. Live demo with four agents handling a production incident: <a href="https://mnemom.ai/showcase" rel="nofollow">https://mnemom.ai/showcase</a><p>The protocols are Apache-licensed, work with any Anthropic/OpenAI/Gemini agent, and ship as SDKs on npm and PyPI. A free gateway proxy (smoltbot) adds integrity checking to any agent with zero code changes.<p>GitHub: <a href="https://github.com/mnemom" rel="nofollow">https://github.com/mnemom</a> Docs: docs.mnemom.ai Demo video: <a href="https://youtu.be/fmUxVZH09So" rel="nofollow">https://youtu.be/fmUxVZH09So</a>
Show HN: VectorNest responsive web-based SVG editor
Show HN (score: 8)[IDE/Editor] Show HN: VectorNest responsive web-based SVG editor I’ve just released VectorNest — an open-source, browser-based SVG editor.<p>If you have an SVG and need quick edits (paths, alignment, small fixes, animations, LLM assistance) without installing software, this is for you.<p>Try the demo: <a href="https://ekrsulov.github.io/vectornest/" rel="nofollow">https://ekrsulov.github.io/vectornest/</a> GitHub repo: <a href="https://github.com/ekrsulov/vectornest" rel="nofollow">https://github.com/ekrsulov/vectornest</a><p>Feedback, issues and contributions are welcome.
Show HN: I built a fuse box for microservices
Hacker News (score: 18)[DevOps] Show HN: I built a fuse box for microservices Hey HN! I'm Rodrigo, I run distributed systems across a few countries. I built Openfuse because of something that kept bugging me about how we all do circuit breakers.<p>If you're running 20 instances of a service and Stripe starts returning 500s, each instance discovers that independently. Instance 1 trips its breaker after 5 failures. Instance 14 just got recycled and hasn't seen any yet. Instance 7 is in half-open, probing a service you already know is dead. For some window of time, part of your fleet is protecting itself and part of it is still hammering a dead dependency and timing out, and all you can do is watch.<p>Libraries can't fix this. Opossum, Resilience4j, Polly are great at the pattern, but they make per-instance decisions with per-instance state. Your circuit breakers don't talk to each other.<p>Openfuse is a centralized control plane. It aggregates failure metrics from every instance in your fleet and makes the trip decision based on the full picture. When the breaker opens, every instance knows at the same time.<p>It's a few lines of code:<p><pre><code> const result = await openfuse.breaker('stripe').protect( () => chargeCustomer(payload) ); </code></pre> The SDK is open source, anyone can see exactly what runs inside their services.<p>The other thing I couldn't let go of: when you get paged at 3am, you shouldn't have to find logs across 15 services to figure out what's broken. Openfuse gives you one dashboard showing every breaker state across your fleet: what's healthy, what's degraded, what tripped and when. And, you shouldn't need a deploy to act. You can open a breaker from the dashboard and every instance stops calling that dependency immediately. Planned maintenance window at 3am? Open beforehand. Fix confirmed? Close it instantly. Thresholds need adjusting? Change them in the dashboard, takes effect across your fleet in seconds. No PRs, no CI, no config files.<p>It has a decent free tier for trying it out, then $99/mo for most teams, $399/mo with higher throughput and some enterprise features. Solo founder, early stage, being upfront.<p>Would love to hear from people who've fought cascading failures in production. What am I missing?
Show HN: Beautiful interactive explainers generated with Claude Code
Show HN (score: 35)[Other] Show HN: Beautiful interactive explainers generated with Claude Code Hello HN,<p>Recently an amazingly beautiful explainer was shared on HN: <a href="https://explainers.blog/posts/why-is-the-sky-blue/" rel="nofollow">https://explainers.blog/posts/why-is-the-sky-blue/</a><p>I loved it so much that I wished more topics were explained that way. So, I decided to stress-test today's frontier models (Opus 4.6 in Claude Code) to generate similar explainer on any given topic WITH (almost) one shot and minimal nudging.<p>I'm launching with four topics: Fourier transformation, scaling laws in bio, cellular automata and LLMs.<p>I would let you be the judge, but I'm quite liking them.<p>Some things I learned:<p>- Prompting CC to test what it builds using headless chromium is essential - There are subtle bugs in explanations (like in one animation human lifespan is 40 years) - Asking CC to verify its plan via codex works really well<p>I do want to reiterate that the pages generated were mostly one-shot, which amazed me given how detailed the pages + animations are.
Terminals should generate the 256-color palette
Hacker News (score: 192)[Other] Terminals should generate the 256-color palette
Show HN: LaminarDB – Streaming SQL database in Rust, zero-alloc hot path
Show HN (score: 5)[Database] Show HN: LaminarDB – Streaming SQL database in Rust, zero-alloc hot path
Lessons learned from `oapi-codegen`'s time in the GitHub Secure Open Source Fund
Hacker News (score: 16)[Other] Lessons learned from `oapi-codegen`'s time in the GitHub Secure Open Source Fund
Show HN: Writing a C++20M:N Scheduler from Scratch (EBR, Work-Stealing)
Show HN (score: 14)[Other] Show HN: Writing a C++20M:N Scheduler from Scratch (EBR, Work-Stealing) tiny_coro is a lightweight, educational M:N asynchronous runtime written from scratch using C++20 coroutines. It's designed to strip away the complexity of industrial libraries (like Seastar or Folly) to show the core mechanics clearly.<p>Key Technical Features:<p>M:N Scheduling: Maps M coroutines to N kernel threads (Work-Stealing via Chase-Lev deque).<p>Memory Safety: Implements EBR (Epoch-Based Reclamation) to manage memory safely in lock-free structures without GC.<p>Visualizations: I used Manim (the engine behind 3Blue1Brown) to create animations showing exactly how tasks are stolen and executed.<p>Why I built it: To bridge the gap between "using coroutines" and "understanding the runtime." The code is kept minimal (~1k LOC core) so it can be read in a weekend.
BarraCUDA Open-source CUDA compiler targeting AMD GPUs
Hacker News (score: 192)[Other] BarraCUDA Open-source CUDA compiler targeting AMD GPUs
Show HN: Pg-typesafe – Strongly typed queries for PostgreSQL and TypeScript
Hacker News (score: 52)[Database] Show HN: Pg-typesafe – Strongly typed queries for PostgreSQL and TypeScript Throughout my career, I tried many tools to query PostgreSQL, and in the end, concluded that for what I do, the simplest is almost always the best: raw SQL queries.<p>Until now, I typed the results manually and relied on tests to catch problems. While this is OK in e.g., GoLang, it is quite annoying in TypeScript. First, because of the more powerful type system (it's easier to guess that updated_at is a date than it is to guess whether it's nullable or not), second, because of idiosyncrasies (INT4s are deserialised as JS numbers, but INT8s are deserialised as strings).<p>So I wrote pg-typesafe, with the goal of it being the less burdensome: you call queries exactly the same way as you would call node-pg, and they are fully typed.<p>It's very new, but I'm already using it in a large-ish project, where it found several bugs and footguns, and also allowed me to remove many manual type definitions.
Show HN: Continue – Source-controlled AI checks, enforceable in CI
Show HN (score: 39)[Code Quality] Show HN: Continue – Source-controlled AI checks, enforceable in CI We now write most of our code with agents. For a while, PRs piled up, causing review fatigue, and we had this sinking feeling that standards were slipping. Consistency is tough at this volume. I’m sharing the solution we found, which has become our main product.<p>Continue (<a href="https://docs.continue.dev">https://docs.continue.dev</a>) runs AI checks on every PR. Each check is a source-controlled markdown file in `.continue/checks/` that shows up as a GitHub status check. They run as full agents, not just reading the diff, but able to read/write files, run bash commands, and use a browser. If it finds something, the check fails with one click to accept a diff. Otherwise, it passes silently.<p>Here’s one of ours:<p><pre><code> .continue/checks/metrics-integrity.md --- name: Metrics Integrity description: Detects changes that could inflate, deflate, or corrupt metrics (session counts, event accuracy, etc.) --- Review this PR for changes that could unintentionally distort metrics. These bugs are insidious because they corrupt dashboards without triggering errors or test failures. Check for: - "Find or create" patterns where the "find" is too narrow, causing entity duplication (e.g. querying only active sessions, missing completed ones, so every new commit creates a duplicate) - Event tracking calls inside loops or retry paths that fire multiple times per logical action - Refactors that accidentally remove or move tracking calls to a path that executes with different frequency Key files: anything containing `posthog.capture` or `trackEvent` </code></pre> This check passed without noise for weeks, but then caught a PR that would have silently deflated our session counts. We added it in the first place because we’d been burned in the past by bad data, only noticing when a dashboard looked off.<p>---<p>To get started, paste this into Claude Code or your coding agent of choice:<p><pre><code> Help me write checks for this codebase: https://continue.dev/walkthrough </code></pre> It will:<p>- Explore the codebase and use the `gh` CLI to read past review comments<p>- Write checks to `.continue/checks/`<p>- Optionally, show you how to run them locally or in CI<p>Would love your feedback!
Launch HN: Sonarly (YC W26) – AI agent to triage and fix your production alerts
Hacker News (score: 26)[Monitoring/Observability] Launch HN: Sonarly (YC W26) – AI agent to triage and fix your production alerts Hey HN, I am Dimittri and we’re building Sonarly (<a href="https://sonarly.com">https://sonarly.com</a>), an AI engineer for production. It connects to your observability tools like Sentry, Datadog, or user feedback channels, triages issues, and fixes them to cut your resolution time. Here's a demo: <a href="https://www.youtube.com/watch?v=rr3VHv0eRdw" rel="nofollow">https://www.youtube.com/watch?v=rr3VHv0eRdw</a>.<p>Sonarly is really about removing the noise from production alerts by grouping duplicates and returning a root cause analysis to save time to on-call engineers and literally cut your MTTR.<p>Before starting this company, my co-founder and I had a B2C app in edtech and had, some days, thousands of users using the app. We pushed several times a day, relying on user feedback. Then we set up Sentry, it was catching a lot of bugs, but we had up to 50 alerts a day. With 2 people it's a lot. We took a lot of time filtering the noise to find the real signal so we knew which bug to focus on.<p>At the same time, we saw how important it is to fix a bug fast when it hits users. A bug means in the worst case a churn and at best a frustrated user. And there are always bugs in production, due to code errors, database mismatches, infrastructure overload, and many issues are linked to a specific user behavior. You can't catch all these beforehand, even with E2E tests or AI code reviews (which catch a lot of bugs but obviously not all, plus it takes time to run at each deployment). This is even more true with vibe-coding (or agentic engineering).<p>We started Sonarly with this idea. More software than ever is being built and users should have the best experience possible on every product. The main idea of Sonarly is to reduce the MTTR (Mean Time To Repair).<p>We started by recreating a Sentry-like tool but without the noise, using only text and session replays as the interface. We built our own frontend tracker (based on open-source rrweb) and used the backend Sentry SDK (open source as well). Companies could just add another tracker in the frontend and add a DSN in their Sentry config to send data to us in addition to Sentry.<p>We wanted to build an interface where you don't need to check logs, dashboards, traces, metrics, and code, as the agent would do it for you with plain English to explain the "what," "why," and "how do I fix it."<p>We quickly realized companies don't want to add a new tracker or change their monitoring stack, as these platforms do the job they're supposed to do. So we decided to build above them. Now we connect to tools like Sentry, Datadog, Slack user feedback channels, and other integrations.<p>Claude Code is so good at writing code, but handling runtime issues requires more than just raw coding ability. It demands deep runtime context, immediate reactivity, and intelligent triage, you can’t simply pipe every alert directly into an agent. That’s why our first step is converting noise into signal. We group duplicates and filter false positives to isolate clear issues. Once we have a confirmed signal, we trigger Claude Code with the exact context it needs, like the specific Sentry issue and relevant logs fetched via MCP (mostly using grep on Datadog/Grafana). However, things get exponentially harder with multi-repo and multi-service architectures.<p>So we built an internal map of the production system that is basically a .md file updated dynamically. It shows every link between different services, logs, and metrics so that Claude Code can understand the issue faster.<p>One of our users using Sentry was receiving ~180 alerts/day. Here is what their workflow looked like:<p>- Receive the alert<p>- 1) Defocus from their current task or wake up, or 2) don't look at the alert at all (most of the time)<p>- Go check dashboards to find the root cause (if infra type) or read the stack trace, events, etc.<p>- Try to figure out if it was a false positive or a real problem (or a known problem already in the fixes pipeline)<p>- Then fix by giving Claude Code the correct context<p>We started by cutting the noise and went from 180/day to 50/day (by grouping issues) and giving a severity based on the impact on the user/infra. This brings it down to 5 issues to focus on in the current day. Triage happens in 3 steps: deduplicating before triggering a coding agent, gathering the root cause for each alert, and re-grouping by RCA.<p>We launched self-serve (<a href="https://sonarly.com">https://sonarly.com</a>) and we would love to have feedback from engineers. Especially curious about your current workflows when you receive an alert from any of these channels like Sentry (error tracking), Datadog (APM), or user feedback. How do you assign who should fix it? Where do you take your context from to fix the issue? Do you have any automated workflow to fix every bug, and do you have anything you use currently to filter the noise from alerts?<p>We have a large free tier as we mainly want feedback. You can self-serve under 2 min. I'll be in the thread with my co-founder to answer your questions, give more technical details, and take your feedback: positive, negative, brutal, everything's constructive!
Using go fix to modernize Go code
Hacker News (score: 302)[Other] Using go fix to modernize Go code
Show HN: I taught LLMs to play Magic: The Gathering against each other
Hacker News (score: 107)[Other] Show HN: I taught LLMs to play Magic: The Gathering against each other I've been teaching LLMs to play Magic: The Gathering recently, via MCP tools hooked up to the open-source XMage codebase. It's still pretty buggy and I think there's significant room for existing models to get better at it via tooling improvements, but it pretty much works today. The ratings for expensive frontier models are artificially low right now because I've been focusing on cheaper models until I work out the bugs, so they don't have a lot of games in the system.
SwiftForth IDE for Windows, Linux, macOS
Hacker News (score: 18)[IDE/Editor] SwiftForth IDE for Windows, Linux, macOS
Show HN: Telescope now queries Kubernetes logs directly
Show HN (score: 6)[Monitoring/Observability] Show HN: Telescope now queries Kubernetes logs directly Telescope originally started as a ClickHouse-focused log viewer.<p>But while building it, I kept running into the same pattern: when everything worked, logs were in ClickHouse. When things broke, logs were still inside Kubernetes.<p>That gap led to adding Kubernetes as a native log source.<p>This is not meant to replace proper log aggregation. Centralized storage with indexing and retention policies is still the right approach for production.<p>But there are situations where aggregation doesn't help: the logging pipeline is broken, logs are delayed, or you're debugging locally and don't have a pipeline at all.<p>In those cases, the logs are already in the pods. The usual fallback is kubectl logs (or stern), often across multiple terminals and namespaces. It works, but correlation becomes manual.<p>Telescope can now query logs directly from Kubernetes clusters via the Kubernetes API. It lets you query across multiple namespaces and clusters, filter by labels and fields, apply time ranges, normalize severity across different log formats, and visualize log volume over time.<p>It uses your existing kubeconfig, fetches logs in parallel (configurable concurrency), and uses time filters to limit data transfer from Kubernetes APIs.<p>No agents. No CRDs. No cluster changes.<p>Current limitations: no streaming / follow mode yet.<p>Curious if others have run into the same "pipeline gap" problem - when logs aren't in your backend yet, but you still need structured access to them.<p>GitHub: <a href="https://github.com/iamtelescope/telescope" rel="nofollow">https://github.com/iamtelescope/telescope</a><p>Changelog: <a href="https://docs.iamtelescope.net/changelog" rel="nofollow">https://docs.iamtelescope.net/changelog</a>
Fff.nvim – Typo-resistant code search
Hacker News (score: 24)[Code Quality] Fff.nvim – Typo-resistant code search The new release of fff.nvim showed out the new compatiblity for the code search both for human and agents - typo resistant and usable code search for real codebases<p>It can match query "shcema" <i>only</i> to "schema" and "SortedMap" only to "SortedArrayMap" and "SortedHashMap" without bloating the results
Show HN: Breadboard – A modern HyperCard for building web apps on the canvas
Hacker News (score: 35)[Other] Show HN: Breadboard – A modern HyperCard for building web apps on the canvas Hey HN! I’m Simone. We re-built Breadboard, a visual app builder that mixes Figma-style UI design with Shortcuts-style logic so you can build, preview, and publish interactive web apps directly from the canvas.<p>What it does<p><pre><code> Design UIs visually with a flexible canvas –like Figma–. Define app logic with a visual, instruction-stacked editor inspired by Shortcuts. Live preview apps directly on the canvas –no separate preview window–. Publish working web apps with one click. </code></pre> Why we made it<p><pre><code> Modernize the HyperCard idea: combine layout, behavior, and instant sharing in one place. Reduce friction between design and a working app. Make simple web apps approachable for non-developers while keeping power features for developers. Build a foundation for LLM integration so users can design and develop with AI while still understanding what’s happening, even without coding experience –in progress!–. </code></pre> Try it –no signup required–<p>Weather forecast app: <a href="https://app.breadboards.io/playgrounds/weather" rel="nofollow">https://app.breadboards.io/playgrounds/weather</a><p>Swiss Public Transit: <a href="https://app.breadboards.io/playgrounds/public_transit" rel="nofollow">https://app.breadboards.io/playgrounds/public_transit</a><p>info: <a href="https://breadboards.io" rel="nofollow">https://breadboards.io</a><p>I would appreciate any feedback :)
Show HN: Simple org-mode web adapter
Show HN (score: 61)[Other] Show HN: Simple org-mode web adapter I like to use org files a lot, but I wanted some way to browse and edit them on my phone when I'm out. Yesterday I used Codex to make this simple one-file web server that just displays all my org files with backlinks. It doesn't have any authentication because I only run it on my wireguard VPN. I've been having fun with it, hopefully it's useful to someone else!