🛠️ All DevTools
Showing 1–20 of 4446 tools
Last Updated
May 03, 2026 at 08:00 AM
Kimi K2.6 just beat Claude, GPT-5.5, and Gemini in a coding challenge
Hacker News (score: 240)[Other] Kimi K2.6 just beat Claude, GPT-5.5, and Gemini in a coding challenge
Voice-AI-for-Beginners – A curated learning path for developers
Hacker News (score: 43)[Other] Voice-AI-for-Beginners – A curated learning path for developers
Show HN: State of the Art of Coding Models, According to Hacker News Commenters
Hacker News (score: 25)[Other] Show HN: State of the Art of Coding Models, According to Hacker News Commenters Hello HN,<p>I was away from my computer for two weeks, and after coming back and reading the latest discussions on HN about coding assistants (models, harnesses), I felt very out of the loop. My normal process would have been to keep reading and figure out the latest and greatest from people's comments, but I wanted to try and automate this process.<p>Basically the goal is to get a quick overview over which coding models are popular on HN. A next iteration could also scan for harnesses that people use, or info on self-hosting or hardware setups.<p>I wrote a short intro on the page about the pipeline that collects and analyzes the data, but feel free to ask for more details or check the Google Sheet for more info.<p><a href="https://hnup.date/hn-sota" rel="nofollow">https://hnup.date/hn-sota</a>
VS Code inserting 'Co-Authored-by Copilot' into commits regardless of usage
Hacker News (score: 576)[Other] VS Code inserting 'Co-Authored-by Copilot' into commits regardless of usage
Flue is a TypeScript framework for building the next generation of agents
Hacker News (score: 41)[Other] Flue is a TypeScript framework for building the next generation of agents
Welcome to Hell Developer
Hacker News (score: 11)[Other] Welcome to Hell Developer
Open Design: Use Your Coding Agent as a Design Engine
Hacker News (score: 141)[Other] Open Design: Use Your Coding Agent as a Design Engine
ShareX/ShareX
GitHub Trending[Other] ShareX is a free and open-source application that enables users to capture or record any area of their screen with a single keystroke. It also supports uploading images, text, and various file types to a wide range of destinations.
Show HN: Mljar Studio – local AI data analyst that saves analysis as notebooks
Hacker News (score: 42)[Other] Show HN: Mljar Studio – local AI data analyst that saves analysis as notebooks Hi HN,<p>I’ve been working on mljar-supervised (open-source AutoML for tabular data) for a few years. Recently I built a desktop app around it called MLJAR Studio.<p>The idea is simple: you talk to your data in natural language, the AI generates Python code, executes it locally, and the whole conversation becomes a reproducible notebook (*.ipynb file). So instead of just chatting with data, you end up with something you can inspect, modify, and rerun.<p>What MLJAR Studio does:<p>- Sets up a local Python environment automatically, runs on Mac, Windows, and Linux<p>- Installs missing packages during the conversation<p>- Built-in AutoML for tabular data (classification, regression, multiclass)<p>- Works with standard Python libraries (pandas, matplotlib, etc.)<p>- Works with any data file: CSV, Excel, Stata, Parquet ...<p>- Connects to PostgreSQL, MySQL, SQL Server, Snowflake, Databricks, and Supabase.<p>For AI: use Ollama locally (zero data egress), bring your own OpenAI key, or use MLJAR AI add-on.<p>I built this because I wanted something between Jupyter Notebook (flexible but manual) and AI tools that generate code but don’t preserve the workflow. Most tools I tried either hide too much or don’t give reproducible results and are cloud based<p>Demos:<p>- 60-second demo: <a href="https://youtu.be/BjxpZYRiY4c" rel="nofollow">https://youtu.be/BjxpZYRiY4c</a><p>- Full 3-minute analysis: <a href="https://youtu.be/1DHMMxaNJxI" rel="nofollow">https://youtu.be/1DHMMxaNJxI</a><p>Pricing is $199 one-time, with a 7-day trial.<p>Curious if this is useful for others doing real data work, or if I’m solving my own problem here.<p>Happy to answer questions.
Show HN: Filling PDF forms with AI using client-side tool calling
Hacker News (score: 19)[Other] Show HN: Filling PDF forms with AI using client-side tool calling Hey HN!<p>I built SimplePDF Copilot: an AI assistant that can interact with the PDF editor. It fills fields, answers questions, focuses on a specific field, adds fields, deletes pages, and so on.<p>It's built on top of SimplePDF that I started 7 years ago, pioneering privacy-respecting client-side pdf editing, now used monthly by 200k+ people.<p>As for the privacy model: the PDF itself never leaves the browser. Parsing, rendering, and field detection all run client-side.<p>The text the model needs (and your messages) goes to whatever LLM you point at. By default that's our demo proxy (DeepSeek V4 Flash, rate-capped), but you can BYOK and point it at any cloud provider, or go fully local (I've been testing with LM Studio).<p>Unlike the existing "Chat with PDF" tools that only retrieve the text/OCR layer, Copilot can act on the PDF: filling fields, adding fields (detected client-side using CommonForms by Joe Barrow [1], jbarrow on HN with some post-processing heuristics I added on top), focusing on fields, deleting pages, and so on.<p>I built this because SimplePDF is mostly used by healthcare customers where document privacy is paramount, and I wanted an AI experience that didn't require shipping PII to a third party. Stack is pretty standard:<p>- Tanstack Start<p>- AI SDK from Vercel<p>- Tailwind (I personally prefer CSS modules, I'm old-school but the goal since I open source it, I figured that Tailwind would be a better fit)<p>The more interesting part is the client-side tool calling: events are passed back and forth via iframe postMessage.<p>If you're not familiar with "tool calling" and "client-side tool calling", a quick primer:<p>Tool calling is what LLMs use to take actions. When Claude runs grep or ls, or hits an MCP server, those are tool calls.<p>Client-side tool calling means the intent to call a tool comes from the LLM, but the execution happens in the browser.<p>That matters for: speed, you can't go faster than client-to-client operations and also gives you the ability to limit the data you expose to the LLM. For the demo I do feed the content of the document to the LLM, but that connection could be severed as simply as removing the tool that exposes the content data.<p>The demo is fully open source, available on Github [2] and the demo is the same as the link of this post [3]<p>What's not open source is SimplePDF itself (loaded as the iframe).<p>I could talk on and on about this, let me know if you have any questions, anything goes!<p>[1] <a href="https://github.com/jbarrow/commonforms" rel="nofollow">https://github.com/jbarrow/commonforms</a><p>[2] <a href="https://github.com/SimplePDF/simplepdf-embed/tree/main/copilot" rel="nofollow">https://github.com/SimplePDF/simplepdf-embed/tree/main/copil...</a><p>[3] <a href="https://copilot.simplepdf.com/?share=a7d00ad073c75a75d493228e6ff7b11eb3f2d945b6175913e87898ec96ca8076&form=w9&lang=en" rel="nofollow">https://copilot.simplepdf.com/?share=a7d00ad073c75a75d493228...</a>
K3k: Kubernetes in Kubernetes
Hacker News (score: 34)[DevOps] K3k: Kubernetes in Kubernetes
Create an MP4 video of a web page scrolling at a steady speed
Hacker News (score: 23)[Other] Create an MP4 video of a web page scrolling at a steady speed
Show HN: Agent-desktop – Native desktop automation CLI for AI agents
Hacker News (score: 87)[CLI Tool] Show HN: Agent-desktop – Native desktop automation CLI for AI agents I've been building computer-use tools for a while, and I quietly launched this about a month ago (122 Stars on GH). I figured it was worth sharing here.<p>Over the last few months, a lot of computer-use agents have come out: Codex, Claude Code, CUA, and others. Most of them seem to work roughly like this: 1. Take a screenshot 2. Have the model predict pixel coordinates 3. Click x,y 4. Take another screenshot 5. Repeat<p>That works, but it's slow, expensive in tokens, and fragile. If the UI shifts a few pixels, things break. And the model still doesn't know what any element actually is.<p>But the OS already exposes structured UI information:<p><pre><code> - macOS: Accessibility API - Windows: UI Automation - Linux: AT-SPI </code></pre> Screen readers have used these APIs for years. On the web, Playwright beat screenshot scraping for the same reason: structured access is just a better abstraction than pixels.<p>So I built a desktop equivalent: agent-desktop.<p>It's a cross-platform CLI for structured desktop automation through the accessibility tree. One Rust binary, about 15 MB, no runtime dependencies. It exposes 53 commands with JSON output, so an LLM can inspect and operate native apps without screenshots or vision models. Inspired by agent-browser by Vercel Labs.<p>A typical loop looks like this:<p><pre><code> agent-desktop snapshot --app Slack -i --compact agent-desktop click @e12 agent-desktop type @e5 "ship it" agent-desktop press cmd+return </code></pre> So the loop becomes:<p><pre><code> 1. Snapshot 2. Decide 3. Act 4. Snapshot again </code></pre> The main design problem was context size.<p>A naive approach would dump the full accessibility tree into the model, but real apps get huge. Slack can easily exceed 50,000 tokens for a full tree dump, which makes the approach impractical.<p>The approach I ended up using is progressive skeleton traversal:<p><pre><code> - First pass: return a shallow tree, typically depth 3, with deeper containers truncated and annotated with children_count - Named containers get references so the agent can request only that subtree - The agent drills down into the relevant region with --root @e3 - References are scoped and invalidated only for that subtree - After acting, the agent can re-query just that region instead of re-snapshotting the whole app </code></pre> In practice, this reduced token usage by about 78% to 96% versus full-tree dumps in Electron apps like Slack, VS Code, and Notion.<p>A few implementation details that may be interesting here:<p><pre><code> - Rust workspace with strict platform/core separation through a PlatformAdapter trait - Accessibility-first activation chain; mouse synthesis is the fallback, not the default - Deterministic element refs like @e1, @e2, with optimistic re-identification across UI shifts - Structured errors with machine-readable codes plus retry suggestions - C ABI via cdylib, so it can be loaded directly from Python, Swift, Go, Node, Ruby, or C without shelling out - Batch operations in a single call - Support for windows, menus, sheets, popovers, alerts, and notifications - Special handling for Chromium/Electron accessibility trees, which can get very deep and noisy </code></pre> Why I think this matters: pixel-based desktop control feels like a leaky abstraction. The OS already knows the UI semantically. Accessibility APIs give you roles, names, actions, hierarchy, focus, selection, and state directly. That seems like a much better substrate for desktop agents than screenshot loops.<p>If you're building your own desktop agent, internal automation tool, or research prototype, this may be useful.<p>Install:<p><pre><code> npm install -g agent-desktop agent-desktop snapshot --app Finder -i </code></pre> Repo: <a href="https://github.com/lahfir/agent-desktop" rel="nofollow">https://github.com/lahfir/agent-desktop</a><p>I'd especially love feedback from people who've built desktop automation before. What are the biggest pain points you've run into, and what would you want a tool like this to support?
Lib0xc: A set of C standard library-adjacent APIs for safer systems programming
Hacker News (score: 48)[API/SDK] Lib0xc: A set of C standard library-adjacent APIs for safer systems programming
Show HN: Omar – A TUI for managing 100 coding agents
Show HN (score: 10)[Other] Show HN: Omar – A TUI for managing 100 coding agents We were both genuinely impressed by Claude Code after it helped each of us fix nasty CI problems overnight. Doing those fixes manually would have taken days.<p>After that experience, we each found ourselves struggling through Ctrl+Tab through multiple Claude Code windows in our terminals. While we enjoyed having agents working for us in parallel, context switching and cycling through each terminal tab was a real pain.<p>So we thought: Can we design a TUI dashboard that manages a large swarm of agents in one place? Even better, can agents manage agents hierarchically, like how companies work?<p>OMAR (Open Multi-Agent Runtime) is the result of this exploration. We spent months building it, and we think it is now ready to show the world. If you find OMAR interesting, give it a try. We would love to hear from you. :)<p>Check out our blog here for more details: <a href="https://omar.tech/blog/introducing-omar/" rel="nofollow">https://omar.tech/blog/introducing-omar/</a><p>Thanks! Karim & Shaokai
Show HN: AI CAD Harness
Hacker News (score: 20)[Other] Show HN: AI CAD Harness Hi HN, I'm Zach, one of the co-founders of Adam (<a href="https://adam.new">https://adam.new</a>).<p>We've been on HN twice before with text-to-CAD/3D experiments [1][2]. The honest takeaway from those threads: prompt-to-3D model web apps are fun, but serious mechanical engineers don't want a black box that spits out an STL. They want help inside the CAD tool they already use, with full visibility and control over the feature tree.<p>So we built that. Adam is now a harness that integrates directly with your CAD. It reads your parts, understands the existing feature tree, and edits it for you agentically. We are now live in beta on Onshape and Fusion! [3]:<p>Install link Autodesk Fusion: <a href="https://fusion.adam.new/install">https://fusion.adam.new/install</a><p>Install link PTC Onshape: <a href="https://cad.onshape.com/appstore/apps/Design%20&%20Documentation/690a8dc864e816c112aa66a0" rel="nofollow">https://cad.onshape.com/appstore/apps/Design%20&%20Documenta...</a><p>Things people are using it for today: - "Merge redundant features and clean up my tree" - "Rename every feature so the tree is actually readable" - "Round all internal edges with a 2mm fillet" - “Parametrize my model”<p>Along with of course, using Adam to generate CAD end-to-end!<p>A few things we care about that aren't obvious from the listing:<p>1. From the start we have always believed in CAD as code as the right abstraction. Our harness leverages Onshape's FeatureScript and Python in Fusion heavily.<p>2. We run an internal CAD benchmark across frontier models. There has been a massive jump in the spatial reasoning capabilities of recently released models. Particularly GPT 5.5 and Opus 4.7 [4] [5]<p>3. We open-sourced our earlier text-to-CAD work [6]<p>A note on the Anthropic Autodesk connector that shipped a couple days ago [7]: We think it's great for the space and validates the direction.<p>Where Adam is different:<p>- Model-agnostic. We pick whichever frontier model is winning on each task type from our own internal bench, instead of being tied to one lab.<p>- We live natively in your CAD apps and are actively building integrations across all programs<p>What would you want an in-CAD agent to do that nothing does today?<p>[1] <a href="https://news.ycombinator.com/item?id=44182206">https://news.ycombinator.com/item?id=44182206</a><p>[2] <a href="https://news.ycombinator.com/item?id=45140921">https://news.ycombinator.com/item?id=45140921</a><p>[3] <a href="https://x.com/adamdotnew/status/2050264512230719980?s=20" rel="nofollow">https://x.com/adamdotnew/status/2050264512230719980?s=20</a><p>[4] <a href="https://x.com/adamdotnew/status/2044859329329893376?s=20" rel="nofollow">https://x.com/adamdotnew/status/2044859329329893376?s=20</a><p>[5] <a href="https://x.com/adamdotnew/status/2047795078912172122?s=20" rel="nofollow">https://x.com/adamdotnew/status/2047795078912172122?s=20</a><p>[6] <a href="https://github.com/Adam-CAD/CADAM" rel="nofollow">https://github.com/Adam-CAD/CADAM</a><p>[7] <a href="https://x.com/claudeai/status/2049143440508616863?s=20" rel="nofollow">https://x.com/claudeai/status/2049143440508616863?s=20</a>
Show HN: My Private GitHub on Postgres
Hacker News (score: 26)[Other] Show HN: My Private GitHub on Postgres
Show HN: GhostBox – Borrow a disposable little machine from the Global Free Tier
Hacker News (score: 103)[CLI Tool] Show HN: GhostBox – Borrow a disposable little machine from the Global Free Tier I built this because I was always creating machines on GH actions to test builds on different OS, and I wanted a tight CLI that could do it. I always saw Actions as this great resources and ephemeral machines you could do dev work in just were a natural way for me to work, so this grew out of that workflow.<p>I didn't expect it to blow up, so it wasn't 100% finished when I posted it. But it should stabilize pretty quickly.<p>Happy to know what you think and talk about it.
Running Adobe's 1991 PostScript Interpreter in the Browser
Hacker News (score: 106)[Other] Running Adobe's 1991 PostScript Interpreter in the Browser
Show HN: Loopsy, a way for terminals and AI agents on different machines to talk
Hacker News (score: 35)[Other] Show HN: Loopsy, a way for terminals and AI agents on different machines to talk I've always had the urge to have my two macbooks communicate. Having one idle while working on the other felt like underutilization of resources. So I built Loopsy. Initially the goal was to do file transfer via local network, and then came running commands. I then tried running coding agents from one machine to the other, and it worked.<p>Later I figured there should be a better way to continue my claude sessions remotely on my phone from the gym. So I did a cloudflare worker that connects to my local machine. I just need to ensure the laptop is plugged in.<p>I know I might be reinventing the wheel, but I love that it just works. Still working on E2E encryption. iOS app still in review.<p>Lemme know your thoughts.