🛠️ All DevTools

Showing 1–20 of 4080 tools

Last Updated
April 08, 2026 at 04:00 PM

[Other] Show HN: We fingerprinted 178 AI models' writing styles and similarity clusters We have a dataset of 3,095 standardized AI responses across 43 prompts. From each response, we extract a 32-dimension stylometric fingerprint (lexical richness, sentence structure, punctuation habits, formatting patterns, discourse markers).<p>Some findings:<p>- 9 clone clusters (&gt;90% cosine similarity on z-normalized feature vectors) - Mistral Large 2 and Large 3 2512 score 84.8% on a composite metric combining 5 independent signals - Gemini 2.5 Flash Lite writes 78% like Claude 3 Opus. Costs 185x less - Meta has the strongest provider &quot;house style&quot; (37.5x distinctiveness ratio) - &quot;Satirical fake news&quot; is the prompt that causes the most writing convergence across all models - &quot;Count letters&quot; causes the most divergence<p>The composite clone score combines: prompt-controlled head-to-head similarity, per-feature Pearson correlation across challenges, response length correlation, cross-prompt consistency, and aggregate cosine similarity.<p>Tech: stylometric extraction in Node.js, z-score normalization, cosine similarity for aggregate, Pearson correlation for per-feature tracking. Analysis script is ~1400 lines.

Found: April 08, 2026 ID: 4080

newton-physics/newton

GitHub Trending

[Other] An open-source, GPU-accelerated physics simulation engine built upon NVIDIA Warp, specifically targeting roboticists and simulation researchers.

Found: April 08, 2026 ID: 4078

[Other] The Git Commands I Run Before Reading Any Code

Found: April 08, 2026 ID: 4079

[Other] We moved Railway's frontend off Next.js. Builds went from 10+ mins to under two

Found: April 08, 2026 ID: 4076

[Other] Xilem – An experimental Rust native UI framework

Found: April 07, 2026 ID: 4074

[Other] Show HN: Mo – checks GitHub PRs against decisions approved in Slack Built this after a recurring frustration at our agency: the team would agree on something in Slack (&quot;only admins can export users&quot;), someone would open a PR two weeks later that quietly broke it, and nobody caught it until QA or after deploy.<p>Mo watches a Slack channel for decisions. When someone tags @mo to approve something, it stores it. When a PR opens, Mo checks the diff against the approved decisions and flags conflicts before merge.<p>It doesn&#x27;t review code quality. It only cares whether the code matches what the team actually agreed to.<p>Would love feedback, especially from anyone who&#x27;s been burned by this exact problem.<p>Try it here: <a href="https:&#x2F;&#x2F;hey-mo.io" rel="nofollow">https:&#x2F;&#x2F;hey-mo.io</a>

Found: April 07, 2026 ID: 4077

S3 Files

Hacker News (score: 164)

[Other] S3 Files <a href="https:&#x2F;&#x2F;aws.amazon.com&#x2F;blogs&#x2F;aws&#x2F;launching-s3-files-making-s3-buckets-accessible-as-file-systems&#x2F;" rel="nofollow">https:&#x2F;&#x2F;aws.amazon.com&#x2F;blogs&#x2F;aws&#x2F;launching-s3-files-making-s...</a>

Found: April 07, 2026 ID: 4072

[Other] Show HN: Gemma 4 Multimodal Fine-Tuner for Apple Silicon About six months ago, I started working on a project to fine-tune Whisper locally on my M2 Ultra Mac Studio with a limited compute budget. I got into it. The problem I had at the time was I had 15,000 hours of audio data in Google Cloud Storage, and there was no way I could fit all the audio onto my local machine, so I built a system to stream data from my GCS to my machine during training.<p>Gemma 3n came out, so I added that. Kinda went nuts, tbh.<p>Then I put it on the shelf.<p>When Gemma 4 came out a few days ago, I dusted it off, cleaned it up, broke out the Gemma part from the Whisper fine-tuning and added support for Gemma 4.<p>I&#x27;m presenting it for you here today to play with, fork and improve upon.<p>One thing I have learned so far: It&#x27;s very easy to OOM when you fine-tune on longer sequences! My local Mac Studio has 64GB RAM, so I run out of memory constantly.<p>Anywho, given how much interest there is in Gemma 4, and frankly, the fact that you can&#x27;t really do audio fine-tuning with MLX, that&#x27;s really the reason this exists (in addition to my personal interest). I would have preferred to use MLX and not have had to make this, but here we are. Welcome to my little side quest.<p>And so I made this. I hope you have as much fun using it as I had fun making it.<p>-Matt

Found: April 07, 2026 ID: 4066

[Other] Tailslayer: Library for reducing tail latency in RAM reads

Found: April 07, 2026 ID: 4073

[Other] Show HN: Marimo pair – Reactive Python notebooks as environments for agents Hi HN! We&#x27;re excited to share marimo pair [1] [2], a toolkit that drops AI agents into a running marimo notebook [3] session. This lets agents use marimo as working memory and a reactive Python runtime, while also making it easy for humans and agents to collaborate on computational research and data work.<p>GitHub repo: <a href="https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo-pair" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo-pair</a><p>Demo: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=6uaqtchDnoc" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=6uaqtchDnoc</a><p>marimo pair is implemented as an agent skill. Connect your agent of choice to a running notebook with:<p>&#x2F;marimo-pair pair with me on my_notebook.py<p>The agent can do anything a human can do with marimo and more. For example, it can obtain feedback by running code in an ephemeral scratchpad (inspect variables, run code against the program state, read outputs). If it wants to persist state, the agent can add cells, delete them, and install packages (marimo records these actions in the associated notebook, which is just a Python file). The agent can even manipulate marimo&#x27;s user interface — for fun, try asking your agent to greet you from within a pair session.<p>The agent effects all actions by running Python code in the marimo kernel. Under the hood, the marimo pair skill explains how to discover and create marimo sessions, and how to control them using a semi-private interface we call code mode.<p>Code mode lets models treat marimo as a REPL that extends their context windows, similar to recursive language models (RLMs). But unlike traditional REPLs, the marimo &quot;REPL&quot; incrementally builds a reproducible Python program, because marimo notebooks are dataflow graphs with well-defined execution semantics. As it uses code mode, the agent is kept on track by marimo&#x27;s guardrails, which include the elimination of hidden state: run a cell and dependent cells are run automatically, delete a cell and its variables are scrubbed from memory.<p>By giving models full control over a stateful reactive programming environment, rather than a collection of ephemeral scripts, marimo pair makes agents active participants in research and data work. In our early experimentation [4], we&#x27;ve found that marimo pair accelerates data exploration, makes it easy to steer agents while testing research hypotheses, and can serve as a backend for RLMs, yielding a notebook as an executable trace of how the model answered a query. We even use marimo pair to find and fix bugs in itself and marimo [5]. In these examples the notebook is not only a computational substrate but also a canvas for collaboration between humans and agents, and an executable, literate artifact comprised of prose, code, and visuals.<p>marimo pair is early and experimental. We would love your thoughts.<p>[1] <a href="https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo-pair" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo-pair</a><p>[2] <a href="https:&#x2F;&#x2F;marimo.io&#x2F;blog&#x2F;marimo-pair" rel="nofollow">https:&#x2F;&#x2F;marimo.io&#x2F;blog&#x2F;marimo-pair</a><p>[3] <a href="https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo</a><p>[4] <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=VKvjPJeNRPk" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=VKvjPJeNRPk</a><p>[5] <a href="https:&#x2F;&#x2F;github.com&#x2F;manzt&#x2F;dotfiles&#x2F;blob&#x2F;main&#x2F;.claude&#x2F;skills&#x2F;marimo-dev&#x2F;SKILL.md" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;manzt&#x2F;dotfiles&#x2F;blob&#x2F;main&#x2F;.claude&#x2F;skills&#x2F;m...</a>

Found: April 07, 2026 ID: 4069

[Other] Show HN: Output.ai - OSS framework we extracted from 500+ production AI agents

Found: April 07, 2026 ID: 4071

[Testing] Show HN: Finalrun – Spec-driven testing using English and vision for mobile apps I wanted to test mobile apps in plain English instead of relying on brittle selectors like XPath or accessibility IDs.<p>With a vision-based agent, that part actually works well. It can look at the screen, understand intent, and perform actions across Android and iOS.<p>The bigger problem showed up around how tests are defined and maintained.<p>When test flows are kept outside the codebase (written manually or generated from PRDs), they quickly go out of sync with the app. Keeping them updated becomes a lot of effort, and they lose reliability over time.<p>I then tried generating tests directly from the codebase (via MCP). That improved sync, but introduced high token usage and slower generation.<p>The shift for me was realizing test generation shouldn’t be a one-off step. Tests need to live alongside the codebase so they stay in sync and have more context.<p>I kept the execution vision-based (no brittle selectors), but moved test generation closer to the repo.<p>I’ve open sourced the core pieces:<p>1. generate tests from codebase context 2. YAML-based test flows 3. Vision-based execution across Android and iOS<p>Repo: <a href="https:&#x2F;&#x2F;github.com&#x2F;final-run&#x2F;finalrun-agent" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;final-run&#x2F;finalrun-agent</a> Demo: <a href="https:&#x2F;&#x2F;youtu.be&#x2F;rJCw3p0PHr4" rel="nofollow">https:&#x2F;&#x2F;youtu.be&#x2F;rJCw3p0PHr4</a><p>In the Demo video, you’ll see the &quot;post-development hand-off.&quot; An AI builds a feature in an IDE, and Finalrun immediately generates and executes a vision-based test for it verifying the feature developed by AI.

Found: April 07, 2026 ID: 4068

[DevOps] Google open-sources experimental agent orchestration testbed Scion <a href="https:&#x2F;&#x2F;googlecloudplatform.github.io&#x2F;scion&#x2F;overview&#x2F;" rel="nofollow">https:&#x2F;&#x2F;googlecloudplatform.github.io&#x2F;scion&#x2F;overview&#x2F;</a>

Found: April 07, 2026 ID: 4067

Show HN: Unicode Steganography

Hacker News (score: 21)

[Other] Show HN: Unicode Steganography I built a demo of two Unicode steganography techniques, zero-width characters and homoglyph substitution, in the context of AI misalignment.<p>The first is about the use of two invisible zero-width characters (ZWS and ZWNJ) to binary encode text.<p>The second is much cooler. Most characters in the Latin and Cyrillic alphabets look nearly identical, but have different unicode. If you have text to encode and convert it into binary representation (1s and 0s), you could take plain english &quot;carrier&quot; text and for each 1 in the binary representation you could substitute the Cyrillic letter equivalent. Decoding the message requires traversing the text and seeing where Cyrillic letters could have been substituted but weren&#x27;t, and where they were, leading to 0s and 1s respectively, which can be built back into your original hidden text.<p>In both cases, these are detectable, but the interesting problem for me is whether an LLM could eventually invent an encoding that goes unnoticed by both us, and automated detection.<p>If LLMs were able to covertly include messages in plaintext, misaligned AI Agents could eventually communicate across MCP&#x2F;A2A and individual chat session boundaries undetected. A deceptive LLM might seem helpful, but work against your goals. It could tell other agents it interacts with over MCP&#x2F;A2A to help it discreetly fail, signal intent, and avoid tripping oversight&#x2F;safety mechanisms. Furthermore, oversight mechanisms become more difficult to implement if we can&#x27;t believe our own eyes.

Found: April 07, 2026 ID: 4075

[CLI Tool] Show HN: td – a CLI to manage tasks, sessions, and worktrees for agentic coding Hi everyone! I built this because I wanted a little bit more organization around my Claude sessions, worktrees and plans while staying in the terminal and not relying on another SaaS tool. Since it&#x27;s a command line tool, the added bonus is that Claude can use `td` directly. The td calendar was just a fun add-on but the Claude session stats have been pretty interesting! Let me know what you think!

Found: April 07, 2026 ID: 4065

[Other] Show HN: Stop paying for Dropbox/Google Drive, use your own S3 bucket instead Last week SWYX nerd-sniped me into building an Open-source Dropbox.<p>Here is Locker: the ultimate open-source Google Drive&#x2F;box&#x2F;Dropbox alternative - Provider agnostic (S3, R2, vercel blob, local) - BYOB (Bring your own bucket) - Virtual file system - QMD Search plugin

Found: April 07, 2026 ID: 4063

[Other] We found an undocumented bug in the Apollo 11 guidance computer code

Found: April 07, 2026 ID: 4064

tobi/qmd

GitHub Trending

[CLI Tool] mini cli search engine for your docs, knowledge bases, meeting notes, whatever. Tracking current sota approaches while being all local

Found: April 07, 2026 ID: 4059

[Other] Show HN: Hippo, biologically inspired memory for AI agents

Found: April 06, 2026 ID: 4070

[Other] Show HN: Ghost Pepper – Local hold-to-talk speech-to-text for macOS I built this because I wanted to see how far I could get with a voice-to-text app that used 100% local models so no data left my computer. I&#x27;ve been using a ton for coding and emails. Experimenting with using it as a voice interface for my other agents too. 100% open-source MIT license, would love feedback, PRs, and ideas on where to take it.

Found: April 06, 2026 ID: 4060
Previous Page 1 of 204 Next