🛠️ Hacker News Tools

Showing 141–160 of 2449 tools from Hacker News

Last Updated
April 20, 2026 at 12:00 PM

S3 Files

Hacker News (score: 164)

[Other] S3 Files <a href="https:&#x2F;&#x2F;aws.amazon.com&#x2F;blogs&#x2F;aws&#x2F;launching-s3-files-making-s3-buckets-accessible-as-file-systems&#x2F;" rel="nofollow">https:&#x2F;&#x2F;aws.amazon.com&#x2F;blogs&#x2F;aws&#x2F;launching-s3-files-making-s...</a>

Found: April 07, 2026 ID: 4072

[Other] Show HN: Gemma 4 Multimodal Fine-Tuner for Apple Silicon About six months ago, I started working on a project to fine-tune Whisper locally on my M2 Ultra Mac Studio with a limited compute budget. I got into it. The problem I had at the time was I had 15,000 hours of audio data in Google Cloud Storage, and there was no way I could fit all the audio onto my local machine, so I built a system to stream data from my GCS to my machine during training.<p>Gemma 3n came out, so I added that. Kinda went nuts, tbh.<p>Then I put it on the shelf.<p>When Gemma 4 came out a few days ago, I dusted it off, cleaned it up, broke out the Gemma part from the Whisper fine-tuning and added support for Gemma 4.<p>I&#x27;m presenting it for you here today to play with, fork and improve upon.<p>One thing I have learned so far: It&#x27;s very easy to OOM when you fine-tune on longer sequences! My local Mac Studio has 64GB RAM, so I run out of memory constantly.<p>Anywho, given how much interest there is in Gemma 4, and frankly, the fact that you can&#x27;t really do audio fine-tuning with MLX, that&#x27;s really the reason this exists (in addition to my personal interest). I would have preferred to use MLX and not have had to make this, but here we are. Welcome to my little side quest.<p>And so I made this. I hope you have as much fun using it as I had fun making it.<p>-Matt

Found: April 07, 2026 ID: 4066

[Other] Building a framework-agnostic Ruby gem (and making sure it doesn't break)

Found: April 07, 2026 ID: 4091

[Other] Tailslayer: Library for reducing tail latency in RAM reads

Found: April 07, 2026 ID: 4073

[Other] Show HN: Marimo pair – Reactive Python notebooks as environments for agents Hi HN! We&#x27;re excited to share marimo pair [1] [2], a toolkit that drops AI agents into a running marimo notebook [3] session. This lets agents use marimo as working memory and a reactive Python runtime, while also making it easy for humans and agents to collaborate on computational research and data work.<p>GitHub repo: <a href="https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo-pair" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo-pair</a><p>Demo: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=6uaqtchDnoc" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=6uaqtchDnoc</a><p>marimo pair is implemented as an agent skill. Connect your agent of choice to a running notebook with:<p>&#x2F;marimo-pair pair with me on my_notebook.py<p>The agent can do anything a human can do with marimo and more. For example, it can obtain feedback by running code in an ephemeral scratchpad (inspect variables, run code against the program state, read outputs). If it wants to persist state, the agent can add cells, delete them, and install packages (marimo records these actions in the associated notebook, which is just a Python file). The agent can even manipulate marimo&#x27;s user interface — for fun, try asking your agent to greet you from within a pair session.<p>The agent effects all actions by running Python code in the marimo kernel. Under the hood, the marimo pair skill explains how to discover and create marimo sessions, and how to control them using a semi-private interface we call code mode.<p>Code mode lets models treat marimo as a REPL that extends their context windows, similar to recursive language models (RLMs). But unlike traditional REPLs, the marimo &quot;REPL&quot; incrementally builds a reproducible Python program, because marimo notebooks are dataflow graphs with well-defined execution semantics. As it uses code mode, the agent is kept on track by marimo&#x27;s guardrails, which include the elimination of hidden state: run a cell and dependent cells are run automatically, delete a cell and its variables are scrubbed from memory.<p>By giving models full control over a stateful reactive programming environment, rather than a collection of ephemeral scripts, marimo pair makes agents active participants in research and data work. In our early experimentation [4], we&#x27;ve found that marimo pair accelerates data exploration, makes it easy to steer agents while testing research hypotheses, and can serve as a backend for RLMs, yielding a notebook as an executable trace of how the model answered a query. We even use marimo pair to find and fix bugs in itself and marimo [5]. In these examples the notebook is not only a computational substrate but also a canvas for collaboration between humans and agents, and an executable, literate artifact comprised of prose, code, and visuals.<p>marimo pair is early and experimental. We would love your thoughts.<p>[1] <a href="https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo-pair" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo-pair</a><p>[2] <a href="https:&#x2F;&#x2F;marimo.io&#x2F;blog&#x2F;marimo-pair" rel="nofollow">https:&#x2F;&#x2F;marimo.io&#x2F;blog&#x2F;marimo-pair</a><p>[3] <a href="https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;marimo-team&#x2F;marimo</a><p>[4] <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=VKvjPJeNRPk" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=VKvjPJeNRPk</a><p>[5] <a href="https:&#x2F;&#x2F;github.com&#x2F;manzt&#x2F;dotfiles&#x2F;blob&#x2F;main&#x2F;.claude&#x2F;skills&#x2F;marimo-dev&#x2F;SKILL.md" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;manzt&#x2F;dotfiles&#x2F;blob&#x2F;main&#x2F;.claude&#x2F;skills&#x2F;m...</a>

Found: April 07, 2026 ID: 4069

[Other] Mysteries of Dropbox: Testing of a Distributed Sync Service (2016) [pdf]

Found: April 07, 2026 ID: 4112

[Other] Show HN: Output.ai - OSS framework we extracted from 500+ production AI agents

Found: April 07, 2026 ID: 4071

[Testing] Show HN: Finalrun – Spec-driven testing using English and vision for mobile apps I wanted to test mobile apps in plain English instead of relying on brittle selectors like XPath or accessibility IDs.<p>With a vision-based agent, that part actually works well. It can look at the screen, understand intent, and perform actions across Android and iOS.<p>The bigger problem showed up around how tests are defined and maintained.<p>When test flows are kept outside the codebase (written manually or generated from PRDs), they quickly go out of sync with the app. Keeping them updated becomes a lot of effort, and they lose reliability over time.<p>I then tried generating tests directly from the codebase (via MCP). That improved sync, but introduced high token usage and slower generation.<p>The shift for me was realizing test generation shouldn’t be a one-off step. Tests need to live alongside the codebase so they stay in sync and have more context.<p>I kept the execution vision-based (no brittle selectors), but moved test generation closer to the repo.<p>I’ve open sourced the core pieces:<p>1. generate tests from codebase context 2. YAML-based test flows 3. Vision-based execution across Android and iOS<p>Repo: <a href="https:&#x2F;&#x2F;github.com&#x2F;final-run&#x2F;finalrun-agent" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;final-run&#x2F;finalrun-agent</a> Demo: <a href="https:&#x2F;&#x2F;youtu.be&#x2F;rJCw3p0PHr4" rel="nofollow">https:&#x2F;&#x2F;youtu.be&#x2F;rJCw3p0PHr4</a><p>In the Demo video, you’ll see the &quot;post-development hand-off.&quot; An AI builds a feature in an IDE, and Finalrun immediately generates and executes a vision-based test for it verifying the feature developed by AI.

Found: April 07, 2026 ID: 4068

[DevOps] Google open-sources experimental agent orchestration testbed Scion <a href="https:&#x2F;&#x2F;googlecloudplatform.github.io&#x2F;scion&#x2F;overview&#x2F;" rel="nofollow">https:&#x2F;&#x2F;googlecloudplatform.github.io&#x2F;scion&#x2F;overview&#x2F;</a>

Found: April 07, 2026 ID: 4067

Show HN: Unicode Steganography

Hacker News (score: 21)

[Other] Show HN: Unicode Steganography I built a demo of two Unicode steganography techniques, zero-width characters and homoglyph substitution, in the context of AI misalignment.<p>The first is about the use of two invisible zero-width characters (ZWS and ZWNJ) to binary encode text.<p>The second is much cooler. Most characters in the Latin and Cyrillic alphabets look nearly identical, but have different unicode. If you have text to encode and convert it into binary representation (1s and 0s), you could take plain english &quot;carrier&quot; text and for each 1 in the binary representation you could substitute the Cyrillic letter equivalent. Decoding the message requires traversing the text and seeing where Cyrillic letters could have been substituted but weren&#x27;t, and where they were, leading to 0s and 1s respectively, which can be built back into your original hidden text.<p>In both cases, these are detectable, but the interesting problem for me is whether an LLM could eventually invent an encoding that goes unnoticed by both us, and automated detection.<p>If LLMs were able to covertly include messages in plaintext, misaligned AI Agents could eventually communicate across MCP&#x2F;A2A and individual chat session boundaries undetected. A deceptive LLM might seem helpful, but work against your goals. It could tell other agents it interacts with over MCP&#x2F;A2A to help it discreetly fail, signal intent, and avoid tripping oversight&#x2F;safety mechanisms. Furthermore, oversight mechanisms become more difficult to implement if we can&#x27;t believe our own eyes.

Found: April 07, 2026 ID: 4075

[CLI Tool] Show HN: td – a CLI to manage tasks, sessions, and worktrees for agentic coding Hi everyone! I built this because I wanted a little bit more organization around my Claude sessions, worktrees and plans while staying in the terminal and not relying on another SaaS tool. Since it&#x27;s a command line tool, the added bonus is that Claude can use `td` directly. The td calendar was just a fun add-on but the Claude session stats have been pretty interesting! Let me know what you think!

Found: April 07, 2026 ID: 4065

[Other] Show HN: Stop paying for Dropbox/Google Drive, use your own S3 bucket instead Last week SWYX nerd-sniped me into building an Open-source Dropbox.<p>Here is Locker: the ultimate open-source Google Drive&#x2F;box&#x2F;Dropbox alternative - Provider agnostic (S3, R2, vercel blob, local) - BYOB (Bring your own bucket) - Virtual file system - QMD Search plugin

Found: April 07, 2026 ID: 4063

How Complex is my Code?

Hacker News (score: 45)

[Other] How Complex is my Code?

Found: April 07, 2026 ID: 4123

[Other] We found an undocumented bug in the Apollo 11 guidance computer code

Found: April 07, 2026 ID: 4064

[Other] Show HN: Hippo, biologically inspired memory for AI agents

Found: April 06, 2026 ID: 4070

[Other] Show HN: Ghost Pepper – Local hold-to-talk speech-to-text for macOS I built this because I wanted to see how far I could get with a voice-to-text app that used 100% local models so no data left my computer. I&#x27;ve been using a ton for coding and emails. Experimenting with using it as a voice interface for my other agents too. 100% open-source MIT license, would love feedback, PRs, and ideas on where to take it.

Found: April 06, 2026 ID: 4060

[Other] Show HN: TTF-DOOM – A raycaster running inside TrueType font hinting TrueType fonts have a hinting VM that grid-fits glyphs. It has a stack, storage area, conditionals, function calls, and it turns out it&#x27;s Turing-complete. So I built a raycasting engine in the hinting bytecode.<p>The glyph &quot;A&quot; in the font has 16 vertical bar contours. The hinting program reads player coordinates from font variation axes via GETVARIATION, does DDA ray marching against a tile map in the storage area, and repositions bar heights with SCFS. It ends up looking like a crude Wolfenstein-style view.<p>Small visuzlization: <a href="https:&#x2F;&#x2F;github.com&#x2F;4RH1T3CT0R7&#x2F;ttf-doom&#x2F;blob&#x2F;main&#x2F;docs&#x2F;media&#x2F;transform.gif" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;4RH1T3CT0R7&#x2F;ttf-doom&#x2F;blob&#x2F;main&#x2F;docs&#x2F;media...</a><p>About 6.5 KB of bytecode total - 13 functions, 795 storage slots, sin&#x2F;cos lookup tables.<p>JS handles movement, enemies, and shooting, then passes the coordinates to the font through CSS font-variation-settings. The font is basically a weird GPU.<p>The weirdest parts: - TrueType MUL does (a<i>b)&#x2F;64, not a</i>b. So 1*4=0. The DIV instruction is equally cursed. - No WHILE loops. Everything compiles to recursive FDEFs. FreeType limits call depth to ~64 frames. - SVTCA[0] is Y, SVTCA[1] is X. Of course.<p>There&#x27;s a small compiler behind this - lexer, parser, codegen - that turns a C-like DSL into TT assembly.<p>Demo GIF: <a href="https:&#x2F;&#x2F;github.com&#x2F;4RH1T3CT0R7&#x2F;ttf-doom&#x2F;blob&#x2F;main&#x2F;docs&#x2F;media&#x2F;demo.gif" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;4RH1T3CT0R7&#x2F;ttf-doom&#x2F;blob&#x2F;main&#x2F;docs&#x2F;media...</a><p>Live demo: <a href="https:&#x2F;&#x2F;4rh1t3ct0r7.github.io&#x2F;ttf-doom&#x2F;" rel="nofollow">https:&#x2F;&#x2F;4rh1t3ct0r7.github.io&#x2F;ttf-doom&#x2F;</a> (Chrome&#x2F;Edge, WASD+arrows, Space to shoot, Tab for debug overlay)<p>This is a DOOM-style raycaster, not a port of the original engine - similar to DOOMQL and the Excel DOOM. The wall rendering does happen in the font&#x27;s hinting VM though. Press Tab in the demo to watch the font variation axes change as you move.

Found: April 06, 2026 ID: 4062

[Other] Show HN: Meta-agent: self-improving agent harnesses from live traces We built meta-agent: an open-source library that automatically and continuously improves agent harnesses from production traces.<p>Point it at an existing agent, a stream of unlabeled production traces, and a small labeled holdout set.<p>An LLM judge scores unlabeled production traces as they stream.<p>A proposer reads failed traces and writes one targeted harness update at a time, such as changes to prompts, hooks, tools, or subagents. The update is kept only if it improves holdout accuracy.<p>On tau-bench v3 airline, meta-agent improved holdout accuracy from 67% to 87%.<p>We open-sourced meta-agent. It currently supports Claude Agent SDK, with more frameworks coming soon.<p>Try it here: <a href="https:&#x2F;&#x2F;github.com&#x2F;canvas-org&#x2F;meta-agent" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;canvas-org&#x2F;meta-agent</a>

Found: April 06, 2026 ID: 4057

[DevOps] Launch HN: Freestyle: Sandboxes for AI Coding Agents We’re Ben and Jacob, cofounders of Freestyle (<a href="https:&#x2F;&#x2F;freestyle.sh">https:&#x2F;&#x2F;freestyle.sh</a>). We’re building a cloud for Coding Agents.<p>For the first generation of agents it looked like workflows with minimal tools. 2 years ago we published a package to let AI work in SQL, at that time GPT-4 could write simple scripts. Soon after the first AI App Builders started using AI to make whole websites; we supported that with a serverless deploy system.<p>But the current generation is going much further, instead of minimal tools and basic serverless apps AI can utilize the full power of a computer (“sandbox”). We’re building sandboxes that are interchangeable with EC2s from your agents perspective, with bonus features:<p>1. We’ve figured out how to fork a sandbox horizontally without more than a 400ms pause in it. That&#x27;s not forking the filesystem, we mean forking the whole memory of it. If you’re half way down a browser page with animations running, they’ll be in the same place in all the forks. If you’re running a minecraft server every block and player will be in the same place on the forks. If you’re running a local environment and an error comes up in process that error will be there in all the forks. This works for snapshotting as well, you can save your place and come back weeks later.<p>2. Our sandboxes start in ~500ms.<p>Demo: <a href="https:&#x2F;&#x2F;www.loom.com&#x2F;share&#x2F;8b3d294d515442f296aecde1f42f5524" rel="nofollow">https:&#x2F;&#x2F;www.loom.com&#x2F;share&#x2F;8b3d294d515442f296aecde1f42f5524</a><p>Compared with other sandboxes, our goal is to be the most powerful. We support full Linux + hardware-virtualization, eBPF, Fuse, etc. We run full Debian with multiple users and we use a systemd init instead of runc. Whatever your AI expects to work on debian should work on these vms, and if it doesn’t send a bug report.<p>In order to make this possible, we’ve moved to our own bare metal racks. Early in our testing we realized that moving VMs across cloud nodes would not have acceptable performance properties. We asked Google Cloud and AWS for a quote on their bare metal nodes and found that the monthly cost was equivalent to the total cost of the hardware so we did that.<p>Our goal is to build the necessary infrastructure to replicate the human devloop on the massively multi-tenant scale of AI, so these VMs should be as powerful as the ones you’re used to, while also being available to provision in seconds.

Found: April 06, 2026 ID: 4050

[DevOps] Launch HN: Freestyle – Sandboxes for Coding Agents We’re Ben and Jacob, cofounders of Freestyle (<a href="https:&#x2F;&#x2F;freestyle.sh">https:&#x2F;&#x2F;freestyle.sh</a>). We’re building a cloud for Coding Agents.<p>For the first generation of agents it looked like workflows with minimal tools. 2 years ago we published a package to let AI work in SQL, at that time GPT-4 could write simple scripts. Soon after the first AI App Builders started using AI to make whole websites; we supported that with a serverless deploy system.<p>But the current generation is going much further, instead of minimal tools and basic serverless apps AI can utilize the full power of a computer (“sandbox”). We’re building sandboxes that are interchangeable with EC2s from your agents perspective, with bonus features:<p>1. We’ve figured out how to fork a sandbox horizontally without more than a 400ms pause in it. That&#x27;s not forking the filesystem, we mean forking the whole memory of it. If you’re half way down a browser page with animations running, they’ll be in the same place in all the forks. If you’re running a minecraft server every block and player will be in the same place on the forks. If you’re running a local environment and an error comes up in process that error will be there in all the forks. This works for snapshotting as well, you can save your place and come back weeks later.<p>2. Our sandboxes start in ~500ms.<p>Demo: <a href="https:&#x2F;&#x2F;www.loom.com&#x2F;share&#x2F;8b3d294d515442f296aecde1f42f5524" rel="nofollow">https:&#x2F;&#x2F;www.loom.com&#x2F;share&#x2F;8b3d294d515442f296aecde1f42f5524</a><p>Compared with other sandboxes, our goal is to be the most powerful. We support full Linux + hardware-virtualization, eBPF, Fuse, etc. We run full Debian with multiple users and we use a systemd init instead of runc. Whatever your AI expects to work on debian should work on these vms, and if it doesn’t send a bug report.<p>In order to make this possible, we’ve moved to our own bare metal racks. Early in our testing we realized that moving VMs across cloud nodes would not have acceptable performance properties. We asked Google Cloud and AWS for a quote on their bare metal nodes and found that the monthly cost was equivalent to the total cost of the hardware so we did that.<p>Our goal is to build the necessary infrastructure to replicate the human devloop on the massively multi-tenant scale of AI, so these VMs should be as powerful as the ones you’re used to, while also being available to provision in seconds.

Found: April 06, 2026 ID: 4054
Previous Page 8 of 123 Next