Show HN: Ducklang: Achieving 100x more requests per second than NextJS

Show HN (score: 6)
Found: January 01, 2026
ID: 2866

Description

Other
Show HN: Ducklang: Achieving 100x more requests per second than NextJS Duck (https://duck-lang.dev) is a statically typed, compiled programming language that combines the best of Rust, TypeScript and Go, aiming to provide an alternative for full-stack-development while being as familiar as possible

Improvements over Rust: - garbage collection simplifies developing network applications - no lifetimes - built-in concurrency runtime and apis for web development

Improvements over bun/node/typescript: - massive performance gains due to Go's support for parallel execution and native code generation - easier deployment since Duck compiles to a statically linked native executable that doesn't need dependencies - reduced complexity and costs since a single duck deployment massively outscales anything that runs javascript - streamlined toolchain management using duckup (compiler version manager) and dargo (build tool)

Improvements over Go: - a more expresive type system supporting Union types, Duck typing and tighter control over mutability - Server Side Rendering with a jsx-like syntax as well as preact components for frontend development - better error handling based on union types - a rust based reimplementation of tailwind that is directly integrated with the language (but optional to use) - type-safe json apis

More from Show

Show HN: Agent MCP Studio – build multi-agent MCP systems in a browser tab

Show HN: Agent MCP Studio – build multi-agent MCP systems in a browser tab I built a browser-only studio for designing and orchestrating MCP agent systems for development and experimental purposes. The whole stack — tool authoring, multi-agent orchestration, RAG, code execution — runs from a single static HTML file via WebAssembly. No backend.<p>The bet: WASM is a hard sandbox for free. When you generate tools with an LLM (or write them by hand), the studio AST-validates the source, registers it lazily, and JIT-compiles into Pyodide on first call. SQL tools run in DuckDB-WASM in a Web Worker. The built-in RAG uses Xenova&#x2F;all-MiniLM-L6-v2 via Transformers.js for on-device embeddings. Nothing leaves the browser; close the tab and the stack is gone. The WASM boundary is what makes it safe to execute LLM-generated code locally — no Docker, no per-tenant container, no server.<p>Above the tool layer sits an agentic system with 10 orchestration strategies:<p>- Supervisor (router → 1 expert) - Mixture of Experts (parallel + synthesizer) - Sequential Pipeline - Plan &amp; Execute (planner decomposes, workers execute) - Swarm (peer handoffs) - Debate (contestants + judge) - Reflection (actor + critic loop) - Hierarchical (manager delegates via ask_&lt;persona&gt; tools) - Round-Robin (panel + moderator) - Map-Reduce (splitter → parallel → aggregator)<p>You build a team visually: drag tool chips onto persona nodes on a service graph, pick a strategy, and the topology reshapes to match. Each persona auto-registers as an MCP tool (ask_&lt;name&gt;), plus an agent_chat(query, strategy?) meta tool. A bundled Node bridge speaks stdio to Claude Desktop and WebSocket to your tab — your browser becomes an MCP server.<p>When you&#x27;re done, Export gives you a real Python MCP server: server.py, agentic.py, tools&#x2F;*.py, Dockerfile, requirements.txt, .env.example. The exported agentic.py is a faithful Python port of the same orchestration logic running in the browser, so the deployable artifact behaves identically to the prototype.<p>Also shipped: Project Packs. Export the whole project as a single .agentpack.json. Auto-detects required external services (OpenAI, GitHub, Stripe, Anthropic, Slack, Notion, Linear, etc.) by scanning tool source for os.environ.get(...) and cross-referencing against the network allowlist. Recipients get an import wizard that prompts for credentials. Manifests are reviewable, sharable, and never carry secrets.<p>Some things I&#x27;m honestly uncertain about:<p>- 10 strategies might be too many. My guess is most users only need Supervisor, Mixture of Experts, and Debate. Open to data on which ones actually pull weight. - Browser cold-starts (Pyodide warm-up on first load) are a real UX hit despite aggressive caching. - bridge.js is the only non-browser piece. A hosted variant is the obvious next step.<p>Built with Pyodide, DuckDB-WASM, Transformers.js, OpenAI Chat Completions (or a local Qwen 1.5 0.5B running in-browser via Transformers for fully offline mode). ~5K lines of HTML&#x2F;CSS&#x2F;JS in one file.<p><a href="https:&#x2F;&#x2F;www.agentmcp.studio" rel="nofollow">https:&#x2F;&#x2F;www.agentmcp.studio</a><p>Genuinely curious whether running this much LLM-generated code in a browser tab feels reasonable to you, or quietly terrifying.

No other tools from this source yet.