Tuesday, 3 March 2026

Show HN: DubTab – Live AI Dubbing in the Browser (Meet/YouTube/Twitch/etc.) https://bit.ly/4u1yiVL

Show HN: DubTab – Live AI Dubbing in the Browser (Meet/YouTube/Twitch/etc.) Hi HN — I’m Ethan, a solo developer. I built DubTab because I spend a lot of time in meetings and watching videos in languages I’m not fluent in, and subtitles alone don’t always keep up (especially when the speaker is fast). DubTab is a Chrome/Edge extension that listens to the audio of your current tab and gives you: 1.Live translated subtitles (optional bilingual mode) 2.Optional AI dubbing with a natural-sounding voice — so you can follow by listening, not just reading The goal is simple: make it easier to understand live audio in another language in real time, without downloading files or doing an upload-and-wait workflow. How you’d use it 1.Open a video call / livestream / lecture / any tab with audio 2.Start DubTab 3.Choose target language (and source language if you know it) 4.Use subtitles only, or turn on natural AI dubbing and adjust the audio mix (keep original, or duck it) What it’s good for 1.Following cross-language meetings/classes when you’re tired of staring at subtitles 2.Watching live content where you can’t pause/rewind constantly 3.Language learners who want bilingual captions to sanity-check meaning 4.Keeping up with live news streams on YouTube when events are unfolding in real time (e.g., breaking international updates like U.S./Iran/Israel-related developments) Link: https://bit.ly/40HBFUo I’ll be in the comments and happy to share implementation details if anyone’s curious. https://bit.ly/40HBFUo March 4, 2026 at 02:04AM

Show HN: I built a LLM human rights evaluator for HN (content vs. site behavior) https://bit.ly/4l4c4yi

Show HN: I built a LLM human rights evaluator for HN (content vs. site behavior) My health challenges limit how much I can work. I've come to think of Claude Code as an accommodation engine — not in the medical-paperwork sense, but in the literal one: it gives me the capacity to finish things that a normal work environment doesn't. Observatory was built in eight days because that kind of collaboration became possible for me. (I even used Claude Code to write this post — but am only posting what resonates with me.) Two companion posts: on the recursive methodology ( https://bit.ly/409tFeD... ) and what 806 evaluated stories reveal ( https://bit.ly/4r7k9DW... ). I built Observatory to automatically evaluate Hacker News front-page stories against all 31 provisions of the UN Universal Declaration of Human Rights — starting with HN because its human-curated front page is one of the few feeds where a story's presence signals something about quality, not just virality. It runs every minute: https://bit.ly/4aKNMpG . Claude Haiku 4.5 handles full evaluations; Llama 4 Scout and Llama 3.3 70B on Workers AI run a lighter free-tier pass. The observation that shaped the design: rights violations rarely announce themselves. An article about a company's "privacy-first approach" might appear on a site running twelve trackers. The interesting signal isn't whether an article mentions privacy — it's whether the site's infrastructure matches its words. Each evaluation runs two parallel channels. The editorial channel scores what the content says about rights: which provisions it touches, direction, evidence strength. The structural channel scores what the site infrastructure does: tracking, paywalls, accessibility, authorship disclosure, funding transparency. The divergence — SETL (Structural-Editorial Tension Level) — is often the most revealing number. "Says one thing, does another," quantified. Every evaluation separates observable facts from interpretive conclusions (the Fair Witness layer, same concept as fairwitness.bot — https://bit.ly/43DzQKs ). You get a facts-to-inferences ratio and can read exactly what evidence the model cited. If a score looks wrong, follow the chain and tell me where the inference fails. Per our evaluations across 805 stories: only 65% identify their author — one in three HN stories without a named author. 18% disclose conflicts of interest. 44% assume expert knowledge (a structural note on Article 26). Tech coverage runs nearly 10× more retrospective than prospective: past harm documented extensively; prevention discussed rarely. One story illustrates SETL best: "Half of Americans now believe that news organizations deliberately mislead them" (fortune.com, 652 HN points). Editorial: +0.30. Structural: −0.63 (paywall, tracking, no funding disclosure). SETL: 0.84. A story about why people don't trust media, from an outlet whose own infrastructure demonstrates the pattern. The structural channel for free Llama models is noisy — 86% of scores cluster on two integers. The direction I'm exploring: TQ (Transparency Quotient) — binary, countable indicators that don't need LLM interpretation (author named? sources cited? funding disclosed?). Code is open source: https://bit.ly/3MJJANP — the .claude/ directory has the cognitive architecture behind the build. Find a story whose score looks wrong, open the detail page, follow the evidence chain. The most useful feedback: where the chain reaches a defensible conclusion from defensible evidence and still gets the normative call wrong. That's the failure mode I haven't solved. My background is math and psychology (undergrad), a decade in software — enough to build this, not enough to be confident the methodology is sound. Expertise in psychometrics, NLP, or human rights scholarship especially welcome. Methodology, prompts, and a 15-story calibration set are on the About page. Thanks! https://bit.ly/4aKNMpG March 4, 2026 at 01:26AM

Show HN: Interactive WordNet Visualizer-Explore Semantic Relations as a Graph https://bit.ly/4l9DCCr

Show HN: Interactive WordNet Visualizer-Explore Semantic Relations as a Graph https://bit.ly/4l7NYTv March 3, 2026 at 10:17PM

Monday, 2 March 2026

Show HN: An Auditable Decision Engine for AI Systems https://bit.ly/4r0ct6d

Show HN: An Auditable Decision Engine for AI Systems https://bit.ly/4rKkhKt March 3, 2026 at 03:03AM

Show HN: PHP 8 disable_functions bypass PoC https://bit.ly/4coTizr

Show HN: PHP 8 disable_functions bypass PoC https://bit.ly/4ckhq6k March 3, 2026 at 02:12AM

Show HN: We filed 99 patents for deterministic AI governance(Prior Art vs. RLHF) https://bit.ly/3OHLRtr

Show HN: We filed 99 patents for deterministic AI governance(Prior Art vs. RLHF) For the last few months, we've been working on a fundamental architectural shift in how autonomous agents are governed. The current industry standard relies almost entirely on probabilistic alignment (RLHF, system prompts, constitutional training). It works until it's jailbroken or the context window overflows. A statistical disposition is not a security boundary. We've built an alternative: Deterministic Policy Gates. In our architecture, the LLM is completely stripped of execution power. It can only generate an "intent payload." That payload is passed to a process-isolated, deterministic execution environment where it is evaluated against a cryptographically hashed constraint matrix (the constitution). If it violates the matrix, it is blocked. Every decision is then logged to a Merkle-tree substrate (GitTruth) for an immutable audit trail. We filed 99 provisional patents on this architecture starting January 10, 2026. Crucially, we embedded strict humanitarian use restrictions directly into the patent claims themselves (The Peace Machine Mandate) so the IP cannot legally be used for autonomous weapons, mass surveillance, or exploitation. I wrote a full breakdown of the architecture, why probabilistic safety is a dead end, and the timeline of how we filed this before the industry published their frameworks: Read the full manifesto here: https://bit.ly/4l5y3Vx... The full patent registry is public here: https://bit.ly/4l1JNbI I'm the founder and solo inventor. Happy to answer any questions about the deterministic architecture, the Merkle-tree state persistence, or the IP strategy of embedding ethics directly into patent claims. March 2, 2026 at 11:56PM

Show HN: Open-Source Postman for MCP https://bit.ly/4l4lxG3

Show HN: Open-Source Postman for MCP https://bit.ly/40EKzC1 March 3, 2026 at 12:40AM

Sunday, 1 March 2026

Show HN: Vibe Code your 3D Models https://bit.ly/4aYHwto

Show HN: Vibe Code your 3D Models Hi HN, I’m the creator of SynapsCAD, an open-source desktop application I've been building that combines an OpenSCAD code editor, a real-time 3D viewport, and an AI assistant. You can write OpenSCAD code, compile it directly to a 3D mesh, and use an LLM (OpenAI, Claude, Gemini, ...) to modify the code through natural language. Demo video: https://www.youtube.com/watch?v=cN8a5UozS5Q A bit about the architecture: - It’s built entirely in Rust. - The UI and 3D viewport are powered by Bevy 0.15 and egui. - It uses a pure-Rust compilation pipeline (openscad-rs for parsing and csgrs for constructive solid geometry rendering) so there are no external tools or WASM required. - Async AI network calls are handled by Tokio in the background to keep the Bevy render loop smooth. Disclaimer: This is a very early prototype. The OpenSCAD parser/compiler doesn't support everything perfectly yet, so you will definitely hit some rough edges if you throw complex scripts at it. I mostly just want to get this into the hands of people who tinker with CAD or Rust. I'd be super happy for any feedback, architectural critiques, or bug reports—especially if you can drop specific OpenSCAD snippets that break the compiler in the GitHub issues! GitHub (Downloads for Win/Mac/Linux): https://bit.ly/3MDl1Cd Happy to answer any questions about the tech stack or the roadmap! https://bit.ly/3MDl1Cd February 27, 2026 at 06:27PM

Show HN: Logira – eBPF runtime auditing for AI agent runs https://bit.ly/3MP5orl

Show HN: Logira – eBPF runtime auditing for AI agent runs I started using Claude Code (claude --dangerously-skip-permissions) and Codex (codex --yolo) and realized I had no reliable way to know what they actually did. The agent's own output tells you a story, but it's the agent's story. logira records exec, file, and network events at the OS level via eBPF, scoped per run. Events are saved locally in JSONL and SQLite. It ships with default detection rules for credential access, persistence changes, suspicious exec patterns, and more. Observe-only – it never blocks. https://bit.ly/4sgvLW1 https://bit.ly/4sgvLW1 March 2, 2026 at 12:25AM