Daily Show HN

Upvote0

Show HN for January 15, 2026

42 items
74

1Code – Open-source Cursor-like UI for Claude Code #

github.com favicongithub.com
48 comments7:20 PMView on HN
Hi, we're Sergey and Serafim. We've been building dev tools at 21st.dev and recently open-sourced 1Code (https://1code.dev), a local UI for Claude Code.

Here's a video of the product: https://www.youtube.com/watch?v=Sgk9Z-nAjC0

Claude Code has been our go-to for 4 months. When Opus 4.5 dropped, parallel agents stopped needing so much babysitting. We started trusting it with more: building features end to end, adding tests, refactors. Stuff you'd normally hand off to a developer. We started running 3-4 at once. Then the CLI became annoying: too many terminals, hard to track what's where, diffs scattered everywhere.

So we built 1Code.dev, an app to run your Claude Code agents in parallel that works on Mac and Web. On Mac: run locally, with or without worktrees. On Web: run in remote sandboxes with live previews of your app, mobile included, so you can check on agents from anywhere. Running multiple Claude Codes in parallel dramatically sped up how we build features.

What’s next: Bug bot for identifying issues based on your changes; QA Agent, that checks that new features don't break anything; Adding OpenCode, Codex, other models and coding agents. API for starting Claude Codes in remote sandboxes.

Try it out! We're open-source, so you can just bun build it. If you want something hosted, Pro ($20/mo) gives you web with live browser previews hosted on remote sandboxes. We’re also working on API access for running Claude Code sessions programmatically.

We'd love to hear your feedback!

34

MailPilot – Freedom to go anywhere while your agents work #

40 comments7:20 AMView on HN
https://mailpilot.chat

What is this? A local TUI (like Claude, Codex, Gemini, OpenCode or Copilot) that wraps your agent and sends each turn to an email in a nice format.

What does it enable? It lets you email your agents. And them email you. So They Keep Working on the tasks you want, and You go and do What You Want. No need to stay at your desk. Be free.

I built this for me, but thought others would find it useful, so I turned it into a product. I want get outside, and away from my desk, but still have the agents work.

The accidental killer feature: You can CC your team. If you forward the thread to a coworker, their reply goes straight to the agent context too. It turns a local session into an async multiplayer thread.

Works out of the box with Claude, Codex, Gemini, Copilot, and OpenCode. Happy to answer questions!

https://mailpilot.chat

33

Tusk Drift – Turn production traffic into API tests #

github.com favicongithub.com
8 comments6:54 PMView on HN
Hi HN! In the past few months my team and I have been working on Tusk Drift, a system that records real API traffic from your service, then replays those requests as deterministic tests. Outbound I/O (databases, HTTP calls, etc.) gets automatically mocked using the recorded data.

Problem we're trying to solve: Writing API tests is tedious, and hand-written mocks drift from reality. We wanted tests that stay realistic because they come from real traffic.

versus mocking libraries: Tools like VCR/Nock intercept HTTP within your tests. Tusk Drift records full request/response traces externally (HTTP, DB, Redis, etc.) and replays them against your running service, no test code or fixtures to write/maintain.

How it works:

1. Add a lightweight SDK (we currently support Python and Node.js)

2. Record traffic in any environment.

3. Run `tusk run`, the CLI sandboxes your service and serves mocks via Unix socket

We run this in CI on every PR. Also been using it as a test harness for AI coding agents, they can make changes, run `tusk run`, and get immediate feedback without needing live dependencies.

Source: https://github.com/Use-Tusk/tusk-drift-cli

Demo: https://github.com/Use-Tusk/drift-node-demo

Happy to answer questions!

31

The Hessian of tall-skinny networks is easy to invert #

github.com favicongithub.com
23 comments8:36 PMView on HN
It turns out the inverse of the Hessian of a deep net is easy to apply to a vector.

Doing this naively takes cubically many operations in the number of layers (so impractical), but it's possible to do this in time linear in the number of layers (so very practical)! This is possible because the Hessian of a deep net has a matrix polynomial structure that factorizes nicely. The Hessian-inverse-product algorithm that takes advantage of this is similar to running backprop on a dual version of the deep net. It echoes an old idea of Pearlmutter's for computing Hessian-vector products.

Maybe this idea is useful as a preconditioner for stochastic gradient descent?

14

Control Claude permissions using cloud-based decision tables #

github.com favicongithub.com
10 comments6:21 PMView on HN
We’ve been building rule engines for a while and had the fun idea lately to see what happens when we use our decision table UI with Claude’s PreToolUse hook.

The result is a surprisingly useful policy/gating layer– these tables let your team:

- Write multi-factor, exception-friendly policies (e.g. deny rm -rf / when --force; allow cleanup only in node_modules; ask on network calls like curl/wget; block kubectl delete or SQL DROP, each with a clear reason)

- Roll out policy changes instantly (mid-run, flip a risky operation from allow → ask; the next attempt across devs and agents is gated immediately– no git pull, agent restart, or coordination)

- Adopt lightweight governance that survives churn (MCP/skills/etc): just add columns/rules as new tools and metadata show up

- Get a quick central utility to understand which tools are being used, which tools get blocked most often, and why

12

Munimet.ro – ML-based status page for the local subways in SF #

munimet.ro faviconmunimet.ro
7 comments6:51 PMView on HN
During a recent subway outage in San Francisco I decided to build a webapp in the spirit of "Do I Need an Umbrella," basically to answer the question "Should I take the subway or a bus?"

In the interest of learning new tools I decided to vibe code it as much as possible.

First, I had Claude Code write a background script to download images of the real-time circuit diagram of the subway, which are publicly available here: http://sfmunicentral.com/

Next I had it build an image labeler tool in tkinter, which turned out to need a lot of manual tweaking before I could even get to the labeling. Seemed like the right tool for the job, but it would have saved time if I'd built it from scratch myself.

The most interesting part was turning the labeled image data into predictions with pytorch. Claude wrote the initial script fairly quickly, but as these things go it required manual tweaking and second guessing myself on the images flagged as outliers. I'll admit I got embarrassingly far along before realizing that Claude hadn't enabled pytorch's GPU support; a real facepalm moment on my part.

For those curious, brave, or crazy enough to dive in the source code is available here under an MIT license: https://github.com/MrEricSir/munimet.ro

11

I modeled personal success as a control system with Bayesian priors #

mondonno.github.io faviconmondonno.github.io
1 comments12:29 PMView on HN
I am a student and co-founder of an accessibility startup. Balancing product development with exams broke my workflow, so I looked for a debugger.

Mathematica Successūs is my attempt to model the "Self" as a control system. It argues that stability is only possible when your regulator's variety matches environmental disturbances (Ashby's Law).

It is a short (59-page) technical manual written in LaTeX. It treats burnout not as a feeling, but as a system constraint violation. It relies on mathematical rigor, not coaching fluff.

10

I spent 10k hours building the perfect language learning app #

phrasing.app faviconphrasing.app
8 comments3:25 PMView on HN
Hello HN -

Over the last few years, I've been building an app to learn languages. It supports ~85 languages and ~40 dialects, and is optimized for learning and maintaining multiple languages.

Learning a language is a ton of work, but really quite simple. Find text you're interested in, dissect and understand it, and review it often. Grind that every day for a few years, and you'll be fluent.

So I built Phrasing for 4 main reasons:

1. To have an app I would actually be excited to use every single day for years.

2. To do all the research, explanation, prioritization, course planning and management for me, so I could just focus on learning

3. To support learning multiple languages as a first class citizen

4. To have a nice, unified interface to learn languages from Spanish to Arabic to Cantonese

I've been using it daily now for almost a year to learn a variety of languages and I've been loving it. I have an awesome group of users in the Telegram channel who have been using it as well to great success.

If you're a polyglot, an avid language learner, or someone who tried to use Anki in the past but found it wasn't for them — I think you might really like Phrasing :)

--

PS: Phrasing uses FSRS under the hood, so a huge thank you to the FSRS team and Jarrett Ye for all the amazing work. Phrasing would not be the product it is without you!

PSS: I'm documenting some of the latest updates on the blog this week, they're 50% product and 50% technical, might be of special interest to the HN crowd: https://phrasing.app/blog/launch-week-1

10

I built a game on my old phone without knowing what I was building #

kikkupico.com faviconkikkupico.com
10 comments11:08 AMView on HN
I'm calling this "Vibe Discovery" — distinct from vibe coding because I didn't know the requirements upfront. Started with "make something with the accelerometer" and discovered through 6 iterations that I wanted a WebGL marble game. The interesting part was the dev setup: Claude Code running in Termux on a Redmi Note 9 (4GB RAM). The same-device feedback loop — code, test accelerometer, react, iterate — made rapid discovery possible in a way that laptop-to-phone deployment wouldn't.
6

See the best daily US stock trades with entry and exit timing #

finsuba.com faviconfinsuba.com
0 comments1:20 PMView on HN
Hi HN,

I built Finsuba Premium to solve a problem I kept running into while trading:

Most sites highlight “top gaining stocks” over a fixed time window (e.g. 9:30–4:00), but that often hides the actual tradeability. A stock can finish the day up big while offering very different outcomes depending on when you entered and exited.

Finsuba Premium sends a daily and weekly email showing the top-performing U.S. stock trades, based on identifiable entry and exit points rather than broad time-based gains. It includes:

- exact entry time & price - exact exit time & price - trade duration and performance

It’s not a signal service or alert system. Everything is processed at the end of the trading day and ran through minute by minute. I built this for traders who want to improve by reviewing winning setups consistently, rather than reacting to alerts in real time.

Link: https://www.finsuba.com/premium/

Happy to answer questions about how the trades are selected, the data pipeline, or the product itself. Feedback very welcome.

4

IMSAI/Altair inspired microcomputer with web emulator #

gzalo.github.io favicongzalo.github.io
1 comments1:57 AMView on HN
I designed and built a physical replica of a 1970s-style front panel microcomputer with 25+ toggle switches, 16 LEDs, and an LCD display. The brain is a Raspberry Pi Pico running an Intel 8080 CPU emulator.

The main twist: I decided to see how far I could get using Claude Code for the firmware. That and the web emulator were written almost entirely using Claude Code (Opus 4.5). I've kept the full prompt history here: https://github.com/gzalo/microcomputer/blob/main/ai_prompts....

It was able to create the emulator in just a few prompts! It really surprised me that it was able to make a WebAssembly version from the same code (compiled with emscripten) and get the physical layout of the panel from a given photo. It also created some simple working examples using 8086 instructions!

Repository: https://github.com/gzalo/microcomputer

4

Flour Hour – I built a bread baking app with Claude Code in 3 hours #

yaninatrekhleb.github.io faviconyaninatrekhleb.github.io
1 comments4:42 AMView on HN
I built this to solve my own problem — sourdough takes ~20 hours from start to finish, and I was tired of setting a dozen alarms and constantly checking "what's next?" Flour Hour lets you pick a recipe, set your start time (or work backwards from when you want to eat), and it calculates every step with exact timestamps. Includes 22 recipes: classic sourdough, same-day sourdough, focaccia, bagels, croissants, brioche, discard pancakes, and more. Each with ingredients. Tech: React + Vite, deployed on GitHub Pages. The interesting part — I built it entirely by talking to Claude Code. Described what I wanted like a client, and it wrote all the code. Took about 3 hours of conversation (not counting rate limits). GitHub: https://github.com/YaninaTrekhleb/flour-hour Happy to answer questions about the build process or the app itself.
3

Matriq – Search inside video files using natural language #

matriq.video faviconmatriq.video
0 comments7:21 AMView on HN
Hello HN, I’m David. I built Matriq because I was frustrated with the "Post-Production Scrub."

I had hours of B-roll and long-form content, but finding a specific 5-second clip (e.g., "a red car passing by" or "a specific quote about React state") meant manually watching the footage at 2x speed.

What I built: Matriq indexes video content visually and aurally. It uses multimodal embeddings to understand scene context, action, and dialogue.

The Use Case:

While it works for general editing, I’ve found the best use case is Content Repurposing. Creators with 100+ hours of archives can now instantly find "viral hooks" or specific topics to turn into Shorts/Reels without re-watching old footage.

It’s in beta. I’d love you to try breaking it or give feedback on the retrieval accuracy.

Link: https://matriq.video

3

Beni AI – video call with your AI companion #

app.thebeni.ai faviconapp.thebeni.ai
0 comments12:56 AMView on HN
Beni AI is a real-time AI companion for natural 1:1 conversations. Unlike typical text-based character chat, Beni focuses on face-to-face interaction, consistent personality, and long-term memory that adapts to you over time.

We’re building Beni for users who want an AI that feels like a real presence rather than a scripted chatbot. Try it here: >> https://thebeni.ai/

3

Chklst – A Minimalist Checklist #

chklst.xyz faviconchklst.xyz
0 comments1:48 AMView on HN
Welp... I finally shipped. This is my first real project. I wanted a checklist app the way I wanted it so I built chklst.

What’s different? Simple, drag & drop reordering, keyboard shortcuts, color labels. There’s a live demo on the landing page so you can try it without signing up.

Free accounts can create 1 list. Premium is $5/month for up to 25 lists + shareable lists. What do you think?

I built it with Next.js 16 + Turso/libSQL + Drizzle + Better Auth + Stripe.

https://www.chklst.xyz/

Would love feedback on onboarding, UX, and pricing.

Thanks everyone!

3

Stash: End-to-end encrypted file sharing with zero friction #

stash-app.xyz faviconstash-app.xyz
1 comments1:54 PMView on HN
I built Stash because I was tired of the usual tradeoff in file sharing: either force recipients to create accounts or give up on privacy.

How it works 1. Upload from the iOS, iPad or Mac app 2. Get a shareable link 3. Recipients download in the browser, no app, no account, nothing to install

Encryption model Files are encrypted on device using AES 256 GCM before upload. The encryption key lives only in the URL fragment (after #), which browsers never send to servers. Even if the link is intercepted in transit, the file cannot be decrypted without the fragment. The server never sees plaintext or keys.

Native Mac app Built with Mac Catalyst and full bidirectional drag and drop. Drag files onto the app to upload. Drag files out of the app to save them anywhere. No save dialogs, no friction. It feels like moving files in Finder.

What I optimized for • No file size limits • No compression, original quality preserved • Links do not expire unless you delete them • Native Apple UX for senders, universal browser access for recipients

The goal was something my non technical family could use to receive files securely without walkthroughs, accounts or app installs.

I would love feedback on the encryption approach and the UX, especially from people who have built or reviewed similar systems.

Web: https://stash-app.xyz

3

Gpu.st – Instant GPU Instances over SSH #

gpu.st favicongpu.st
0 comments4:01 PMView on HN
Recently I was inspired by exe.dev to finally finish a project I'd been procrastinating on for half a year. You can simply `ssh gpu.st new` and instantly create instances with a variety of GPUs attached. Instances boot a custom image with PyTorch, Jupyter, and the CUDA toolkit installed by default.

Currently we give access to 3090s and 4090s for free users; paid users get access to all GPUs.

2

I built a satellite forensic engine to detect fraud in Carbon Markets #

0 comments7:23 AMView on HN
I got tired of reading PDF reports for Carbon Offsets that claimed to be "Protected Forests" but looked suspicious on Google Maps.

So I built Silent Earth—a standalone forensic engine (Python/React) to audit them mathematically.

The Stack: * Ingestion: Scrapes Verra/Gold Standard registries for project coordinates. * Telemetry: Queries the Sentinel-2 Satellite Constellation (via Microsoft Planetary Computer). * Analysis: Calculates NDVI (Biomass Density) and performs spectral analysis on the claimed area. * AXIOM Kernel: Cross-references industrial "shutdown" claims against Tropospheric NO2 levels (Sentinel-5P).

The Result: I routinely find "Verified" assets with an NDVI < 0.2 (Bare Soil).

The tool is currently local-only for privacy, but I'm moving the data stream to Snowflake.

Happy to answer questions about the geospatial stack (Rasterio/Shapely) or the satellite APIs.

2

GoGen – A simple template-based file generator written in Go #

github.com favicongithub.com
1 comments8:36 PMView on HN
Hi HN, I recently open-sourced GoGen (https://github.com/zaheershaikh936/gogen), a tool to automate file generation using templates. I'm trying to grow the project and reach my first 10 contributors to get listed on various open-source directories. If you have 5 minutes, I'd love feedback on: The ease of use of the CLI. The roadmap (I'm planning to add nested templates and better variable support). I've marked a few issues as "Good First Issue" for anyone who wants to contribute to a small, fresh Go project.
2

Why Neural Networks Need He Init, Clipping, and Momentum #

sbondaryev.dev faviconsbondaryev.dev
0 comments2:12 PMView on HN
Hi HN,

I made an interactive guide showing why neural networks need specific training techniques.

Start with a simple network trying to learn x². Each section introduces a problem like gradient explosion or dead neurons and shows the fix that solves it using ReLU activation, He initialization, gradient clipping, and momentum.

All code is editable and runs in the browser.

2

FrontendCheck – Learn enterprise patterns by building weird SaaS #

frontendcheck.com faviconfrontendcheck.com
0 comments1:03 PMView on HN
Built this because I kept seeing devs learn from todo app tutorials, then struggle when they hit real patterns like multi-tenancy or RBAC on the job.

The approach: you pick a standard domain (CRM, task manager) and make it weird. Your CRM becomes a cat fashion store. Your task manager becomes a heist planner. You learn the same enterprise patterns (multi-tenancy, permissions, theming, i18n) but end up with something that's not the 500th identical portfolio project.

The first challenge is free. Curious what HN thinks about this approach to teaching architecture patterns, and whether the "pick your weird domain" angle actually helps with motivation or just adds confusion.

2

Ghostty Ambient – Terminal theme switcher that learns your preferences #

github.com favicongithub.com
1 comments10:29 PMView on HN
I built a theme selector for Ghostty that learns which color schemes you prefer in different contexts.

  It reads your Mac's ambient light sensor and considers:
  - Time of day (morning, afternoon, evening, night)
  - Ambient light level (dim, office, bright, daylight)
  - System dark/light mode
  - Weather conditions
  - Power source (AC vs battery)

  A background daemon watches which themes you use and builds a Bayesian model of your preferences. Over time it learns things like "in dim evening light, they prefer high-contrast dark themes with muted colors."

  Once it has enough data, it can generate a custom theme optimized for your current context.

  ghostty-ambient              # Interactive picker with recommendations
  ghostty-ambient --ideal      # Generate optimal theme for current context
  ghostty-ambient --stats      # See what it's learned about you

  Install: uv tool install ghostty-ambient or pip install ghostty-ambient

  GitHub: https://github.com/gezibash/ghostty-ambient

  Built for Ghostty (https://ghostty.org) but the Bayesian preference learning could apply to any theming system.
2

I'm building an open-source AI agent runtime using Firecracker microVMs #

github.com favicongithub.com
3 comments5:18 PMView on HN
Hello Hacker News! I'm Mark. I'm building Moru, an open-source runtime for AI agents that runs each session in an isolated Firecracker microVM. It started as a fork of E2B, and most of the low-level Firecracker runtime is still from upstream.

It lets you run agent harnesses like Claude Code or Codex in the cloud, giving each session its own isolated microVM with filesystem and shell access.

The repo is: https://github.com/moru-ai/moru

Each VM is a snapshot of a Docker build. You define a Dockerfile, CPU, memory limits, and Moru runs the build inside a Firecracker VM, then pauses and saves the exact state: CPU, dirty memory pages, and changed filesystem blocks.

When you spawn a new VM, it resumes from that template snapshot. Memory snapshot is lazy-loaded via userfaultfd, which helps sandboxes start within a second.

Each VM runs on Firecracker with KVM isolation and a dedicated kernel. Network uses namespaces for isolation and iptables for access control.

From outside, you talk to the VM through the Moru CLI or TypeScript/Python SDK. Inside, it's just Linux. Run commands, read/write files, anything you'd do on a normal machine.

I've been building AI apps since the ChatGPT launch. These days, when an agent needs to solve complex problems, I just give it filesystem + shell access. This works well because it (1) handles large data without pushing everything into the model context window, and (2) reuses tools that already work (Python, Bash, etc.). This has become much more practical as frontier models have gotten good at tool use and multi-step workflows.

Now models run for hours on real tasks. As models get smarter, the harness should give models more autonomy, but with safe guardrails. I want Moru to help developers focus on building agents, not the underlying runtime and infra.

You can try the cloud version without setting up your own infra. It's fully self-hostable including the infra and the dashboard. I'm planning to keep this open like the upstream repo (Apache 2.0).

Give it a spin: https://github.com/moru-ai/moru Let me know what you think!

Next features I'm working toward:

- Richer streaming: today it's mostly stdin/stdout. That pushes me to overload print/console.log for control-plane communication, which gets messy fast. I want a separate streaming channel for structured events and coordination with the control plane (often an app server), while keeping stdout/stderr for debugging.

- Seamless deployment: a deploy experience closer to Vercel/Fly.io.

- A storage primitive: save and resume sessions without always having to manually sync workspace and session state.

Open to your feature requests or suggestions.

I'm focusing on making it easy to deploy and run local-first agent harnesses (e.g., Claude Agent SDK) inside isolated VMs. If you've built or are building those, I'd appreciate any notes on what's missing, or what you'd prioritize first.

1

Grain Synonym: Create multi-aesthetic OS variants #

codeberg.org faviconcodeberg.org
0 comments8:40 PMView on HN
Hi HN, I'm Keaton (@kae3g on GitHub and @kae3g_ on Twitter). I'm sharing Grain Synonym, a system for creating multi-aesthetic OS variants with distinct semantic identities while maintaining a shared technical codebase.

Grain OS and Clutch OS are the same operating system—a minimal, safety-first OS in pure Zig targeting RISC-V—with different aesthetic and philosophical identities. Grain OS (teamcarry11/xy-mathematics) emphasizes growth and cultivation, inspired by Aquarius energy from classical Vedic astrology. Clutch OS (teambrick06/yz) emphasizes mechanical precision and framework-enabling, inspired by Virgo Gemini Cancer energy. The technical codebase is shared; the human-facing semantics differ.

Inspired by TigerBeetle's precision, SixOS's (NixOS with s6) service management, Urbit's identity model, Plan 9's simplicity, vegan food forests, Y2K culture, traditional urbanism, and Modern Money Theory. The goal is to enable parallel evolution of ideas with different philosophical foundations while maintaining technical compatibility.

I'm working on this with Glow G2, an AI assistant that helps coordinate the 24-agent development team. I'm also working on turning this project into a nonprofit 501(c)(3) and am interested in Y Combinator's program. The specification document outlines the architecture, rationale, and spiritual foundation: Grain Synonym Variant System Specification

Would love feedback on the variant system concept and whether semantic diversity in OS development resonates with this community. Thanks!

1

RepoEngine – turn repos and docs into a shared knowledge layer #

repoengine.com faviconrepoengine.com
0 comments3:29 PMView on HN
Hello HN,

I am building RepoEngine to solve a problem I encountered as a backend engineer. Once a codebase grows into multiple repositories, documents, and teams, it becomes tough to track context or find the code for a specific feature/functionality. Even if it becomes harder to know in which context a decision was taken to implement it that way.

To overcome these issues, I have started building a systematic approach in RepoEngine, which unifies your repositories, docs, and decisions into a knowledge layer you can actually build on (onboarding flows, internal agents, automated docs, etc.).

What it does: - Multiple Repositories at once. ( that's from an organisation's repositories or even from personal profiles' selected repositories) - Ask questions across repos + docs and get answers with citations. - Every answer links to the exact sources (files, Notion pages, Slack threads, Jira tickets, etc.). - It indexes structure/meaning but doesn’t store your source code. - Connect GitHub, GitLab, Notion, Google Drive, Dropbox, Confluence, Slack, Jira, and Linear.

This is early, and I’m actively seeking feedback from people working with large or messy codebases. I really want a product that captures the shared context of decisions/directions taken to build a feature.

Would love feedback on the product direction, integrations you want, and how teams want to use this day‑to‑day. https://repoengine.com

1

Bldr – polyglot build system for monorepos #

github.com favicongithub.com
0 comments8:40 PMView on HN
I built bldr because I was frustrated with the Bazel/Buck2 tradeoff: powerful but painful.

Key innovations:

- Runtime dependency discovery: Actions can extend the build graph during execution. Your protobuf compiler discovers outputs naturally. No preprocessing hacks.

- Economic optimization: First build system to compute Pareto-optimal build plans across cost/time. A 2-min build at $12 vs 8-min at $2 – both valid depending on your constraints.

- Built in D with SIMD-accelerated BLAKE3 hashing, lock-free Chase-Lev work-stealing, arena allocators.

- 28 languages supported with unified handler architecture.

- Migration tools for Bazel, CMake, Maven, Gradle, Make, Cargo, npm.

Install: brew tap GriffinCanCode/bldr && brew install bldr

Happy to answer questions about the architecture!

1

Crystal Unified – Search compressed archives without decompressing #

github.com favicongithub.com
0 comments8:37 PMView on HN
Crystal Unified lets you grep through compressed archives (700MB logs, etc.) in milliseconds without extracting.

Other features: - Binary delta patching for firmware/OTA updates - Block-level random access - 10x+ compression on logs - Genomic data: 4:1 on 2-bit encoded DNA, 111,000:1 with reference compression - Parallel compression, 1GB/s+ decompression

https://github.com/powerhubinc/crystal-unified-public

1

VerityNgn–Open-source AI that fact-checks YouTube videos #

hotchilianalyticsllc.mintlify.app faviconhotchilianalyticsllc.mintlify.app
0 comments3:28 PMView on HN
I built an open-source system that generates truthfulness reports for YouTube videos using multimodal AI and a counter-intelligence approach.

*Live demo:* https://verityngn.streamlit.app

*Documentation:* https://hotchilianalyticsllc.mintlify.app

*Repo:* https://github.com/hotchilianalytics/verityngn-oss

*Substack Article:* https://ajjcop.substack.com/p/i-built-an-ai-that-fact-checks...

### The Problem

Existing fact-checking tools only analyze text transcripts. They miss on-screen graphics, visual demonstrations, and the multimodal nature of video. Worse, when you search for evidence, you often get promotional press releases that confirm false claims because the misinformation ecosystem is SEO-optimized.

### How VerityNgn Works

1. *Multimodal analysis*: Uses Gemini 2.5 Flash (1M token context) to analyze video frames at 1 FPS — audio, OCR, visuals, and transcript together. 2. *Intelligent Segmentation*: Automatically calculates optimal segment sizes based on model context windows (reduces API calls by 86% for typical 30-min videos). 3. *Enhanced claim extraction*: Multi-pass extraction with specificity scoring (0-100) and "absence claim" generation (identifying what's NOT mentioned, like missing FDA disclaimers). 4. *Counter-intelligence*: Actively hunts for contradiction — searching for YouTube review/debunking videos and detecting self-referential press releases (94% precision). 5. *Probabilistic output*: A calibrated THREE-state distribution (TRUE/FALSE/UNCERTAIN) based on Bayesian aggregation of source validation power.

### Results (200-claim test set)

- 75% accuracy vs. ground truth (95% CI: 61-85%) - +18% improvement from counter-intel on misleading content - Well-calibrated (Brier score = 0.12, ECE = 0.04) - Cost: $0.50–$2.00 per video

### Tech Stack

- Python 3.12, Gemini 2.5 Flash via Vertex AI - LangChain/LangGraph for orchestration - Streamlit UI, Cloud Run backend - yt-dlp for video download - Google Custom Search + YouTube Data API

### Honest Limitations

- English only - YouTube only (no TikTok/Instagram yet) - ~25% error rate (75% accuracy means 25% wrong) - Susceptible to coordinated fake review campaigns - No human-in-the-loop

### Why Open Source

Misinformation is too important to solve behind closed doors. The methodology needs to be transparent and auditable. Full research papers with step-by-step calculations are in the `papers/` directory.

Looking for feedback on the approach and contributions (especially: multi-language support, additional platforms, expanded evidence sources).