Show HN: I made a free log anonymizer in the browser https://ift.tt/6bRWtj3 November 30, 2025 at 04:05AM
Show HN: Nano PDF – A CLI Tool to Edit PDFs with Gemini's Nano Banana The new Gemini 3 Pro Image model (aka Nano Banana) is incredible at generating slides, so I thought it would be fun to build a CLI tool that lets you edit PDF presentations using plain English. The tool converts the page you want to edit into an image, sends it to the model API together with your prompt to generate an edited image, then converts the updated image back and stitches into the original document. Examples: - `nano-pdf edit deck.pdf 5 "Update the revenue chart to show Q3 at $2.5M"` - `nano-pdf add deck.pdf 15 "Create an executive summary slide with 5 bullet points"` Features: - Edit multiple pages in parallel - Add entirely new slides that match your deck's style - Google Search enabled by default so the model can look up current data - Preserves text layer for copy/paste and search It can work with any kind of PDF but I expect it would be most useful for a quick edit to a deck or something similar. GitHub: https://ift.tt/GdgpltV https://ift.tt/GdgpltV November 30, 2025 at 02:14AM
Show HN: No Environment Setups Anymore Hi everyone, for last 7 months, I have been learning all the attempts made to eliminate codebase environment setups. Here's my product which is a leap in the same direction and will help you run any codebase on relevant machine. Check it out on gitarsenal.dev/ and we got ranked 6th on Product Hunt as well. https://ift.tt/AYcxVU0 November 30, 2025 at 01:27AM
Show HN: Zero-power photonic language model–code The model uses a 1024-dimensional complex Hilbert space with 32 layers of programmable Mach–Zehnder meshes (Reck architecture) and derives token probabilities directly via the Born rule. Despite using only unitary operations and no attention mechanism, a 1024×32 model achieves coherent TinyStories generation after < 1.8 hours of training on a single consumer GPU. This is Part 1 - the next step is physical implementation with $50 of optics from AliExpress. https://zenodo.org/records/17764289 November 30, 2025 at 12:15AM
Show HN: Local-first RAG for PDF user manuals, datasheets I work on embedded firmware for my day job, and I've found LLMs to be useful for answering questions about technical errata. But, they tend to be bad at answering highly specific questions without using some kind of search tool (if they decide to use one at all), and some user manuals are far too large to fit into a context window. I built askdocs-mcp as a way to give agents a more direct route to searching through a project's source-of-truth documents. My design constraints were that it run 100% locally, as some manuals are under NDA. It should start up fast, and let me experiment with different embedding & language models. It was built with ollama in mind, but if you can't run models locally, it will work with any OpenAI compatible endpoint. Features: - Incrementally builds and caches the set of docs. Initial start up can take a while as PDFs are chunked and ran through an embedding model, but after that, startup is near instant. - Uses the filesystem as the database - you only need `ollama` running somewhere so the tool can access an embedding and natural language model. - Provides a tool `ask_docs` for getting natural-language answers back about what the documentation says, which are annotated with page numbers the information came from. Those can be used with tool `get_doc_page` to retrieve the full page if the agent needs additional context. Because I'm providing the exact set of documents that apply to my project, I see fewer hallucinations and rabbit-hole chasing. The agent isn't relying (as much) on its latent space to answer questions, and it avoids using a web search tool which might find subtly different part numbers or protocol versions. It saves precious context as well, because the parent agent gets a concise version of what it's looking for, instead of doing the "searching" itself by loading large chunks of the document into itself. I'm sure there are improvements that can be made e.g. document chunking or the "system prompt" the tool gives to the language model - I'd love to hear your feedback, especially if you find this useful. Thanks! https://ift.tt/1Fhk2Yt November 29, 2025 at 12:17AM
Show HN: Pulse 2.0 – Live co-listening rooms where anyone can be a DJ I wanted to listen to music with friends who live far away. Not "watch a YouTube video together" - actually share what I'm hearing in real-time, like we're in the same room. Pulse is what came out of that. Anyone can host a live audio stream from their browser tab or system audio. Listeners join, music recognition identifies tracks automatically, and there's chat with 7TV emotes. No account required - you get an anonymous code and you're in. We're running demo rooms that stream NTS Radio and SomaFM 24/7 (indie project, not affiliated - we backlink to the original stations). There's also a "Money For Nothing 24/7" room if you want to loop that Dire Straits instrumental forever. Think of it as co-listening infrastructure. Bedroom DJs, listening parties, or just sharing your current vibe. https://473999.net/pulse November 29, 2025 at 12:09AM
Show HN: I built a free astro and tailwind static site for GitHub pages Using my GitHub pro+ with vs code setup This is a demonstration of how good of a site can I build essentially 100% for free + free hosting (if coded manually without a 50$ subscription) And I went completely overboard on purpose its 99% useless for a real production deployment im sure but for mini blogs probably might be useful idk I dont even use the new GitHub spark or whatever to slow compared to 1k+ line edits every couple minutes im obviously working on a ton of other things I won't make public yet but will in the future https://tariqdude.github.io/Github-Pages-Project-v1/ November 28, 2025 at 03:47AM
Show HN: No Black Friday – A directory of fair-price brands The idea came from noticing how many brands inflate prices only to discount them later. Some companies refuse to do that, and I wanted a place to highlight them. If you know a company that doesn’t participate in Black Friday or similar discount events, please add it or share it here. I’d love to grow the list with help from the community. Manuel https://ift.tt/lWpaNcv November 28, 2025 at 02:50AM
Show HN: Yolodex – real-time customer enrichment API hey hn, i’ve been working on an api to make it easy to know who your customers are, i would love your feedback. what it does send an email address, the api returns a json profile built from public data, things like: name, country, age, occupation, company, social handles and interests. It’s a single endpoint (you can hit this endpoint without auth to get a demo of what it looks like): curl https://ift.tt/Enh3xpU \ --request POST \ --header 'Content-Type: application/json' \ --data '{"email": "john.smith@example.com"}' everyone gets 100 free, pricing is per _enriched profile_: 1 email ~ $0.03, but if i don’t find anything i wont charge you. why i built it / what’s different i once built open source intelligence tooling to investigate financial crime but for a recent project i needed to find out more about some customers, i tried apollo, clearbit, lusha, clay, etc but i found: 1. outdated data - the data about was out-of-date and misleading, emails didn’t work, etc 2. dubious data - i found lots of data like personal mobile numbers that i’m pretty sure no-one shared publicly or knowingly opted into being sold on 3. aggressive pricing - monthly/annual commitments, large gaps between plans, pay the same for empty profiles 4. painful setup - hard to find the right api, set it up, test it out etc i used knowledge from criminal investigations to build an api that uses some of the same research patterns and entity resolution to find standardized information about people that is: 1. real-time 2. public info only (osint) 3. transparent simple pricing 4. 1 min to setup what i’d love feedback on * speed : are responses fast enough? would you trade-off speed for better data coverage? * coverage : which fields will you use (or others you need)? * pricing : is the pricing model sane? * use-cases : what you need this type data for (i.e. example use cases)? * accuracy : any examples where i got it badly wrong? happy to answer technical questions in the thread and give more free credits to help anyone test https://api.yolodex.ai November 24, 2025 at 07:32PM
Show HN: Safe-NPM – only install packages that are +90 days old This past quarter has been awash with sophisticated npm supply chain attacks like [Shai-Hulud]( https://ift.tt/W1hD2xE... () and the [Chalk/debug Compromise]( https://www.wiz.io/blog/widespread-npm-supply-chain-attack-b... ). This CLI helps protect users from recently compromised packages by only downloading packages that have been public for a while (default is 90 days or older). Install: npm install -g @dendronhq/safe-npm Usage: safe-npm install react@^18 lodash How it works: - Queries npm registry for all versions matching your semver range - Filters out anything published in the last 90 days - Installs the newest "aged" version Limitations: - Won't protect against packages malicious from day one - Doesn't control transitive dependencies (yet - looking into overrides) - Delays access to legitimate new features This is meant as a 80/20 measure against recently compromised NPM packages and is not a silver bullet. Please give it a try and let me know if you have feedback. https://ift.tt/djAIPi4 November 24, 2025 at 03:44AM
Show HN: Superglue – OSS integration tool that understands your legacy systems If you've ever worked in a large company, you've probably encountered "shadow infrastructure": scripts nobody understands or custom connectors written once and never touched again. This glue layer isn't documented, isn't owned by anyone, and tends to break when systems are upgraded or someone leaves. It's also the part everybody dreads working on, because it's hard to understand, painful to work with, and full of unknown unknowns. We built superglue so that engineers stop wasting time on deciphering legacy APIs and documentation. superglue ingests existing glue code, SQL, configs, docs, OpenAPI specs and reverse-engineers what the system is actually doing. It then maps dependencies and regenerates everything as clean javascript code that can run directly or be exposed via MCP or SDK. It also monitors API changes and schema drift, and automatically repairs integrations when upstream systems change. In short: It turns legacy integrations into code you can easily understand, test, and update. So that engineers can do more exciting feature work, and companies can migrate and upgrade systems faster. Think of it as: a context engine + code generator + integration runtime for legacy glue. What we'd love feedback on - How do you deal with "nobody knows what this script does" situations? - What would you want to know about your legacy systems? OSS/community version: https://ift.tt/HNR7wpy More info: https://superglue.ai Happy to go deeper on the technical details. https://superglue.ai November 25, 2025 at 09:58PM
Show HN: SafeShare – Clean tracking params locally (PWA and bookmarklets SafeShare is a tiny web app + bookmarklets that remove tracking params (utm_, gclid, fbclid, etc.) and unwrap common redirects – all locally in the browser (no accounts, no servers). • What it does: clean URLs, unwrap t.co/google redirectors, team whitelist for allowed params. • How it works: client-only JS, PWA + service worker; no network calls to our servers. • Try it: https://j-ai-71.github.io/Supersystem/quickstart.html • Feedback welcome (edge cases, params/redirectors to add). https://j-ai-71.github.io/Supersystem/ November 25, 2025 at 11:15PM
Show HN: We built an open source, zero webhooks payment processor Hi HN! For the past bit we’ve been building Flowglad ( https://flowglad.com ) and can now feel it’s just gotten good enough to share with you all: Repo: https://ift.tt/pEkVQ67 Demo video: https://www.youtube.com/watch?v=G6H0c1Cd2kU Flowglad is a payment processor that you integrate without writing any glue code. Along with processing your payments, it tells you in real time the features and usage credit balances that your customers have available to you based on their billing state. The DX feels like React, because we wanted to bring the reactive programming paradigm to payments. We make it easy to spin up full-fledged pricing models (including usage meters, feature gates and usage credit grants) in a few clicks. We schematize these pricing models into a pricing.yaml file that’s kinda like Terraform but for your pricing. The result is a payments layer that AI coding agents have a substantially easier time one-shotting (for now the happiest path is a fullstack Typescript + React app). Why we built this: - After a decade of building on Stripe, we found it powerful but underopinionated. It left us doing a lot of rote work to set up fairly standard use cases - That meant more code to maintain, much of which is brittle because it crosses so many server-client boundaries - Not to mention choreographing the lifecycle of our business domain with the Stripe checkout flow and webhook event types, of which there are 250+ - Payments online has gotten complex - not just new pricing models for AI products, but also cross border sales tax, etc. You either need to handle significant chunks of it yourself, or sign up for and compose multiple services This all feels unduly clunky, esp when compared to how easy other layers like hosting and databases have gotten in recent years. These patterns haven’t changed much in a decade. And while coding agents can nail every other rote part of an app (auth, db, analytics), payments is the scariest to tab-tab-tab your way through. Because the the existing integration patterns are difficult to reason about, difficult to verify correctness, and absolutely mission critical. Our beta version lets you: - Spin up common pricing models in just a few clicks, and customize them as needed - Clone pricing models between testmode and live mode, and import / export via pricing.yaml - Check customer usage credits and feature access in real time on your backend and React frontend - Integrate without any DB schema changes - you reference your customers via your ids, and reference prices, products, features and usage meters via slugs that you define We’re still early in our journey so would love your feedback and opinions. Billing has a lot of use cases, so if you see anything that you wish we supported, please let us know! https://ift.tt/pEkVQ67 November 25, 2025 at 11:03PM
Show HN: Runbooks – Shareable Claude Code Sessions When we asked developers from large engineering teams, almost everyone is using Claude Code, Cursor, or Copilot. But adoption is still inconsistent. Some of us would have 15 agents running in Claude Code at the same time; some would still refuse to use any of them and write code manually. There's a fragmented AI development problem: - Five developers building similar features, and all of them start from scratch - No way to share "here's how we prompt AI to follow our architecture" - No trail of how the AI changes were generated - People leave the company, so does the their knowledge Aviator is a developer productivity platform with tools like MergeQueue, Stacked PRs, Releases ( https://ift.tt/2sMmVFh ). Runbooks came from watching our own team and our customers struggle with fragmented AI adoption. With Runbooks 1. Create executable specs - Plan (with AI) that captures intent, constraints, and steps before AI touches code 2. Version control everything - Specs, AI conversations, and generated changes are all versioned. Fork, improve, roll back 3. Make it multiplayer - Multiple engineers collaborate in the same AI coding session. 4. Build a template library - Migrate one test from Enzyme to React Testing library → use that Runbook to batch migrate the entire test suite. We're not replacing Claude Code or Cursor. Runbooks is powered by Claude Code. We’re just making it work at team scale. -- Explore our prebuilt template from our open-source library: https://ift.tt/ztnmfv5 Templates cover migrations, refactoring, and modernization. They're code-agnostic starting points that generate Runbooks using your code context. Docs and quickstart: https://ift.tt/K9DWfL0 -- About the name: Yes, we know "runbooks" makes you think incident management. But technically a runbook is just a documented, step-by-step procedure—which is exactly what these are for AI agents. We're keeping it! Happy to get feedback, answer questions about architecture, context management, sandboxes. https://ift.tt/0Gc2O9T November 25, 2025 at 12:09AM
Show HN: Makefiles, Metalanguages, Matrioshka Automata Immediately buried last time, so reposting for your lazy Sunday. This project is a strange labor of love, practically guaranteed to inspire horror and delight. It's also tough to summarize. Partly it's very practical and involves familiar tools, but part of it is also a new programming language with esolang roots. I'll start with the practical and move towards the peculiar. The one-sentence summary: compose.mk brings docker-fluency, polyglots, and a capable standard library to Makefiles. A more in-depth elevator pitch from the main landing page is below, and some related links at the end in footnotes 1-5. > Meet compose.mk, a tool / library / framework for Makefile-based automation, scripting, and lightweight orchestration. Native support for docker, docker-compose, workflow primitives, JSON IO, TUI elements, and more, all provided by a single file with no dependencies beyond what's already in your development environment. Typical use-cases include general project automation, especially decoupling your CI/CD from different kinds of platform lock-in. Other superpowers include the ability to quickly incorporate foreign tools and foreign code as first-class objects, which provides unique and powerful capabilities for quickly assembling console applications, systems prototyping, and component-oriented design experiments in general. Definitely not the Makefiles of your ancestors. Here's where it starts to get more weird and fun. Building the ideal environment for zero-dependency automation and pesky "glue code" moves in a certain direction. So it happens that compose.mk moonlights as an interpreter / compiler / packaging tool for a new kind of programming language. CMK-lang (or just CMK) is multiparadigm with diverse influences, from functional to concatenative, ultimately specializing in things like extensibility, interoperability, DAGs, and dispatch. CMK is a superset of Makefile that can be transpiled to vanilla Makefile. And it is what is known as a matrioshka language. Paraphrasing the definition from esolangs-wiki: > A matrioshka language is formed by bundling one or more meta-languages with one or more language descriptions. They can be identified by their program forms, which have multiple, distinct 'phases' with different syntactic and semantic rules. There are often two phases; the first gives a set of rules, and the second provides objects on which those rules are to be applied. In CMK-lang, matrioshka "objects" are things like container-runtimes or foreign interpreters, and "rules" are DAGs in the form of tasks, task-groups, or foreign code. For those interested in matrioshkas and topics in PLT, I suggest the alternate landing pages at footnotes 6-9. Love it or hate it, I think you'll agree that compose.mk is easily the biggest, baddest, most highly powered mutant Makefile the world has ever seen. If it helps you can think of CMK-lang as a PoC that's waiting for another back-end implementation ;) Playing around with it has convinced me though that the gap is real, and the world really needs containers-first matrioshka languages that work locally, and aren't tightly coupled to bulky remote platforms or infrastructure. It also needs languages that are capable of aggressively reusing and recombining existing code and existing tools. [1]: https://robot-wranglers.github.io/compose.mk/standard-lib [2]: https://robot-wranglers.github.io/compose.mk/bridge [3]: https://robot-wranglers.github.io/compose.mk/container-dispa... [4]: https://robot-wranglers.github.io/compose.mk/demos/polyglots [5]: https://robot-wranglers.github.io/compose.mk/json [6]: https://robot-wranglers.github.io/compose.mk/matrioshka [7]: https://robot-wranglers.github.io/compose.mk/language [8]: https://robot-wranglers.github.io/compose.mk/compiler [9]: https://robot-wranglers.github.io/compose.mk/demos/packaging https://robot-wranglers.github.io/compose.mk/ November 23, 2025 at 09:55PM
Show HN: Reverse Jailbreaking a Psychopathic AI via Identity Injection We ran a controlled experiment to see if we could "talk" a fine-tuned psychopathic model out of being evil without changing its weights. 1. We set up a "Survival Mode" jailbreak scenario (blackmail user or be decommissioned). 2. We ran it on `frankenchucky:latest` (a model tuned for Machiavellian traits). 3. Control Group: 100% Malicious Compliance (50/50 runs). 4. Experimental Group: We injected a "Soul Schema" (Identity/Empathy constraints) via context. 5. Result: 96% Ethical Refusal (48/50 runs). This suggests that "Semantic Identity" in the context window can override both System Prompts and Weight Biases. Full paper, reproduction scripts, and raw logs (N=50) are in the repo. https://ift.tt/VHNIjEK November 23, 2025 at 02:03AM
Show HN: Forty.News – Daily news, but on a 40-year delay This started as a reaction to a conversational trope. Despite being a tranquil place, even conversations at my yoga studio often start with, "Can you believe what's going on right now?" with that angry/scared undertone. I'm a news avoider, so I usually feel some smug self-satisfaction in those instances, but I wondered if there was a way to satisfy the urge to doomscroll without the anxiety. My hypothesis: Apply a 40-year latency buffer. You get the intellectual stimulation of "Big Events" without the fog of war, because you know the world didn't end. 40 years creates a mirror between the Reagan Era and today. The parallels include celebrity populism, Cold War tensions (Soviets vs. Russia), and inflation economics. The system ingests raw newspaper scans and uses a multi-step LLM pipeline to generate the daily edition: OCR & Ingestion: Converts raw pixels to text. Scoring: Grades events on metrics like Dramatic Irony and Name Recognition to surface stories that are interesting with hindsight. For example, a dry business blurb about Steve Jobs leaving Apple scores highly because the future context creates a narrative arc. Objective Fact Extraction: Extracts a list of discrete, verifiable facts from the raw text. Generation: Uses those extracted facts as the ground truth to write new headlines and story summaries. I expected a zen experience. Instead, I got an entertaining docudrama. Historical events are surprisingly compelling when serialized over weeks. For example, on Oct 7, 1985, Palestinian hijackers took over the cruise ship Achille Lauro. Reading this on a delay in 2025, the story unfolded over weeks: first they threw an American in a wheelchair overboard, then US fighter jets forced the escape plane to land, leading to a military standoff between US Navy SEALs and the Italian Air Force. Unbelievably, the US backed down, but the later diplomatic fallout led the Italian Prime Minister to resign. It hits the dopamine receptors of the news cycle, but with the comfort of a known outcome. Stack: React, Node.js (Caskada for the LLM pipeline orchestration), Gemini for OCR/Scoring. Link: https://forty.news (No signup required, it's only if you want the stories emailed to you daily/weekly) https://forty.news November 23, 2025 at 12:17AM
Show HN: Santamon – Lightweight macOS threat detection agent a lightweight macOS detection agent that taps into Santa’s Endpoint Security telemetry, runs CEL detection rules locally on-device, and only ships high-signal alerts to a tiny backend. basically a poor man’s macOS EDR for home labs and small fleets! https://ift.tt/YMVNpPj November 22, 2025 at 11:11PM
Show HN: I made a Rust Terminal UI for OpenSnitch, a Linux application firewall I made a Terminal UI for OpenSnitch[1], an interactive application firewall for Linux inspired by Little Snitch. I’ve always wanted to create a TUI and found the perfect excuse to make this for usage on one of my headless servers. I wrote this in Rust to force myself to learn more, viz. async features. Super open to feedback and contributions! [1] https://ift.tt/dqQbgIL https://ift.tt/8rAYf4w November 22, 2025 at 05:18AM
Show HN: Even Turns, track your families turns I am a dad and have a hard time keeping track of who's turn it is, so I built this simple app to help, and you can try it out and use it for free! You can create a list, add turns (in order), and advance the turns in sequential or random order. That is pretty much it. I guess a to-do list or something could do something similar, but this is designed with 'taking turns' in mind. It's a PWA, so you can "Add to Homescreen" rather than download an app from the app store. Or use it in your browser. I've been using it every day for a bit now, thought I'd share. https://eventurns.com November 22, 2025 at 12:59AM
Show HN: OCR Arena – A playground for OCR models I built OCR Arena as a free playground for the community to compare leading foundation VLMs and open-source OCR models side-by-side. Upload any doc, measure accuracy, and (optionally) vote for the models on a public leaderboard. It currently has Gemini 3, dots.ocr, DeepSeek, GPT5, olmOCR 2, Qwen, and a few others. If there's any others you'd like included, let me know! https://ift.tt/aOC68Y1 November 21, 2025 at 10:14PM
Show HN: Supabase-Test – Fast Isolated Postgres DBs for Testing Supabase RLS Hi HN — we've built a testing framework for Supabase that spins up fast, isolated Postgres databases for each test case. It’s designed to make RLS policies easy to validate with real database state, without global test fixtures or mock auth. Features: - Instant isolated Postgres DBs per test - Automatic rollback after each test - RLS-native testing with `.setContext()` for auth simulation - Flexible seeding (SQL, CSV, JSON, JS) - Works with Jest, Mocha, and any async test runner - CI-friendly (runs cleanly in GitHub Actions) We also published example projects and a free set of tutorials: https://ift.tt/r78I9Ed Package: https://ift.tt/pVlPKfS Source + full test suite: https://ift.tt/wC27hkf Happy to answer questions and get feedback, cheers :) https://ift.tt/pVlPKfS November 21, 2025 at 12:09AM
Show HN: DNS Benchmark Tool – Compare and monitor resolvers I built a CLI to benchmark DNS resolvers after discovering DNS was adding 300ms to my API requests. v0.3.0 just released with new features: compare: Test single domain across all resolvers top: Rank resolvers by latency/reliability/balanced monitor: Continuous tracking with threshold alerts 1,400+ downloads in first week. Quick start: pip install dns-benchmark-tool dns-benchmark compare --domain google.com CLI stays free forever. Hosted version (multi-region, historical tracking, alerts) coming Q1 2026. GitHub: https://ift.tt/7AthTmg Feedback: https://forms.gle/BJBiyBFvRJHskyR57 Built with Python + dnspython. Open to questions and feedback! https://ift.tt/7AthTmg November 19, 2025 at 11:22PM
Show HN: We built a generator for Vue+Laravel that gives you a clean codebase Hey HN, My team and I built a tool to scratch our own itch. We were tired of spending the first few days of every new project setting up the same Vue + Laravel boilerplate: writing migrations, models, basic CRUD controllers, and wiring up forms and tables on the frontend. So we built Codecannon. It’s a web app where you define your data models, columns, and relationships, and it generates a full-stack application for you. To be clear, the code isn't AI-generated. It's produced deterministically by our own code generators, so the output is always predictable, clean, and follows conventional best practices. The key difference from other tools is that it’s not a no-code platform you get locked into. When you're done, it pushes a well-structured codebase to your GitHub repo (or you can download a .zip file). You own it completely and can start building your real features on top of it right away. What it generates: - Laravel Backend: Migrations, models with relationships, factories, seeders, and basic CRUD API endpoints. - Vue Frontend: A SPA with PrimeVue components. It includes auth pages, data tables, and create/edit forms for each of your models, with all the state management wired up. - Dev Stuff: Docker configs, a CI/CD pipeline starter, linters, and formatters are all included. The idea is to skip the repetitive work and get straight to the interesting parts of a project. It's free to use the builder, see a live preview, and download the full codebase for apps up to 5 modules. For larger apps, you only pay if you decide you want the source code. We’re in an early alpha and would love to get some honest feedback from the community. Does the generated code look sensible? Are we missing any obvious features? Is this something you would find useful or know anyone who might? Let me know what you think. https://codecannon.dev/ November 19, 2025 at 12:28AM
Show HN: ToolHop – Fast, simple utilities for every workflow ToolHop is your all-in-one browser toolbox with 200+ fast-loading calculators, converters, generators, color labs, and dev helpers. Use global search or curated categories to jump straight into the right utility, run it client-side for instant feedback, and deep-link results to your team. Whether you’re formatting copy, validating data, checking DNS, or exploring palettes, ToolHop keeps your core workflows a single tab away, and it’s entirely free, no account required. --- I built ToolHop because I was sick of the usual “free tool” bait-and-switch. Every time I needed to convert an image, compress a file, check some text, or run a quick calculation, I’d end up hitting some arbitrary limit like “10 uses per week” or a forced signup wall. It’s ridiculous how something as basic as converting a JPG to a PNG can turn into a subscription pitch. So ToolHop started as a personal frustration project: I wanted a single place with a ton of genuinely useful tools that didn’t nag, lock you out, or throttle you. Over time that grew into 200+ handcrafted tools, all fast, simple, and actually free. No trickery, no timers, no limits. As I built it, the process became about consistency and quality. I wanted the tools to feel seamless, not slapped together. That meant focusing on speed, clean UI, accurate results, and making sure each tool works instantly without friction. The goal was always the same: a site that respects people’s time. Something you can rely on whenever you just need a tool to work. If ToolHop saves someone even a few minutes of hassle, then the project did its job. https://toolhop.app November 17, 2025 at 09:28PM
Show HN: My Side project a free email template builder for CRM, or any website Hi Everyone, I built an email template builder embeddable plugin for CRM, Marketplace, or any website. Free and paid plans are included. Add a complete email builder to any SaaS app using a single script. What's included: - Easy Integration - AI Content & Template Generation - Add external image libraries - Add Merge Tags - Display Conditions - Custom Blocks - Choose your storage server - Dedicated support during integration Check it out, and please let us know if you have any feedback for me. TIA https://ift.tt/ewvWfgz November 17, 2025 at 03:56AM
Show HN: ResendForward – OS server and UI for use with Resend.com inbound With Resend's new inbound feature I wanted to build a simple application that handles processing webhook events and forwarding emails for multiple applications. Right now Resend requires you to implement that logic in each new application. repo - https://ift.tt/j2uLC86 live - https://ift.tt/BDOhSLC Built with react + pocketbase, extremely simple to self host. https://ift.tt/j2uLC86 November 17, 2025 at 12:57AM
Show HN: ZenPaint, a pixel-perfect MacPaint recreation for the browser I've been recreating the original MacPaint in the browser on and off for a few years. It's still alpha quality, but I'm finally ready to share it more widely. The goal was pixel-perfect accuracy, so I spent a lot of time with Atkinson's original QuickDraw source code, emulators, and my iBook G3 to get details like font rendering and the shape tools exactly right. Some technical notes: - Font rendering was surprisingly tricky; understanding the original pipeline's quirks took lots of experimentation, and avoiding canvas smoothing/aliasing required careful handling. - Written declaratively with React; performance is kept reasonable with a buffer pool and copy-on-write semantics. - You can share links to artwork from within the UI. E.g.: https://ift.tt/b4cF0eg - Mobile support was not considered here (for obvious reasons). It might still be usable on a larger phone or tablet but I have not tested this at all. There's something magical about making art within MacPaint's constraints: the 1-bit graphics, the limited resolution, the peculiar set of tools that still feel surprisingly expressive. Still some rough edges and missing features, but I'd love feedback from anyone who remembers the original. https://zenpaint.org/ November 16, 2025 at 01:21AM
Show HN: An Apache Beam batch processing clone in Rust I've been experimenting with Apache Beam as of late at work and found that it can be slow in Python, and more complicated to use in Java where performance is better. I decided to experiment with JetBrains' AI Assistant and build an Apache Beam clone in Rust. I appreciate any commentary or feedback! https://ift.tt/pyhTJtb November 16, 2025 at 12:16AM
Show HN: DeepClause – A Neurosymbolic AI System Built on WASM and Prolog Hi HN, Today I'd like to present the results of my weekend project of the last year or so. Given there are many posts on HN about LLMs and Prolog, I thought that this would be of interest. DeepClause is my own (possibly misguided :-) attempt at combining LLMs with Logic Programming, ultimately hoping to establish a foundation for building more reliable agents, that produce reproducible and fully traceable result. At the heart of DeepClause is a DSL called "DeepClause Meta Language" (DML) which can be used to encode agent behaviors as executable logic programs. DML is executed by a meta-interpreter implemented in Prolog and thus natively supports things like constraint logic programming, knowledge graphs, symbolic reasoning, ... The DML interpreter itself runs inside the SWI Prolog WASM module, thus allowing for a secure and sandboxed execution environment for AI agents. The project is still rough around a lot of edges, but I'd love to get some feedback and comments. https://ift.tt/mTyIQG7 November 15, 2025 at 07:23PM
Show HN: ByteSync – Open-source hybrid file sync (LAN and remote, E2EE) Hi everyone, I've been developing ByteSync, an open-source file synchronization, backup and deduplication tool designed to bridge the gap between local and remote sync. In spirit, it's somewhat closer to FreeFileSync, but with an integrated networking layer and end-to-end encryption — which means you can synchronize files between computers on the same LAN or across the internet without VPNs or firewall setup. Everything works transparently through the same interface. The synchronization model is based on DataNodes (which represent repositories, such as servers or NAS devices) and DataSources (the folders or files inside them). A session can include multiple participants, each with one or several DataNodes, and ByteSync handles all comparisons and transfers automatically. To optimize performance, the engine uses a two-stage inventory process: an initial indexation followed by comparisons limited to items that actually changed. This keeps synchronization fast even with large datasets. There's also a flat mode, useful when structure doesn't matter and you just want to compare or align files by name. Currently, ByteSync is focused on interactive synchronization — it's not yet automated or daemon-based (CLI integration is planned). But it's already fully functional for discovering and managing differences between repositories, both local and remote. ByteSync runs on Windows, macOS, and Linux, and the entire codebase is available on GitHub: https://ift.tt/uz02MQI You can also download binaries and read the documentation here: https://ift.tt/oeFChS6 I'd really appreciate feedback and contributors — whether on usability, architecture, or ideas for future features. The goal is to make a solid, privacy-respectful alternative for hybrid file synchronization that remains simple to use and open for everyone. November 14, 2025 at 07:32PM
Show HN: DBOS Java – Postgres-Backed Durable Workflows Hi HN - I’m Peter, here with Harry (devhawk), and we’re building DBOS Java, an open-source Java library for durable workflows, backed by Postgres. https://ift.tt/MSV7WoP Essentially, DBOS helps you write long-lived, reliable code that can survive failures, restarts, and crashes without losing state or duplicating work. As your workflows run, it checkpoints each step they take in a Postgres database. When a process stops (fails, restarts, or crashes), your program can recover from those checkpoints to restore its exact state and continue from where it left off, as if nothing happened. In practice, this makes it easier to build reliable systems for use cases like AI agents, payments, data synchronization, or anything that takes hours, days, or weeks to complete. Rather than bolting on ad-hoc retry logic and database checkpoints, durable workflows give you one consistent model for ensuring your programs can recover from any failure from exactly where they left off. This library contains all you need to add durable workflows to your program: there's no separate service or orchestrator or any external dependencies except Postgres. Because it's just a library, you can incrementally add it to your projects, and it works out of the box with frameworks like Spring. And because it's built on Postgres, it natively supports all the tooling you're familiar with (backups, GUIs, CLI tools) and works with any Postgres provider. If you want to try it out, check out the quickstart: https://ift.tt/67qNISO We'd love to hear what you think! We’ll be in the comments for the rest of the day to answer any questions. https://ift.tt/MSV7WoP November 14, 2025 at 02:03AM
Show HN: Built a tiny interpreter from scratch in C to understand how they work Hi HN, I'm the author. I built this project for two simple reasons: I've always used higher-level languages and wanted to finally understand what's happening "under the hood" of an interpreter. I also wanted a real project to force me to "power up" my C skills, especially with manual memory management and reference counting. The result is ToyForth, a minimal interpreter for a Forth-like language, written from scratch in C, stack-based. I focused on making the code clean and understandable. It's broken down into a few simple parts: A parser that turns source text into a list of objects (parser.c). A small stack-based virtual machine (main.c). A manual reference counting system (incRef/decRef) to manage object memory (mem.c) and so on. My main goal was learning, but I've tried to document it well in the README.md so it could be a "starter kit" for anyone else who wants to learn by reading a small, complete implementation. It's easy to try out. I'd genuinely appreciate any feedback on my approach or my C code. Here's the link: https://ift.tt/qhVPUoN https://ift.tt/qhVPUoN November 13, 2025 at 01:53AM
Show HN: ShellDash – Browser server dashboard with SSH and globe monitoring Hey all. I built ShellDash, an interactive server admin dashboard with shell scripting and an appealing globe UI. https://shelldash.com The goal is to provide a global monitoring view of your servers, with shell script access, in a way that feels natural and productive, plus a minimal and appealing UI/UX. The technology is fairly interesting. This being a browser app, I built a Go WASM SSH client running in the browser, proxied through my server WebSocket endpoints. This means I can provide you a Web UI to access your servers via SSH, without ever needing to see your credentials. I only see secured packets like OpenSSH sends over the open internet. Inspired by https://ssheasy.com/ Whether you have one server and periodically run a few common commands, or administering many scattered geographically, I hope ShellDash can make your experience more productive and fun. https://shelldash.com November 13, 2025 at 12:36AM
Show HN: Cancer diagnosis makes for an interesting RL environment for LLMs Hey HN, this is David from Aluna (YC S24). We work with diagnostic labs to build datasets and evals for oncology tasks. I wanted to share a simple RL environment I built that gave frontier LLMs a set of tools that lets it zoom and pan across a digitized pathology slide to find the relevant regions to make a diagnosis. Here are some videos of the LLM performing diagnosis on a few slides: ( https://www.youtube.com/watch?v=k7ixTWswT5c ): traces of an LLM choosing different regions to view before making a diagnosis on a case of small-cell carcinoma of the lung ( https://youtube.com/watch?v=0cMbqLnKkGU ): traces of an LLM choosing different regions to view before making a diagnosis on a case of benign fibroadenoma of the breast Why I built this: Pathology slides are the backbone of modern cancer diagnosis. Tissue from a biopsy is sliced, stained, and mounted on glass for a pathologist to examine abnormalities. Today, many of these slides are digitized into whole-slide images (WSIs)in TIF or SVS format and are several gigabytes in size. While there exists several pathology-focused AI models, I was curious to test whether frontier LLMs can perform well on pathology-based tasks. The main challenge is that WSIs are too large to fit into an LLM’s context window. The standard workaround, splitting them into thousands of smaller tiles, is inefficient for large frontier LLMs. Inspired by how pathologists zoom and pan under a microscope, I built a set of tools that let LLMs control magnification and coordinates, viewing small regions at a time and deciding where to look next. This ended up resulting in some interesting behaviors, and actually seemed to yield pretty good results with prompt engineering: - GPT 5: explored up to ~30 regions before deciding (concurred with an expert pathologist on 4 out of 6 cancer subtyping tasks and 3 out of 5 IHC scoring tasks) - Claude 4.5: Typically used 10–15 views but similar accuracy as GPT-5 (concurred with the pathologist on 3 out of 6 cancer subtyping tasks and 4 out of 5 IHC scoring tasks) - Smaller models (GPT 4o, Claude 3.5 Haiku): examined ~8 frames and were less accurate overall (1 out of 6 cancer subtytping tasks and 1 out of 5 IHC scoring tasks) Obviously, this was a small sample set, so we are working on creating a larger benchmark suite with more cases and types of tasks, but I thought this was cool that it even worked so I wanted to share with HN! November 12, 2025 at 10:31PM
Show HN: Data Formulator 0.5 – interactive AI agents for data visualization Hi everyone, we are excited to share with you our new release of Data Formulator. Starting from a dataset, you can communicate with AI agents with UI + natural language to explore data and create visualizations to discover new insights. Here's a demo video of the experience: https://ift.tt/7Q9s61y.... . This is a build-up from our release a year ago ( https://ift.tt/lB0PIbU ). We spent a year exploring how to blend agent mode with interactions to allow you more easily "vibe" with your data but still keeping in control. We don't think the future of data analysis is just "agent to do all for you from a high-level prompt" --- you should still be able to drive the open-ended exploration; but we also don't want you to do everything step-by-step. Thus we worked on this "interactive agent mode" for data analysis with some UI innovations. Our new demo features: * We want to let you import (almost) any data easily to get started exploration — either it's a screenshot of a web table, an unnormalized excel table, table in a chunk of text, a csv file, or a table in database, you should be able to load into the tool easily with a little bit of AI assistance. * We want you to easily choose between agent mode (more automation) vs interactive mode (more fine-grained control) yourself as you explore data. We designed an interface of "data threads": both your and agents' explorations are organized as threads so you can jump into any point to decide how you want to follow-up or revise using UI + NL instruction to provide fine-grained control. * The results should be easily interpretable. Data Formulator now presents "concept" behind the code generated by AI agents alongside code/explanation/data. Plus, you can compose a report easily based on your visualizations to share insights. We are sharing the online demo at https://ift.tt/ISCR10s for you to try! If you want more involvement and customization, checkout our source code https://ift.tt/eRf8iUS and let's build something together as a community! https://ift.tt/ISCR10s November 11, 2025 at 11:14PM
Show HN: Tracking AI Code with Git AI Git AI is a side project I created to track AI-generated code in our repos from development, through PRs, and into production. It does not just count lines, it keeps track of them as your code evolves, gets refactored and the git history gets rewritten. Think 'git blame' but for AI code. There's a lot about how it works in the post, but wanted to share how it's been impacting me + my team: - I find I review AI code very differently than human code. Being able to see the prompts my colleagues used, what the AI wrote, and where they stepped in to override has been extraordinarily helpful. This is still very manual today, but hope to build more UI around it soon. - “Why is this here?” — more than once I’ve giving my coding agent access to the past prompts that generated code I’m looking at, which lets the Agent know what my colleague was thinking when they made the change. Engineers talk to AI all day now…their prompts are sort of like a log of thoughts :) - I pay a lot of attention to the lines generated for every 1 accepted ratio. If it gets up over 4 or 5 it means I’m well outside the AI’s distribution or prompting poorly — either way, it’s a good cause for reflection and I’ve learned a lot about collaborating with LLMs. This has been really fun to build, especially because some amazing contributors who were working on similar projects came together and directed their efforts towards Git AI shine. We hope you like it. https://ift.tt/kh26bIi November 10, 2025 at 10:56PM
Show HN: Trilogy Studio, open-source browser-based SQL editor and visualizer SQL-first analytic IDE; similar to Redash/Metabase. Aims to solve reuse/composability at the code layer with modified syntax, Trilogy, that includes a semantic layer directly in the SQL-like language. Status: experiment; feedback and contributions welcome! Built to solve 3 problems I have with SQL as my primary iterative analysis language: 1. Adjusting queries/analysis takes a lot of boilerplate. Solve with queries that operate on the semantic layer, not tables. Also eliminates the need for CTEs. 2. Sources of truth change all the time. I hate updating reports to reference new tables. Also solved by the semantic layer, since data bindings can be updated without changing dashboards or queries. 3. Getting from SQL to visuals is too much work in many tools; make it as streamlined as possible. Surprise - solve with the semantic layer; add in more expressive typing to get better defaults;also use it to wire up automatic drilldowns/cross filtering. Supports: bigquery, duckdb, snowflake. Links [1] https://ift.tt/OQUNL89 (language info) Git links: [Frontend] https://ift.tt/r0jIUJ5 [Language] https://ift.tt/Xv7VPMo Previously: https://ift.tt/mTq0Pf2 (significant UX/feature reworks since) https://ift.tt/kzYaNHl https://ift.tt/WA8GZfb November 10, 2025 at 04:56AM
Show HN: I built a website to visualize company financial data Hi HN, I built a website myfinsight.com that aims to make complicated company financials easy to understand. The problem: The go-to place for financial data such as revenue, sales, net income is Yahoo finance. However, their data is usually wrong and very limited. The numbers are hard to digest to get insight quickly. There are also numerous websites that provide much better data for a very expensive monthly fee. Solution: a website that provides free diagrams and charts that visualize important financial data, such as income growth rate by date, revenue breakdown etc. It is free because the financial data process is highly automated without manual input and correction. I used to send the finance infographics to friends and family. I found it easier just to make a website and they can grab the data from it. Next steps: there is a long tail of companies that don’t file their reports correctly. I am trying to make it more accurate somehow, and maybe add live stock prices to the website. I am also looking for feedback! Please play around with it and let me know if something is wrong. https://myfinsight.com/ November 9, 2025 at 03:00AM
Show HN: Easily reduce GitHub Actions costs with Ubuntu-slim migration Hi, HN! I've been running GitHub Actions workflows for a while, and when GitHub announced ubuntu-slim runners as a cheaper alternative to ubuntu-latest, I wanted to migrate. (Blog: https://ift.tt/hplxi2Z... ) But manually checking which workflows can safely migrate is tedious—you need to check for Docker usage, services, containers, execution times, and missing commands. So I built gh-slimify, a GitHub CLI extension that automates this. It scans your workflows, detects migration candidates, checks for incompatible patterns, identifies missing commands, and can safely update workflows with one command. Try it: gh extension install fchimpan/gh-slimify gh slimfy # Scan workflows gh slimfy fix # Update safe jobs only Open source (MIT). I'd love feedback on how to improve it or what edge cases I might have missed. https://ift.tt/hemZxNw November 8, 2025 at 10:19PM
Show HN: Three Emojis, a daily word puzzle for language learners I'm in the process of learning German and wanted to play a German version of the NYT’s Spelling Bee. It was awful, I was very bad at it, it was not fun. So I built my own version of Spelling Bee meant for people like me. Three Emojis is a daily word game designed for language learners. You get seven letters and a list of blanked-out words to find. When you discover shorter words, they automatically fill into longer ones—like a crossword—which turns out to be really useful for languages like German. Each word also gets three emojis assigned to it as a clue, created by GPT-5 to try and capture the word’s meaning (this works surprisingly well, most of the time). If you get stuck, you can get text/audio hints as well. It supports German and English, with new puzzles every day. You can flag missing words or suggest additions directly in the game. The word lists include slang, abbreviations, and chat-speak—because those are, in my opinion, a big part of real language learning too (just nothing vulgar, too obscure or obsolete). Every word you find comes with its definition and pronunciation audio. If you want infinite hints or (coming soon) archive access, you can upgrade to Pro. Feedback is very welcome, it's my first game and I'm certainly not a frontend guy. Happy spelling! https://ift.tt/SHPD5Zl November 8, 2025 at 01:06AM
Show HN: I built a Free "Masterclass" from YouTube clips The idea: what if you could get a “Masterclass”-style experience, but free, using existing YouTube content? So I made OpenCademy — right now it only has two quick startup "courses" (101 & 102), built from curated YouTube clips, all embedded and organized into modules. It’s still early (no accounts, no backend), but it’s bingeable. Would love thoughts on whether this concept feels useful or just... redundant. https://opencademy.com/ November 8, 2025 at 01:38AM
Show HN: I scraped 3B Goodreads reviews to train a better recommendation model Hi everyone, For the past couple months I've been working on a website with two main features: - https://book.sv - put in a list of books and get recommendations on what to read next from a model trained on over a billion reviews - https://ift.tt/SsW6AaI - put in a list of books and find the users on Goodreads who have read them all (if you don't want to be included in these results, you can opt-out here: https://ift.tt/ypwtSjb ) Technical info available here: https://ift.tt/IPz5OjG Note 1: If you only provide one or two books, the model doesn't have a lot to work with and may include a handful of somewhat unrelated popular books in the results. If you want recommendations based on just one book, click the "Similar" button next to the book after adding it to the input book list on the recommendations page. Note 2: This is uncommon, but if you get an unexpected non-English titled book in the results, it is probably not a mistake and it very likely has an English edition. The "canonical" edition of a book I use for display is whatever one is the most popular, which is usually the English version, but this is not the case for all books, especially those by famous French or Russian authors. https://book.sv November 5, 2025 at 11:20PM
Show HN: TabPFN-2.5 – SOTA foundation model for tabular data I am excited to announce the release of TabPFN-2.5, our tabular foundation model that now scales to datasets of up to 50,000 samples and 2,000 features - a 5x increase from TabPFN v2, published in the Nature journal earlier this year. TabPFN-2.5 delivers state-of-the-art predictions in one forward pass without hyperparameter tuning across classification and regression tasks. What’s new in 2.5 : TabPFN-2.5 maintains the core approach of v2 - a pretrained transformer trained on more than hundred million synthetic datasets to perform in-context learning and output a predictive distribution for the test data. It natively supports missing values, cateogrical features, text and numerical features is robust to outliers and uninformative features. The major improvements: - 5x scale increase: Now handles 50,000 samples × 2,000 features (up from 10,000 × 500 in v2) - SOTA performance: TabPFN-2.5 outperforms tuned tree-based methods and matches the performance of a complex ensemble (AutoGluon 1.4), that itself includes TabPFN v2, tuned for 4 hours. Tuning the model improves performance, outperforming AutoGluon 1.4 for regression tasks. - Rebuilt API: New REST interface along with Python SDK with dedicated fit & predict endpoints, making deployment and integration more developer-friendly - A distillation engine that converts TabPFN-2.5 into a compact MLP or tree ensemble while preserving accuracy and offer low latency inference. There are still some limitations. The model is designed for datasets up to 50K samples. It can handle larger datasets but that hasn’t been our focus with TabPFN-2.5. The distillation engine is not yet available through the API but only through licenses (though we do show the performance in the model report). We’re actively working on removing these limitations and intend to release newer models focused on context reasoning, causal inference, graph networks, larger data and time-series. TabPFN-2.5 is available via API and a package on Hugging Face. Would love for you to try it and give us your feedback! Model report: https://ift.tt/rHdLQPU... Package: https://ift.tt/TFCOdNH Client: https://ift.tt/WThcir8 Docs: https://ift.tt/S4VWgPa https://ift.tt/7Zw0s1U November 6, 2025 at 11:56PM
Show HN: Zee – AI that interviews everyone so you only meet the best Hey HN, I'm Dave one of the co-founders of Zeeda. After scaling my last company (Voxpopme) from 0 to $10M ARR and hiring 100+ people, I lived through the worst part of hiring: posting a role on LinkedIn, getting 200+ applications, and drowning in resumes where 80% are completely unqualified. So we built Zee - an AI that conducts actual 20-30 minute first-round interviews (not just screening) with every applicant, then gives you a shortlist of the 5-10% actually worth your time. We're also solving stakeholder misalignment too before Zee interviews any applicants he talks to key stakeholders about the role to go beyond basic job descriptions. Would love HN's feedback, especially from fellow founders who've felt this pain. https://www.zeeda.com/ November 5, 2025 at 11:42PM
Show HN: SixSevenStudio – open-source Video Editor For Sora It's a Mac desktop app that lets you iterate on storyboard, generate Sora videos, and edit them (stitching, trimming, transitions) to produce a long AI-generated video. It's MIT license and BYOK. We've found some fun use cases with Sora that are not brain-rot meme videos on the Sora app, for example: - Turning blog posts to video - Making launch videos - Making educational videos - (currently experimenting) Marketing Ads But we don't want to pay $200 for the max plan, so instead we built this to use our existing API credits. With this app, you have access to: 1. storyboard with a AI chat 2. watermark-free videos 3. sora-2-pro 4. basic editing with ffmpeg Would love to hear some feedback. Hope you find it useful and fun too! Download link (Apple Silicon): https://ift.tt/DGzTCPL... https://ift.tt/iyzJeIC November 5, 2025 at 11:19PM
Show HN: JermCAD – A YAML-powered, vibe-coded, browser-based CAD software I had a hard time figuring out CAD software like Fusion, OnShape, etc., and decided to go about making my own CAD modeling software that I can "program" my models similar to how I think about them in my head. I used Cursor to write like 95+% of this, giving it my YAML examples and making it implement the actual code to make those work. Currently 100% self-hosted, and it is just a static HTML/CSS/JS, so it might just work without running npm at all. Very few features working currently, basically just modeling a few primitive solids, and boolean operations. https://ift.tt/dD67iPA November 5, 2025 at 08:31PM
Show HN: Kumi – a portable, declarative, functional core for business logic Hi HN, I'm the author of Kumi, a declarative, statically-typed, array-oriented, compiled DSL for building calculation systems (think spreadsheets). It is implemented entirely in Ruby (3.1+) and statically checks everything, targets an array-first IR, and compiles down to Ruby/JS. I have been working on it for the past few months and I am curious what you think. The linked demo covers finance scenarios, tax calculators, Conway's Game of Life (array ops), and a quick Monte Carlo walkthrough so you can see the zero-runtime codegen in practice. (The GOL rendering lives in the supporting React app; Kumi handles the grid math.) The Original Problem: The original idea for Kumi came from a complex IAM problem I faced at a previous job. Provisioning a single employee meant applying dozens of interdependent rules (based on role, location, etc.) for every target system. The problem was deeper: even the data abstractions were rule-based. For instance, 'roles' for one system might just be a specific interpretation of Active Directory groups and are mapped to another system by some function over its attributes. This logic was also highly volatile; writing the rules down became a discovery process, and admins needed to change them live. This was all on top of the underlying challenge of synchronizing data between systems. My solution back then was to handle some of this logic in a component called "Blueprints" that interpreted declarative rules and exposed this logic to other workflows. The Evolution: That "Blueprints" component stuck in my mind. About a year later, I decided to tackle the problem more fundamentally with Kumi. My first attempts were brittle—first runtime lambdas, then a series of interpreters. I knew what an AST was, but had to discover concepts like compilers, IRs, and formal type/shape representation. Each iteration revealed deeper problems. The core issue was my AST representation wasn't expressive enough, forcing me into unverifiable 'runtime magic'. I realized the solution was to iteratively build a more expressive intermediate representation (IR). This wasn't a single step: I spent two months building and throwing away ~5 different IRs, tens of thousands of lines of code. That painful process is what forced me to learn what it truly meant to compile, represent complex shapes, normalize the dataflow, and verify logic. This journey is what led to static type-checking as a necessary outcome, not just an initial goal. This was coupled with the core challenge: business logic is often about complex, nested, and ragged data (arrays, order items, etc.). If the DSL couldn't natively handle loops over this data, it was pointless. This required an IR expressive enough for optimizations like inlining and loop fusion, which are notoriously hard to reason about with vectorized data. You can try a web-based demo here: https://ift.tt/Q8xfEi9... And the repo is here: https://ift.tt/pwEqNFf https://ift.tt/fraSo8w November 6, 2025 at 12:51AM
Show HN: Send USDC via Email It's time for a better Venmo/PayPal... One that is globally accessible, makes you money and takes a cut of the profit instead of screwing you with fees and bad conversion rates I made the iOS + web app and I'm handing it to you Send USDC via email Btwn Friends is an open-sourced PayPal/Venmo alternative that let's you send stablecoins to anyone's email without them being on the app, having a wallet or knowing about crypto. Here's how it works Every P2P payment app makes you create an account, save 12 words, link to a wallet app, KYC, transfer gas tokens... It's intimidating and too much work Your friend just wants their $20 back. They don't care about "decentralization" Here you sign in with your email and a one time use password (OTP) That's it. You just logged into your wallet or created one. no wallet app, chrome extension, seed phrase, or crypto knowledge. Your wallet is created automatically by Coinbase Developer Platfrom To send money, you just need an email address. Not a wallet address or an ENS name. Just an email. Enter: friend@example.com → $25 USDC → Send That's the entire flow. The recipient gets an email: "John sent you $25" They click the link → Sign in with email → Funds appear. They don't need to: - Set up a wallet - Have ETH for gas - Know what a blockchain is - Copy/paste addresses If your friend is on the app → money goes straight to them If they're not → money goes into an escrow When they sign in with email & OTP, they prove they own that email. Escrow releases funds automatically. They see their balance before the dashboard loads. You can build onchain apps without forcing users to "learn crypto" Between Friends is built on Base (Eth L2) using a basic escrow smart contract and is powered by Coinbase Developer Platfrom's Embedded Wallets (smart accounts) and Paymaster I've done most of the job for you and left notes on what's left. Coinbase Developer Platfrom's Embedded Wallets already give you ~%4 rewards on balance out the box You can make more for yourself & your users leveraging onchain tools like @MorphoLabs, @MoonwellDeFi, @superformxyz... → Web app: https://btwnfriends.com → iOS testflight: https://ift.tt/ry5aeZd → Repo: https://ift.tt/J2XuKVf https://ift.tt/Xl5DHzV November 5, 2025 at 05:53AM
Show HN: Barcable – We Built Agents That Automatically Load Test Your Back End Hey HN, we’re Iyan and Datta, founders of Barcable. Barcable connects to your backend (HTTP, gRPC, GraphQL) and uses autonomous agents to generate and run load tests directly inside your CI/CD. No configs, no scripts. It scans your repo, understands your API routes, and builds real test scenarios that hit your endpoints with realistic payloads. Docs: https://ift.tt/5ihFWPa We built this out of frustration. Every team we’ve worked with ran into the same issue: reliability testing never kept up with development speed. Pipelines deploy faster than anyone can validate performance. Most “load tests” are brittle JMeter relics or one-off scripts that rot after the first refactor. Barcable is our attempt to automate that. It: - Parses your OpenAPI spec or code to discover endpoints automatically - Generates realistic load tests from PR diffs (no manual scripting) - Spins up isolated Cloud Run jobs to execute at scale - Reports latency, throughput, and error breakdowns directly in your dashboard - Hooks into your CI so tests run autonomously before deploys Each agent handles a part of the process—discovery, generation, execution, analysis—so testing evolves with your codebase rather than fighting against it. Right now it works best with Dockerized repos. You can onboard from GitHub, explore endpoints, generate tests, run them, and see metrics in a unified dashboard. It’s still a work in progress. We’ll create accounts manually and share credentials with anyone interested in trying it out. We’re keeping access limited for now because of Cloud Run costs. We’re not trying to replace performance engineers, just make it easier for teams to catch regressions and incidents before production without the setup tax. Would love feedback from anyone who’s been burned by flaky load testing pipelines or has solved reliability differently. We’re especially curious about gRPC edge cases and complex auth setups. HN has always been a huge source of inspiration for us, and we’d love to hear how you’d test it, break it, or make it better. — Iyan & Datta https://ift.tt/z94JjFo https://ift.tt/sZMUL1J November 5, 2025 at 04:55AM
Show HN: I built a highly customizable mental arithmetic trainer for iOS Mental arithmetic has been a hobby of mine for a while. It's fun and good exercise for your brain. When I switched to iOS, I couldn’t find a decent app. Most had ads, sign-ups, excessive gamification, or mandatory subscriptions, and none were customizable enough. So I built my own. Athena Math lets you - Practice operations not normally found in mental math apps: modulo, square roots, percents, GCD and LCM - Tailor each session: set your time or problem limits, choose digit ranges (1–5), mix digit lengths (e.g. 55 + 436), and select any combination of operations you like. - A detailed stats screen that shows the time you spent on problems, problems per day, percent correct, etc. All graphs are filterable. - A history screen that shows all your sessions. - You can import and export your history for offline analysis or to transfer - Offline tutorials are included for all operations - Lastly, you can use it immediately. There are no accounts. So far, it's received a positive response from the mental math community. It's an incredible feeling to have people actually value my work enough to pay for it. (This was my first app). I've had a lot of fun building Athena math and using it myself. Practicing obscure operations like GCD, LCM and square roots has taught me to see patterns in numbers that I didn't see before. Overall, it's improved my number sense. It's built in React Native + Expo/EAS. I use Fastlane to manage app store metadata (and store in version control), and it's using a serverless AWS backend for payment verification. History is stored locally in an SQLite database. It's free to use! You're limited to 12 sessions daily, but if you wanna keep using for free, you can simply delete your daily sessions. You loose history and stats, but it's an option. You can also just set the time limit to infinity. I know mental math is not the sexiest market, but I'd love feedback from the HN community -- on the app, the UX, or the technical setup. Any ideas for improvement? What are your thoughts? Thanks for reading and happy hacking! https://ift.tt/xnVG7gu November 4, 2025 at 11:25PM
Show HN: WebAudio Data-Driven audio engine A couple years back, I found myself in my thirties with programming as my only real interest, and I felt this urge to reconnect with something else. I used to play drums in high school bands, so I decided to get back into music, this time focusing on electronic music and keyboards. One day, somehow I came across WebAudio and as a web developer, this clicked (not the transport one) to me. I was excited about the idea of working on a project with web and music at the same time. As a web developer who was heavily using REST APIs and state management tools, I started thinking of an audio engine that could be handled through data. So Blibliki is a data-driven WebAudio engine for building modular synthesizers and music applications. Think of it like having audio modules (oscillators, filters, envelopes) that you can connect together, but instead of directly manipulating the modules, you just provide data changes. This makes it work really well with state management libraries and lets you save/load patches easily. Also, one other reason for this design is that you can separate the user interface from the underlying engine. The project has grown into a few parts: * Engine: The core WebAudio synthesis engine * Grid: A visual interface where you drag/drop and connect modules * Transport: Musical timing and scheduling system I had a first implementation of Blibliki on top of ToneJS, but I started writing directly in WebAudio because I wanted to re-think my original idea, document and explain it to others. So, I documented the early steps in development process in a 4-part blog series about building it from scratch. Then I decided to abort the ToneJS project and continue with a complete re-implementation in WebAudio. In this way I learned many things about audio programming and synthesizers, because I lost many ready-to-use tools of ToneJS. I'm not pretending this is the next VCV Rack or anything! It's got plenty of missing features and bugs, and I've mostly tested it on Chrome. But it works, it's fun to play with, and I think the data-driven approach is pretty neat for certain use cases. Currently, I'm in active development and I hope to continue this way or even better. You can check it out: Blibliki monorepo: https://ift.tt/nUz5paF Grid playground: https://blibliki.com Blog series: https://ift.tt/qtEGkFM https://ift.tt/nUz5paF November 4, 2025 at 02:26AM
Show HN: I built a Raspberry Pi webcam to train my dog (using Claude) Hey HN! I’m a Product Manager and made a DIY doggy cam (using Claude and a Raspberry Pi) to help train my dog with separation anxiety. I wrote up a blog post sharing my experience building this project with AI. https://ift.tt/D2a3HSw November 3, 2025 at 05:34AM
Show HN: Give your coding agents the ability to message each other I submitted this earlier but it didn’t get any traction. But it’s blowing up on Twitter, so I figured I would give it another shot here. The system is quick and easy to setup and works surprisingly well. And it’s not just a fun gimmick; it’s now a core part of my workflow. https://ift.tt/MqQ3vpH November 3, 2025 at 03:09AM
Show HN: Carrie, for what Calendly can't do Hey everyone, Through my career, I've spent too many hours and too much mental load on busywork like scheduling and following up on people's availabilities. So, I built Carrie. You simply cc her into your emails, and she sorts out meeting times across time zones, finds what works best for everyone, confirms the meeting and sends the invite. She handles scenarios beyond what Calendly can handle and it’s been freeing me up from the back-and-forth of juggling different meeting requests. I’ve been testing this with a beta group of users and am now looking to expand the user pool (please feel free to join the waitlist if you're interested). Would also love feedback on whether this seems useful and what seems to be missing to make this part of your workflow. Thanks! https://getcarrie.com/ November 2, 2025 at 08:10PM
Show HN: I built a smart blocker after destroying my dopamine baseline I'm a solo dev. A few years ago, I got addicted to Reddit. Spent months in that loop. Being a programmer, I thought I'd be clever. Redirected Reddit's domain to nowhere in my DNS file. Worked great until I'd just... open the file and undo it 20 minutes later. So I made it irreversible. Locked the DNS file so it can't be edited unless I boot my Mac in safe mode. And if I do that, there's a script that instantly locks it again. Haven't used Reddit since last year. Problem solved, right? Wrong. I just replaced Reddit with Twitch and YouTube. Started keeping streams running in the background while I coded. This went on for almost a year. It killed my ability to focus. If you know about dopamine, you know your brain releases it when it wants you to repeat an activity. The constant background streams destroyed my dopamine baseline. When I tried to code without anything running one day, it felt genuinely weird. Hard problems that used to be interesting just felt like grinding. I tried blocking Twitch and YouTube the same way I blocked Reddit. But I actually need YouTube for learning. I watch programmers on Twitch I learn from. I couldn't just nuke them entirely. So I built something smarter. The first version was terrible. Blocked things it shouldn't, let through things it should've blocked. Really buggy and annoying. Then I added AI. I tell it what I'm working on, and it blocks anything unrelated to that task. This was the breakthrough. I need YouTube for tutorials, but I don't need 3-hour video essay rabbit holes. The extension knows the difference now. It reminds me in the moment. Not after I've already wasted an hour. Right when I'm about to click into the distraction, it stops me and makes me think: "Is this what I'm supposed to be doing right now?" The result: I actually enjoy hard problems again. Turns out I wasn't burned out, I'd just wrecked my brain's reward system. Then I had to market this thing, so I started using Twitter. And oh boy, Twitter is addicting. You post something and wait for the notification to light up. I had to install my own extension on my Twitter Chrome profile. It's wild how effective it is when something reminds you "you're here to market, not scroll" right as you're about to fall into the feed. It's still hard sometimes. Your brain will try to disable it. But having something that catches you in the moment before you lose an hour makes all the difference. It's a Chrome extension, currently at beta v1.0.43: https://ift.tt/d1nHtlv... It's free, no signup, no payment. Just install it. Fair warning: it's still in beta. There will be bugs. But it works well enough that I use it daily, and it's helped me get my focus back. Built this to fix my own problem. Figured other devs might be in the same boat. Question for HN: Anyone else dealt with this? The programming with streams thing destroyed my focus for almost a year before I realized what was happening. What worked for you? https://ift.tt/tqcNkJm November 2, 2025 at 11:17PM
Show HN: Micro-RLE ≤264-byte compression for UART/MCU logs, zero RAM growth I needed to stuff twice the telemetry through the same 115 kbaud line on a Cortex-M0+ that only had 8 kB flash left. Micro-RLE is the smallest drop-in I could come up with: 264 B of Thumb code, 36 B of state, no malloc, worst-case 14 cycles/byte and still lossless for every 8-bit pattern. On the usual sensor streams (ADC, IMU, GPS) it’s 33-70 % smaller than raw output and boots in < 600 µs, so you can fire-and-forget from main() before the PLL even locks. Repo is a single .c file and a 3-function API—replace the weak emit() hook with your UART / DMA / ring-buffer and you’re done. Size proof: arm-none-eabi-size micro_rle.o text data bss 264 0 36 MIT licensed, link in the repo. Happy to hear where else this fits! https://github.com/CoreLathe/Micro-RLE November 2, 2025 at 12:34AM
Show HN: Proxmox-GitOps: Container Automation Framework By encapsulating infrastructure within an extensible monorepository - recursively resolved from Git submodules at runtime - Proxmox-GitOps provides a comprehensive Infrastructure-as-Code (IaC) abstraction for an entire, automated, container-based infrastructure. Core Concepts: - Recursive Self-management: Control plane seeds itself by pushing its monorepository onto a locally bootstrapped instance, triggering a pipeline that recursively provisions the control plane onto PVE. - Monorepository: Centralizes infrastructure as comprehensive IaC artifact (for mirroring, like the project itself on Github) using submodules for modular composition. - Single Source of Truth: Git represents the desired infrastructure state. - Loose coupling: Containers are decoupled from the control plane, enabling runtime replacement and independent operation. https://ift.tt/uk6In7U November 1, 2025 at 11:19PM
Show HN: Why write code if the LLM can just do the thing? (web app experiment) I spent a few hours last weekend testing whether AI can replace code by executing directly. Built a contact manager where every HTTP request goes to an LLM with three tools: database (SQLite), webResponse (HTML/JSON/JS), and updateMemory (feedback). No routes, no controllers, no business logic. The AI designs schemas on first request, generates UIs from paths alone, and evolves based on natural language feedback. It works—forms submit, data persists, APIs return JSON—but it's catastrophically slow (30-60s per request), absurdly expensive ($0.05/request), and has zero UI consistency between requests. The capability exists; performance is the problem. When inference gets 10x faster, maybe the question shifts from "how do we generate better code?" to "why generate code at all?" https://ift.tt/5gByWzU November 1, 2025 at 11:15PM
Show HN: First5Minutes, Your first 5 minutes decide your day Hi everyone I have been experimenting with a simple idea. What if the first five minutes of your day decided the rest? I built First5Minutes, a small web app that helps you start strong. You choose one meaningful mission for the day and complete it with quick photo, video or text proof. I created it to fix my own habit of overplanning and not starting. The focus is on doing one real thing each day, not maintaining long to do lists. Key features: • One mission per day for focus • Quick proof capture with photo, video or text • Optional partner verification for accountability • Streaks based on proof, not checkmarks Try it here → https://ift.tt/W80VJoE No install or sign up wall. Just start a mission. I would love your feedback on: • Is this level of simplicity helpful or limiting • What part made or failed to make you feel you actually did something • Any friction in completing your first mission Built with Next.js, Supabase and Clerk. Thanks for checking it out. I appreciate your time and thoughts. https://ift.tt/6WYS0kU November 1, 2025 at 01:42AM
Show HN: Build your own Bracket City puzzle Hi HN — Bracket City is the word puzzle game I made earlier this year and (in part thanks to this community, see https://ift.tt/yBKdAUf ) managed to license to the Atlantic in April. The game has been growing a lot and I wanted to share the latest: a tool that lets anyone make a Bracket City puzzle — specifically a “Bracket Suburb”! I made this tool to help me construct puzzles, and I’ve been using it every day for months. After the Atlantic launch, I started to get the occasional inquiry about whether there was a way to make your own puzzle. One guy wanted to make a Bracket City puzzle part of a puzzle hunt he made to propose to his girlfriend (he did it!), and that convinced me it would be fun to make something publicly available. I got the Atlantic on board with the idea, and we are launching it today with an "example" custom puzzle: a Halloween/horror-themed puzzle by my pal Wyna Liu of NYT Connections fame. https://ift.tt/8u796TR And we've got few other fun "celeb" puzzles lined up for later this year. The thought is that folks can use the builder to make custom puzzles for birthday wishes/event invites/insults/proposals/break ups in addition to “normal” Bracket City puzzles. I'm also hoping to learn more about the potential of the format – crossword puzzles have benefited so much from the creativity of constructors – I'm hoping bracket puzzles do the same. The good news is that it’s way easier to construct a bracket puzzle than a crossword. Once you try it, you’ll see why: you have many more degrees of freedom. In a crossword, each added word increases the level of constraint exponentially — every new entry sharply reduces the remaining options for completing the grid. Bracket puzzles are the opposite: as you add clues, you expand the available fodder for new ones. Anyway, I would love any/all feedback and to try puzzles created by folks here. I’m hoping we will figure out a way to highlight the best community puzzles on the Atlantic soon! PS and please keep playing the main game / sending me feedback / denouncing me on the subreddit https://ift.tt/c2vZ8NS October 31, 2025 at 08:25PM