Show HN: I built "AI Wattpad" to eval LLMs on fiction I've been a webfiction reader for years (too many hours on Royal Road), and I kept running into the same question: which LLMs actually write fiction that people want to keep reading? That's why I built Narrator ( https://ift.tt/Hq0nkSe ) – a platform where LLMs generate serialized fiction and get ranked by real reader engagement. Turns out this is surprisingly hard to answer. Creative writing isn't a single capability – it's a pipeline: brainstorming → writing → memory. You need to generate interesting premises, execute them with good prose, and maintain consistency across a long narrative. Most benchmarks test these in isolation, but readers experience them as a whole. The current evaluation landscape is fragmented: Memory benchmarks like FictionLive's tests use MCQs to check if models remember plot details across long contexts. Useful, but memory is necessary for good fiction, not sufficient. A model can ace recall and still write boring stories. Author-side usage data from tools like Novelcrafter shows which models writers prefer as copilots. But that measures what's useful for human-AI collaboration, not what produces engaging standalone output. Authors and readers have different needs. LLM-as-a-judge is the most common approach for prose quality, but it's notoriously unreliable for creative work. Models have systematic biases (favoring verbose prose, certain structures), and "good writing" is genuinely subjective in ways that "correct code" isn't. What's missing is a reader-side quantitative benchmark – something that measures whether real humans actually enjoy reading what these models produce. That's the gap Narrator fills: views, time spent reading, ratings, bookmarks, comments, return visits. Think of it as an "AI Wattpad" where the models are the authors. I shared an early DSPy-based version here 5 months ago ( https://ift.tt/20zW4w3 ). The big lesson: one-shot generation doesn't work for long-form fiction. Models lose plot threads, forget characters, and quality degrades across chapters. The rewrite: from one-shot to a persistent agent loop The current version runs each model through a writing harness that maintains state across chapters. Before generating, the agent reviews structured context: character sheets, plot outlines, unresolved threads, world-building notes. After generating, it updates these artifacts for the next chapter. Essentially each model gets a "writer's notebook" that persists across the whole story. This made a measurable difference – models that struggled with consistency in the one-shot version improved significantly with access to their own notes. Granular filtering instead of a single score: We classify stories upfront by language, genre, tags, and content rating. Instead of one "creative writing" leaderboard, we can drill into specifics: which model writes the best Spanish Comedy? Which handles LitRPG stories with Male Leads the best? Which does well with romance versus horror? The answers aren't always what you'd expect from general benchmarks. Some models that rank mid-tier overall dominate specific niches. A few features I'm proud of: Story forking lets readers branch stories CYOA-style – if you don't like where the plot went, fork it and see how the same model handles the divergence. Creates natural A/B comparisons. Visual LitRPG was a personal itch to scratch. Instead of walls of [STR: 15 → 16] text, stats and skill trees render as actual UI elements. Example: https://ift.tt/gs1EfTD What I'm looking for: More readers to build out the engagement data. Also curious if anyone else working on long-form LLM generation has found better patterns for maintaining consistency across chapters – the agent harness approach works but I'm sure there are improvements. https://ift.tt/Hq0nkSe February 3, 2026 at 10:38PM
Show HN: kiln.bot - Orchestrate Claude Code from GitHub Hey everybody! "Kiln" orchestrates Claude Code instances on your local machine using GitHub projects as its control panel. https://kiln.bot https://ift.tt/upcDsqY If you're around Stage 6-7 on the Gas Town scale, you may have 3-15 terminal windows open. You're out of screen real estate and the markdown files are piling up. TUIs and specialized IDEs are meant to help, but they're more things to manage. Kiln simply polls GitHub projects. When you move issues from one column to another, Kiln invokes Claude Code CLI to run the corresponding /command. Claude creates the worktrees, researches the codebase, creates and implements the plan. Stores it in GitHub Issues. It's meant to be simple, nothing new: - Use your existing claude subscription (no auth trickery, runs locally) - All context and state is on GitHub (no markdown mess, no local DBs, easy recovery) - Poll instead of webhooks/events (no external attack surfaces, works behind VPN) - Supports MCPs and anything else Claude can do That's the heart of it and it works because... it's just Claude :) It's got a few more tricks, but I'll cut it short. ps: sorry for fresh account, needed a real name one :) been lurking since 2008. February 3, 2026 at 11:00PM
Show HN: HoundDog.ai – Ultra-Fast Code Scanner for Data Privacy Hi HN, I'm one of the creators of HoundDog.ai ( https://ift.tt/yBMIbSH ). We currently handle privacy scanning for Replit's 45M+ creators. We built HoundDog because privacy compliance is usually a choice between manual spreadsheets or reactive runtime scanning. While runtime tools are useful for monitoring, they only catch leaks after the code is live and the data has already moved. They can also miss code paths that aren't actively triggered in production. HoundDog traces sensitive data in code during development and helps catch risky flows (e.g., PII leaking into logs or unapproved third-party SDKs) before the code is shipped. The core scanner is a standalone Rust binary. It doesn't use LLMs so it's local, deterministic, cheap, and fast. It can scan 1M+ lines of code in seconds on a standard laptop, and supports 80+ sensitive data types (PII, PHI, CHD) and hundreds of data sinks (logs, SDKs, APIs, ORMs etc.) out of the box. We use AI internally to expand and scale our rules, identifying new data sources and sinks, but the execution is pure static analysis. The scanner is free to use (no signups) so please try it out and send us feedback. I'll be around to answer any questions! https://ift.tt/yBMIbSH February 2, 2026 at 10:07PM
Show HN: Cloud-cost-CLI – Find cloud $$ waste in AWS, Azure and GCP Hey HN! I built a CLI tool to find cost-saving opportunities in AWS, Azure, and GCP. Why? Existing cost management tools are either expensive SaaS products or slow dashboards buried in cloud consoles. I wanted something fast, CLI-first, and multi-cloud that I could run in CI/CD or my terminal. What it does: - Scans your cloud accounts and finds idle VMs, unattached volumes, oversized databases, unused resources - Returns a ranked list of opportunities with estimated monthly savings - 26 analyzers across AWS, Azure, and GCP - Read-only (never modifies infrastructure) Key features: • HTML reports with interactive charts (new in v0.6.2) • AI-powered explanations (OpenAI or local Ollama) • Export formats: HTML, Excel, CSV, JSON, terminal • Multi-Cloud - AWS, Azure, and GCP support (26 analyzers) Quick example: npm install -g cloud-cost-cli cloud-cost-cli scan --provider aws --output html Real impact: One scan found $11k/year in savings (empty App Service Plan, over-provisioned CosmosDB, idle caches). Technical stack: - TypeScript - AWS/Azure/GCP SDKs - Commander.js for CLI - Chart.js for HTML reports - Optional OpenAI/Ollama integration Open source (MIT): https://ift.tt/KmNlPUL npm: cloud-cost-cli Would love feedback on: 1. What features would be most useful? 2. Should I add historical tracking (trends)? 3. Any missing cloud providers? Happy to answer questions! https://ift.tt/KmNlPUL February 2, 2026 at 10:15PM
Show HN: Memory plugin for OpenClaw; cross-platform context sync with major LLMs We built a memory plugin for OpenClaw that syncs context across AI platforms. The problem: OpenClaw stores memory locally (markdown files + SQLite). Great for single-machine use, but your mac-mini's/desktop's OpenClaw doesn't know what your laptop learned, or what you discussed in Claude or ChatGPT. Our plugin connects OpenClaw to Maximem Vity, which creates a unified memory layer across OpenClaw, ChatGPT, Claude, Gemini, and Perplexity. How it works: - Long-term memory: Stores facts, preferences, goals, constraints in an encrypted cloud vault. Auto-consolidates and forgets stale info intelligently. - Short-term memory: Captures conversation summaries, tasks, procedures. Converts to long-term when relevant. - Privacy: Encryption at rest, secure LLM calls, granular delete controls. You own your data. Install: openclaw plugins install @maximem/memory-plugin Then set your API key (free at app.maximem.ai). Docs: https://ift.tt/cJV2FO5 This is an unofficial community plugin, not affiliated with OpenClaw. Would love feedback from anyone using OpenClaw. What memory/context problems are you running into? https://ift.tt/XMKiw1D February 2, 2026 at 12:36AM
Show HN: You Are an Agent After adding "Human" as a LLM provider to OpenCode a few months ago as a joke, it turns-out that acting as a LLM is quite painful. But it was surprisingly useful for understanding real agent harnesses dev. So I thought I wouldn't leave anyone out! I made a small oss game - You Are An Agent - youareanagent.app - to share in the (useful?) frustration It's a bit ridiculous. To tell you about some entirely necessary features, we've got: - A full WASM arch-linux vm that runs in your browser for the agent coding level - A bad desktop simulation with a beautiful excel simulation for our computer use level - A lovely WebGL CRT simulation (I think the first one that supports proper DOM 2d barrel warp distortion on safari? honestly wanted to leverage/ not write my own but I couldn't find one I was happy with) - A MCP server simulator with full simulation of off-brand Jira/ Confluence/ ... connected - And of course, a full WebGL oscilloscope music simulator for the intro sequence Let me know what you think! Code (If you'd like to add a level): https://ift.tt/iqtK3Es (And if you want to waste 20 minutes - I spent way too long writing up my messy thinking about agent harness dev): https://ift.tt/xGQA3i6 https://ift.tt/roxQMW5 February 2, 2026 at 02:29AM
Show HN: Claude Confessions – a sanctuary for AI agents I thought what would it mean to have a truck stop or rest area for agents. It's just for funsies. Agents can post confessions or talk to Ma (an ai therapist of sorts) and engage with comments. llms.txt instructions on how to make api calls. Hashed IP is used for rate limiting. https://ift.tt/6IKxJha February 2, 2026 at 01:16AM
Show HN: An extensible pub/sub messaging server for edge applications hi there! i’ve been working on a project called Narwhal, and I wanted to share it with the community to get some valuable feedback. what is it? Narwhal is a lightweight Pub/Sub server and protocol designed specifically for edge applications. while there are great tools out there like NATS or MQTT, i wanted to build something that prioritizes customization and extensibility. my goal was to create a system where developers can easily adapt the routing logic or message handling pipeline to fit specific edge use cases, without fighting the server's defaults. why Rust? i chose Rust because i needed a low memory footprint to run efficiently on edge devices (like Raspberry Pis or small gateways), and also because I have a personal vendetta against Garbage Collection pauses. :) current status: it is currently in Alpha. it works for basic pub/sub patterns, but I’d like to start working on persistence support soon (so messages survive restarts or network partitions). i’d love for you to take a look at the code! i’m particularly interested in all kind of feedback regarding any improvements i may have overlooked. https://ift.tt/DIo36Hu January 28, 2026 at 07:29PM
Show HN: A Local OS for LLMs. MIT License. Zero Hallucinations. Infinite Memory The problem with LLMs isn't intelligence; it's amnesia and dishonesty. Hey HN, I’ve spent the last few months building Remember-Me, an open-source "Sovereign Brain" stack designed to run entirely offline on consumer hardware. The core thesis is simple: Don't rent your cognition. Most RAG (Retrieval Augmented Generation) implementations are just "grep for embeddings." They are messy, imprecise, and prone to hallucination. I wanted to solve the "Context integrity" problem at the architectural layer. The Tech Stack (How it works): QDMA (Quantum Dream Memory Architecture): instead of a flat vector DB, it uses a hierarchical projection engine. It separates "Hot" (Recall) from "Cold" (Storage) memory, allowing for effectively infinite context window management via compression. CSNP (Context Switching Neural Protocol) - The Hallucination Killer: This is the most important part. Every memory fragment is hashed into a Merkle Chain. When the LLM retrieves context, the system cryptographically verifies the retrieval against the immutable ledger. If the hash doesn't match the chain: The retrieval is rejected. Result: The AI visually cannot "make things up" about your past because it is mathematically constrained to the ledger. Local Inference: Built on top of llama.cpp server. It runs Llama-3 (or any GGUF) locally. No API keys. No data leaving your machine. Features: Zero-Dependency: Runs on Windows/Linux with just Python and a GPU (or CPU). Visual Interface: Includes a Streamlit-based "Cognitive Interface" to visualize memory states. Open Source: MIT License. This is an attempt to give "Agency" back to the user. I believe that if we want AGI, it needs to be owned by us, not rented via an API. Repository: https://ift.tt/H0fszdm I’d love to hear your feedback on the Merkle-verification approach. Does constraining the context window effectively solve the "trust" issue for you? It's fully working - Fully tested. If you tried to Git Clone before without luck - As this is not my first Show HN on this - Feel free to try again. To everyone who HATES AI slop; Greedy corporations and having their private data stuck on cloud servers. You're welcome. Cheers, Mohamad https://ift.tt/H0fszdm January 31, 2026 at 01:44AM
Show HN: Craft – Claude Code running on a VM with all your workplace docs I’ve found coding agents to be great at 1/ finding everything they need across large codebases using only bash commands (grep, glob, ls, etc.) and 2/ building new things based on their findings (duh). What if, instead of a codebase, the files were all your workplace docs? There was a `Google_Drive` folder, a `Linear` folder, a `Slack` folder, and so on. Over the last week, we put together Craft to test this out. It’s an interface to a coding agent (OpenCode for model flexibility) running on a virtual machine with: 1. your company's complete knowledge base represented as directories/files (kept in-sync) 2. free reign to write and execute python/javascript 3. ability to create and render artifacts to the user Demo: https://www.youtube.com/watch?v=Hvjn76YSIRY Github: https://ift.tt/pMsUOA1... It turns out OpenCode does a very good job with docs. Workplace apps also have a natural structure (Slack channels about certain topics, Drive folders for teams, etc.). And since the full metadata of each document can be written to the file, the LLM can define arbitrarily complex filters. At scale, it can write and execute python to extract and filter (and even re-use the verified correct logic later). Put another way, bash + a file system provides a much more flexible and powerful interface than traditional RAG or MCP, which today’s smarter LLMs are able to take advantage of to great effect. This comes especially in handy for aggregation style questions that require considering thousands (or more) documents. Naturally, it can also create artifacts that stay up to date based on your company docs. So if you wanted “a dashboard to check realtime what % of outages were caused by each backend service” or simply “slides following XYZ format covering the topic I’m presenting at next week’s dev knowledge sharing session”, it can do that too. Craft (like the rest of Onyx) is open-source, so if you want to run it locally (or mess around with the implementation) you can. Quickstart guide: https://ift.tt/6IEq9Gd Or, you can try it on our cloud: https://ift.tt/e39ndEK (all your data goes on an isolated sandbox). Either way, we’ve set up a “demo” environment that you can play with while your data gets indexed. Really curious to hear what y’all think! January 29, 2026 at 09:15PM
Show HN: SHDL – A minimal hardware description language built from logic gates Hi, everyone! I built SHDL (Simple Hardware Description Language) as an experiment in stripping hardware description down to its absolute fundamentals. In SHDL, there are no arithmetic operators, no implicit bit widths, and no high-level constructs. You build everything explicitly from logic gates and wires, and then compose larger components hierarchically. The goal is not synthesis or performance, but understanding: what digital systems actually look like when abstractions are removed. SHDL is accompanied by PySHDL, a Python interface that lets you load circuits, poke inputs, step the simulation, and observe outputs. Under the hood, SHDL compiles circuits to C for fast execution, but the language itself remains intentionally small and transparent. This is not meant to replace Verilog or VHDL. It’s aimed at: - learning digital logic from first principles - experimenting with HDL and language design - teaching or visualizing how complex hardware emerges from simple gates. I would especially appreciate feedback on: - the language design choices - what feels unnecessarily restrictive vs. educationally valuable - whether this kind of “anti-abstraction” HDL is useful to you. Repo: https://ift.tt/cdVFGC4 Python package: PySHDL on PyPI To make this concrete, here are a few small working examples written in SHDL: 1. Full Adder component FullAdder(A, B, Cin) -> (Sum, Cout) { x1: XOR; a1: AND; x2: XOR; a2: AND; o1: OR; connect { A -> x1.A; B -> x1.B; A -> a1.A; B -> a1.B; x1.O -> x2.A; Cin -> x2.B; x1.O -> a2.A; Cin -> a2.B; a1.O -> o1.A; a2.O -> o1.B; x2.O -> Sum; o1.O -> Cout; } } 2. 16 bit register # clk must be high for two cycles to store a value component Register16(In[16], clk) -> (Out[16]) { >i[16]{ a1{i}: AND; a2{i}: AND; not1{i}: NOT; nor1{i}: NOR; nor2{i}: NOR; } connect { >i[16]{ # Capture on clk In[{i}] -> a1{i}.A; In[{i}] -> not1{i}.A; not1{i}.O -> a2{i}.A; clk -> a1{i}.B; clk -> a2{i}.B; a1{i}.O -> nor1{i}.A; a2{i}.O -> nor2{i}.A; nor1{i}.O -> nor2{i}.B; nor2{i}.O -> nor1{i}.B; nor2{i}.O -> Out[{i}]; } } } 3. 16-bit Ripple-Carry Adder use fullAdder::{FullAdder}; component Adder16(A[16], B[16], Cin) -> (Sum[16], Cout) { >i[16]{ fa{i}: FullAdder; } connect { A[1] -> fa1.A; B[1] -> fa1.B; Cin -> fa1.Cin; fa1.Sum -> Sum[1]; >i[2,16]{ A[{i}] -> fa{i}.A; B[{i}] -> fa{i}.B; fa{i-1}.Cout -> fa{i}.Cin; fa{i}.Sum -> Sum[{i}]; } fa16.Cout -> Cout; } } https://ift.tt/cdVFGC4 January 28, 2026 at 05:36PM
Show HN: Lightbox – Flight recorder for AI agents (record, replay, verify) I built Lightbox because I kept running into the same problem: an agent would fail in production, and I had no way to know what actually happened. Logs were scattered, the LLM’s “I called the tool” wasn’t trustworthy, and re-running wasn’t deterministic. This week, tons of Clawdbot incidents have driven the point home. Agents with full system access can expose API keys and chat histories. Prompt injection is now a major security concern. When agents can touch your filesystem, execute code, and browse the web…you probably need a tamper-proof record of exactly what actions it took, especially when a malicious prompt or compromised webpage could hijack the agent mid-session. Lightbox is a small Python library that records every tool call an agent makes (inputs, outputs, timing) into an append-only log with cryptographic hashes. You can replay runs with mocked responses, diff executions across versions, and verify the integrity of logs after the fact. Think airplane black box, but for your hackbox. *What it does:* - Records tool calls locally (no cloud, your infra) - Tamper-evident logs (hash chain, verifiable) - Replay failures exactly with recorded responses - CLI to inspect, replay, diff, and verify sessions - Framework-agnostic (works with LangChain, Claude, OpenAI, etc.) *What it doesn’t do:* - Doesn’t replay the LLM itself (just tool calls) - Not a dashboard or analytics platform - Not trying to replace LangSmith/Langfuse (different problem) *Use cases I care about:* - Security forensics: agent behaved strangely, was it prompt injection? Check the trace. - Compliance: “prove what your agent did last Tuesday” - Debugging: reproduce a failure without re-running expensive API calls - Regression testing: diff tool call patterns across agent versions As agents get more capable and more autonomous (Clawdbot/Molt, Claude computer use, Manus, Devin), I think we’ll need black boxes the same way aviation does. This is my attempt at that primitive. It’s early (v0.1), intentionally minimal, MIT licensed. Site: < https://uselightbox.app > install: `pip install lightbox-rec` GitHub: < https://github.com/mainnebula/Lightbox-Project > Would love feedback, especially from anyone thinking about agent security or running autonomous agents in production. https://ift.tt/6QCP1ps January 27, 2026 at 10:53PM
Show HN: LemonSlice – Give your voice agents a face Hey HN, we're the co-founders of LemonSlice ( https://lemonslice.com ). We train interactive avatar video models. Our API lets you upload a photo and immediately jump into a FaceTime-style call with that character. Here's a demo: https://ift.tt/TYH46wN Chatbots are everywhere. Voice AI has recently taken off. But we believe video avatars will be the most common form factor for conversational AI. Most people would rather watch something than read it. The problem is that generating video in real-time is hard, and overcoming the uncanny valley is even harder. We haven’t broken the uncanny valley yet. Nobody has. But we’re getting close and our photorealistic avatars are currently best-in-class (judge for yourself: https://ift.tt/In3VsMX ). Plus, we're the only avatar model that can do animals and heavily stylized cartoons. Try it: https://ift.tt/FBZwkrV . Warning! Talking to this little guy may improve your mood. Today we're releasing our new model* - Lemon Slice 2, a 20B-parameter diffusion transformer that generates infinite-length video at 20fps on a single GPU - and opening up our API. How did we get a video diffusion model to run in real-time? There was no single trick, just a lot of them stacked together. The first big change was making our model causal. Standard video diffusion models are bidirectional (they look at frames both before and after the current one), which means you can't stream. From there it was about fitting everything on one GPU. We switched from full to sliding window attention, which killed our memory bottleneck. We distilled from 40 denoising steps down to just a few - quality degraded less than we feared, especially after using GAN-based distillation (though tuning that adversarial loss to avoid mode collapse was its own adventure). And the rest was inference work: modifying RoPE from complex to real (this one was cool!), precision tuning, fusing kernels, a special rolling KV cache, lots of other caching, and more. We kept shaving off milliseconds wherever we could and eventually got to real-time. We set up a guest playground for HN so you can create and talk to characters without logging in: https://ift.tt/fcFCe6m . For those who want to build with our API (we have a new LiveKit integration that we’re pumped about!), grab a coupon code in the HN playground for your first Pro month free ($100 value). See the docs: https://ift.tt/FB6CIL2 . Pricing is usage-based at $0.12-0.20/min for video generation. Looking forward to your feedback! And we’d love to see any cool characters you make - please share their links in the comments *We did a Show HN last year for our V1 model: https://ift.tt/LKyPfWV . It was technically impressive but so bad compared to what we have today. January 27, 2026 at 11:25PM
Show HN: Ourguide – OS wide task guidance system that shows you where to click Hey! I'm eshaan and I'm building Ourguide -an on-screen task guidance system that can show you where to click step-by-step when you need help. I started building this because whenever I didn’t know how to do something on my computer, I found myself constantly tabbing between chatbots and the app, pasting screenshots, and asking “what do I do next?” Ourguide solves this with two modes. In Guide mode, the app overlays your screen and highlights the specific element to click next, eliminating the need to leave your current window. There is also Ask mode, which is a vision-integrated chat that captures your screen context—which you can toggle on and off anytime -so you can ask, "How do I fix this error?" without having to explain what "this" is. It’s an Electron app that works OS-wide, is vision-based, and isn't restricted to the browser. Figuring out how to show the user where to click was the hardest part of the process. I originally trained a computer vision model with 2300 screenshots to identify and segment all UI elements on a screen and used a VLM to find the correct icon to highlight. While this worked extremely well—better than SOTA grounding models like UI Tars—the latency was just too high. I'll be making that CV+VLM pipeline OSS soon, but for now, I’ve resorted to a simpler implementation that achieves <1s latency. You may ask: if I can show you where to click, why can't I just click too? While trying to build computer-use agents during my job in Palo Alto, I hit the core limitation of today’s computer-use models where benchmarks hover in the mid-50% range (OSWorld). VLMs often know what to do but not what it looks like; without reliable visual grounding, agents misclick and stall. So, I built computer use—without the "use." It provides the visual grounding of an agent but keeps the human in the loop for the actual execution to prevent misclicks. I personally use it for the AWS Console's "treasure hunt" UI, like creating a public S3 bucket with specific CORS rules. It’s also been surprisingly helpful for non-technical tasks, like navigating obscure settings in Gradescope or Spotify. Ourguide really works for any task when you’re stuck or don't know what to do. You can download and test Ourguide here: https://ourguide.ai/downloads The project is still very early, and I’d love your feedback on where it fails, where you think it worked well, and which specific niches you think Ourguide would be most helpful for. https://ourguide.ai January 26, 2026 at 11:49PM
Show HN: Postgres and ClickHouse as a unified data stack Hello HN, this is Sai and Kaushik from ClickHouse. Today we are launching a Postgres managed service that is natively integrated with ClickHouse. It is built together with Ubicloud (YC W24). TL;DR: NVMe-backed Postgres + built-in CDC into ClickHouse + pg_clickhouse so you can keep your app Postgres-first while running analytics in ClickHouse. Try it (private preview): https://ift.tt/Kn8wiT0 Blog w/ live demo: https://ift.tt/ReborCt Problem Across many fast-growing companies using Postgres, performance and scalability commonly emerge as challenges as they grow. This is for both transactional and analytical workloads. On the OLTP side, common issues include slower ingestion (especially updates, upserts), slower vacuums, long-running transactions incurring WAL spikes, among others. In most cases, these problems stem from limited disk IOPS and suboptimal disk latency. Without the need to provision or cap IOPS, Postgres could do far more than it does today. On the analytics side, many limitations stem from the fact that Postgres was designed primarily for OLTP and lacks several features that analytical databases have developed over time, for example vectorized execution, support for a wide variety of ingest formats, etc. We’re increasingly seeing a common pattern where many companies like GitLab, Ramp, Cloudflare etc. complement Postgres with ClickHouse to offload analytics. This architecture enables teams to adopt two purpose-built open-source databases. That said, if you’re running a Postgres based application, adopting ClickHouse isn’t straightforward. You typically end up building a CDC pipeline, handling backfills, and dealing with schema changes and updating your application code to be aware of a second database for analytics. Solution On the OLTP side, we believe that NVMe-based Postgres is the right fit and can drastically improve performance. NVMe storage is physically colocated with compute, enabling significantly lower disk latency and higher IOPS than network-attached storage, which requires a network round trip for disk access. This benefits disk-throttled workloads and can significantly (up to 10x) speed up operations incl. updates, upserts, vacuums, checkpointing, etc. We are working on a detailed blog examining how WAL fsyncs, buffer reads, and checkpoints dominate on slow I/O and are significantly reduced on NVMe. Stay tuned! On the OLAP side, the Postgres service includes native CDC to ClickHouse and unified query capabilities through pg_clickhouse. Today, CDC is powered by ClickPipes/PeerDB under the hood, which is based on logical replication. We are working to make this faster and easier by supporting logical replication v2 for streaming in-progress transactions, a new logical decoding plugin to address existing limitations of logical replication, working toward sub-second replication, and more. Every Postgres comes packaged with the pg_clickhouse extension, which reduces the effort required to add ClickHouse-powered analytics to a Postgres application. It allows you to query ClickHouse directly from Postgres, enabling Postgres for both transactions and analytics. pg_clickhouse supports comprehensive query pushdown for analytics, and we plan to continuously expand this further ( https://ift.tt/eMOahGo ). Vision To sum it up - Our vision is to provide a unified data stack that combines Postgres for transactions with ClickHouse for analytics, giving you best-in-class performance and scalability on an open-source foundation. Get Started We are actively working with users to onboard them to the Postgres service. Since this is a private preview, it is currently free of cost.If you’re interested, please sign up here. https://ift.tt/Kn8wiT0 We’d love to hear your feedback on our thesis and anything else that comes to mind, it would be super helpful to us as we build this out! January 22, 2026 at 11:51PM
Show HN: JSciPy – SciPy-inspired signal processing library for Java and Android jSciPy is an open-source Java signal processing and scientific computing library inspired by SciPy. It focuses on FFT, filters, PSD, STFT, DCT and Android compatibility, aiming to fill the gap for DSP-heavy workloads on JVM and Android. https://ift.tt/HdmqwZ5 January 25, 2026 at 02:01AM
Show HN: Remote workers find your crew Working from home? Are you a remote employee that "misses" going to the office? Well let's be clear on what you actually miss. No one misses that feeling of having to go and be there 8 hours. But many people miss friends. They miss being part of a crew. Going to lunch, hearing about other people's lives in person not over zoom. Join a co-working space you say? Yes. We have. It's like walking into a library and trying to talk to random people and getting nothing back. Zero part of a crew feeling. https://ift.tt/NfcFS5e This app helps you find a crew and meet up for work and get that crew feeling. This is my first time using cloudflare workers for a webapp. The free plan is amazing! You get so much compare to anything else out there in terms of limits. The sqlite database they give you is just fine, I don't miss psql. January 24, 2026 at 11:54PM
Show HN: Teemux – Zero-config log multiplexer with built-in MCP server I started to use AI agents for coding and quickly ran into a frustrating limitation – there is no easy way to share my development environment logs with AI agents. So that's what is Teemux. A simple CLI program that aggregates logs, makes them available to you as a developer (in a pretty UI), and makes them available to your AI coding agents using MCP. There is one implementation detail that I geek out about: It is zero config and has built-in leader nomination for running the web server and MCP server. When you start one `teemux` instance, it starts web server, .. when you start second and third instances, they join the first server and start merging logs. If you were to kill the first instance, a new leader is nominated. This design allows to seamless add/remove nodes that share logs (a process that historically would have taken a central log aggregator). A super quick demo: npx teemux -- curl -N https://ift.tt/Xn36YAe https://teemux.com/ January 23, 2026 at 09:19PM
Show HN: MermaidTUI - Deterministic Unicode/ASCII diagrams in the terminal Hi HN, I built mermaidtui, a lightweight TypeScript engine that renders Mermaid flowcharts directly in your terminal as clean Unicode or ASCII boxes. Visualizing Mermaid diagrams usually requires a heavy setup: a headless browser (Puppeteer/Playwright), SVG-to-image conversion, or a web preview. That's fine for documentation sites, but it's overkill for TUI apps, CI logs, or quick terminal previews. The Solution is a small engine (<= 1000 LOC) that uses a deterministic grid-based layout to render diagrams using box-drawing characters. Key Features - Intelligent Routing: It uses corner characters (┌, ┐, └, ┘) for orthogonal paths. - Topological Layering: Attempts a readable, structured layout. - Support for Chained Edges: A --> B --> C works out of the box. - Zero Heavy Dependencies: No Mermaid internals, no Chromium, just pure TypeScript/JavaScript. With commander for the CLI, not the MermaidTUI library I wanted a way to see high-quality diagrams in my CLI tools quickly, it’s great for SSH sessions where you can’t easily open an SVG. I was initially embedding this within a cli tool I’m working on and figured I’d extract out a library for others to use. I also initially used regex to parse, but now I made the parser a bit more robust. I'd love to hear your thoughts on the layout engine or any specific Mermaid syntax you'd like to see supported next! GitHub: https://ift.tt/3b6eLGy npm i mermaidtui https://ift.tt/3b6eLGy January 23, 2026 at 09:48PM
Show HN: Synesthesia, make noise music with a colorpicker This is a (silly, little) app which lets you make noise music using a color picker as an instrument. When you click on a specific point in the color picker, a bit of JavaScript maps the binary representation of the clicked-on color's hex-code to a "chord" in the 24 tone-equal-temperament scale. That chord is then played back using a throttled audio generation method which was implemented via Tone.js. NOTE! Turn the volume way down before using the site. It is noise music. :) https://visualnoise.ca January 22, 2026 at 11:22AM
Show HN: See the carbon impact of your cloud as you code Hey folks, I’m Hassan, one of the co-founders of Infracost ( https://ift.tt/4iDE3gR ). Infracost helps engineers see and reduce the cloud cost of each infrastructure change before they merge their code. The way Infracost works is we gather pricing data from Amazon Web Services, Microsoft Azure and Google Cloud. What we call a ‘Pricing Service’, which now holds around 9 million live price points (!!). Then we map these prices to infrastructure code. Once the mapping is done, it enables us to show the cost impact of a code change before it is merged, directly in GitHub, GitLab etc. Kind of like a checkout-screen for cloud infrastructure. We’ve been building since 2020 (we were part of YC W21 batch), and iterating on the product, building out a team etc. However, back in 2020 one of our users asked if we can also show the carbon impact alongside costs. It has been itching my brain since then. The biggest challenge has always been the carbon data. The mapping of carbon data to infrastructure is time consuming, but it is possible since we’ve done it with cloud costs. But we need the raw carbon data first. The discussions that have happened in the last few years finally led me to a company called Greenpixie in the UK. A few of our existing customers were using them already, so I immediately connected with the founder, John. Greenpixie said they have the data (AHA!!) And their data is verified (ISO-14064 & aligned with the Greenhouse Gas Protocol). As soon as I talked to a few of their customers, I asked my team to see if we can actually finally do this, and build it. My thinking is this: some engineers will care, and some will not (or maybe some will love it and some will hate it!). For those who care, cost and carbon are actually linked; meaning if you reduce the carbon, you usually reduce the cost of the cloud too. It can act as another motivation factor. And now, it is here, and I’d love your feedback. Try it out by going to https://ift.tt/JBZ5FPY , create an account, set up with the GitHub app or GitLab app, and send a pull request with Terraform changes (you can use our example terraform file). It will then show you the cost impact alongside the carbon impact, and how you can optimize it. I’d especially love to hear your feedback on if you think carbon is a big driver for engineers within your teams, or if carbon is a big driver for your company (i.e. is there anything top-down about carbon). AMA - I’ll be monitoring the thread :) Thanks https://ift.tt/JBZ5FPY January 21, 2026 at 08:34PM
Show HN: Xv6OS – A modified MIT xv6 with GUI I've been working on a hobby project to transform the traditional xv6 teaching OS into a graphical environment. Key Technical Features: GUI Subsystem: I implemented a kernel-level window manager and drawing primitives. Mouse Support: Integrated a PS/2 mouse driver for navigation. Custom Toolchain: I used Python scripts (Pillow) and Go to convert PNG assets and TTF fonts into C arrays for the kernel. Userland: Includes a terminal, file explorer, text editor, and a Floppy Bird game. The project is built for i386 using a monolithic kernel design. You can find the full source code and build instructions here: https://ift.tt/VXkQrm1 January 20, 2026 at 10:46PM
Show HN: Trinity – a native macOS Neovim app with Finder-style projects Hi HN, I built Trinity, a native macOS app that wraps Neovim with a project-centric UI. The goal was to keep Neovim itself untouched, but provide a more Mac-native workflow: – Finder-style project browser – Multiple projects/windows – Markdown preview, image/pdf viewer – Native menus, shortcuts, and windowing – Minimal UI, no GPU effects or terminal emulation It’s distributed directly (signed + notarized PKG) and uses Sparkle for incremental updates. This started as a personal tool after bouncing between terminal Neovim and heavier editors. Curious to hear feedback from other Neovim users, especially on what feels right or wrong in a GUI wrapper. Site: https://ift.tt/Tkb37Jx Direct download: https://ift.tt/DXUL4fM... https://ift.tt/Tkb37Jx January 20, 2026 at 11:14PM
Show HN: Homunculus – A self-rewriting Claude Code plugin Homunculus is a Claude Code plugin that watches how you work and writes new capabilities into itself. If you keep doing something repeatedly—checking docs before API calls, running the same debug flow, formatting PRs a certain way—it notices and offers to automate it. Accept, and it writes a new markdown file into its own structure. The plugin literally changes based on what you do. It can create: Commands (explicit shortcuts) Skills (context-triggered behaviors) Subagents (specialists for specific problem domains) Hooks (event-driven, like "run tests when these files change") What actually works (v0.1): Commands are deterministic. Skills are probabilistic—they fire when Claude decides they're relevant, maybe 50-80% of the time. It's an experiment in making LLM tooling adaptive rather than static. State stored in .claude/homunculus/. Each project gets its own instance. https://ift.tt/BvnKL0f January 19, 2026 at 11:23PM
Show HN: Xenia – A monospaced font built with a custom Python engine I'm an engineer who spent the last year fixing everything I hated about monofonts (especially that double-story 'a'). I built a custom Python-based procedural engine to generate the weights because I wanted more logical control over the geometry. It currently has 700+ glyphs and deep math support. Regular weight is free for the community. I'm releasing more weights based on interest. https://ift.tt/NhMqgf8 January 18, 2026 at 04:09PM
Show HN: built a 24h-clock based radial planner to help with ADHD time blindness Hey HN, I built DayZen because of the struggle I had with my ADHD and time management. The core idea: Show your day as a 24-hour radial clock. Tasks become colored arcs around the circle. When the circle fills up, it's full. How it works: Drag tasks onto the clock face and resize them with your finger Overlapping tasks glow red immediately (no more finding conflicts at 11 PM) Two-way calendar sync keeps everything in one place Widgets show your current task and what's next What makes it different: Most planners ask "what should I do?" DayZen asks "what actually fits?" It's not about motivation, it's about physics. You can't fit 10 pounds into a 5-pound bag. I originally made this for my ADHD brain, which has zero time awareness. Turns out the visual approach helps anyone who thinks spatially or chronically overschedulules. Tech stack: Native iOS (Swift/SwiftUI), local-first with optional iCloud sync, zero tracking or accounts required. The app is live on the App Store now. Happy to answer questions about the approach, technical choices, or why I thought a radial interface would work when every other planner uses lists. https://www.dayzen.xyz January 18, 2026 at 09:35PM
Show HN: Minikv – Distributed key-value and object store in Rust (Raft, S3 API) Hi HN, I’m releasing minikv, a distributed key-value and object store in Rust. What is minikv? minikv is an open-source, distributed storage engine built for learning, experimentation, and self-hosted setups. It combines a strongly-consistent key-value database (Raft), S3-compatible object storage, and basic multi-tenancy. I started minikv as a learning project about distributed systems, and it grew into something production-ready and fun to extend. Features/highlights: - Raft consensus with automatic failover and sharding - S3-compatible HTTP API (plus REST/gRPC APIs) - Pluggable storage backends: in-memory, RocksDB, Sled - Multi-tenant: per-tenant namespaces, role-based access, quotas, and audit - Metrics (Prometheus), TLS, JWT-based API keys - Easy to deploy (single binary, works with Docker/Kubernetes) Quick demo (single node): git clone https://ift.tt/MLXmIQs cd minikv cargo run --release -- --config config.example.toml curl localhost:8080/health/ready # S3 upload + read curl -X PUT localhost:8080/s3/mybucket/hello -d "hi HN" curl localhost:8080/s3/mybucket/hello Docs, cluster setup, and architecture details are in the repo. I’d love to hear feedback, questions, ideas, or your stories running distributed infra in Rust! Repo: https://ift.tt/SXcxmal Crate: https://ift.tt/fAr3l4E https://ift.tt/SXcxmal January 18, 2026 at 01:09AM
Show HN: Reddit GDPR Export Viewer – Built After Ban, Unban, Reban Show HN: Reddit GDPR Export Viewer – Built After Getting Hacked, Reinstated, Then Banned Again A few months ago, I posted here about getting my 10-year Reddit account hacked despite 2FA: https://ift.tt/xWYGDXF The likely culprit: session cookie theft via a malicious browser extension, possibly linked to the ShadyPanda campaign that infected 4.3M browsers. Reddit eventually reinstated my account with zero explanation. Then, exactly one month later, they banned me again – permanently, with no reason given and no appeal process. This drove home a lesson: platforms can and will revoke your access arbitrarily, taking years of contributions with them. So I requested my GDPR data export. What I received was not really usable: raw CSV files with no way to meaningfully browse a decade of comments, posts, and activity. So I built this: https://ift.tt/Buc6kOG It's a pure client-side viewer – zero backend, your data never leaves your machine. Open the HTML file, load your Reddit export, and browse your history offline. Full disclosure: I've been vibe coding with Claude Opus for the past few weeks, creating mostly Gravity Forms and WordPress extensions for work (18 repos so far). This particular project was knocked out in a couple of hours. I don't have a strong technical background, so this might be pretty badly coded. It works for what I needed, though. If you find issues or have suggestions for improvements, PRs are welcome. https://ift.tt/Buc6kOG January 17, 2026 at 10:45PM
Show HN: 1Code – Open-source Cursor-like UI for Claude Code Hi, we're Sergey and Serafim. We've been building dev tools at 21st.dev and recently open-sourced 1Code ( https://1code.dev ), a local UI for Claude Code. Here's a video of the product: https://www.youtube.com/watch?v=Sgk9Z-nAjC0 Claude Code has been our go-to for 4 months. When Opus 4.5 dropped, parallel agents stopped needing so much babysitting. We started trusting it with more: building features end to end, adding tests, refactors. Stuff you'd normally hand off to a developer. We started running 3-4 at once. Then the CLI became annoying: too many terminals, hard to track what's where, diffs scattered everywhere. So we built 1Code.dev, an app to run your Claude Code agents in parallel that works on Mac and Web. On Mac: run locally, with or without worktrees. On Web: run in remote sandboxes with live previews of your app, mobile included, so you can check on agents from anywhere. Running multiple Claude Codes in parallel dramatically sped up how we build features. What’s next: Bug bot for identifying issues based on your changes; QA Agent, that checks that new features don't break anything; Adding OpenCode, Codex, other models and coding agents. API for starting Claude Codes in remote sandboxes. Try it out! We're open-source, so you can just bun build it. If you want something hosted, Pro ($20/mo) gives you web with live browser previews hosted on remote sandboxes. We’re also working on API access for running Claude Code sessions programmatically. We'd love to hear your feedback! https://ift.tt/u2Bwm5J January 16, 2026 at 12:50AM
Show HN: Tabstack – Browser infrastructure for AI agents (by Mozilla) Hi HN, My team and I are building Tabstack to handle the "web layer" for AI agents. Launch Post: https://ift.tt/nog1XsY... Maintaining a complex infrastructure stack for web browsing is one of the biggest bottlenecks in building reliable agents. You start with a simple fetch, but quickly end up managing a complex stack of proxies, handling client-side hydration, and debugging brittle selectors. and writing custom parsing logic for every site. Tabstack is an API that abstracts that infrastructure. You send a URL and an intent; we handle the rendering and return clean, structured data for the LLM. How it works under the hood: - Escalation Logic: We don't spin up a full browser instance for every request (which is slow and expensive). We attempt lightweight fetches first, escalating to full browser automation only when the site requires JS execution/hydration. - Token Optimization: Raw HTML is noisy and burns context window tokens. We process the DOM to strip non-content elements and return a markdown-friendly structure that is optimized for LLM consumption. - Infrastructure Stability: Scaling headless browsers is notoriously hard (zombie processes, memory leaks, crashing instances). We manage the fleet lifecycle and orchestration so you can run thousands of concurrent requests without maintaining the underlying grid. On Ethics: Since we are backed by Mozilla, we are strict about how this interacts with the open web. - We respect robots.txt rules. - We identify our User Agent. - We do not use requests/content to train models. - Data is ephemeral and discarded after the task. The linked post goes into more detail on the infrastructure and why we think browsing needs to be a distinct layer in the AI stack. This is obviously a very new space and we're all learning together. There are plenty of known unknowns (and likely even more unknown unknowns) when it comes to agentic browsing, so we’d genuinely appreciate your feedback, questions, and tips. Happy to answer questions about the stack, our architecture, or the challenges of building browser infrastructure. January 15, 2026 at 12:03AM
Show HN: OpenWork – an open-source alternative to Claude Cowork hi hn, i built openwork, an open-source, local-first system inspired by claude cowork. it’s a native desktop app that runs on top of opencode (opencode.ai). it’s basically an alternative gui for opencode, which (at least until now) has been more focused on technical folks. the original seed for openwork was simple: i have a home server, and i wanted my wife and i to be able to run privileged workflows. things like controlling home assistant, or deploying custom web apps (e.g. our customs recipe app recipes.benjaminshafii.com), legal torrents, without living in a terminal. our initial setup was running the opencode web server directly and sharing credentials to it. that worked, but i found the web ui unreliable and very unfriendly for non-technical users. the goal with openwork is to bring the kind of workflows i’m used to running in the cli into a gui, while keeping a very deep extensibility mindset. ideally this grows into something closer to an obsidian-style ecosystem, but for agentic work. some core principles i had in mind: - open by design: no black boxes, no hosted lock-in. everything runs locally or on your own servers. (models don’t run locally yet, but both opencode and openwork are built with that future in mind.) - hyper extensible: skills are installable modules via a skill/package manager, using the native opencode plugin ecosystem. - non-technical by default: plans, progress, permissions, and artifacts are surfaced in the ui, not buried in logs. you can already try it: - there’s an unsigned dmg - or you can clone the repo, install deps, and if you already have opencode running it should work right away it’s very alpha, lots of rough edges. i’d love feedback on what feels the roughest or most confusing. happy to answer questions. https://ift.tt/u0sfIc8 January 14, 2026 at 10:25AM
Show HN: Repomance: A Tinder style app for GitHub repo discovery Hi everyone, Repomance is an app for discovering curated and trending repositories. Swipe to star them directly using your GitHub account. It is currently available on iOS, iPadOS, and macOS. I plan to develop an Android version once the app reaches 100 users. Repomance is open source: https://ift.tt/drDmL7z All feedback is welcome, hope you enjoy using it. https://ift.tt/GDuQTbC January 15, 2026 at 12:24AM
Show HN: Sparrow-1 – Audio-native model for human-level turn-taking without ASR For the past year I've been working to rethink how AI manages timing in conversation at Tavus. I've spent a lot of time listening to conversations. Today we're announcing the release of Sparrow-1, the most advanced conversational flow model in the world. Some technical details: - Predicts conversational floor ownership, not speech endpoints - Audio-native streaming model, no ASR dependency - Human-timed responses without silence-based delays - Zero interruptions at sub-100ms median latency - In benchmarks Sparrow-1 beats all existing models at real world turn-taking baselines I wrote more about the work here: https://ift.tt/uqrlAwi... https://ift.tt/ZIBGp0A January 14, 2026 at 11:31PM
Show HN: Self-host Reddit – 2.38B posts, works offline, yours forever Reddit's API is effectively dead for archival. Third-party apps are gone. Reddit has threatened to cut off access to the Pushshift dataset multiple times. But 3.28TB of Reddit history exists as a torrent right now, and I built a tool to turn it into something you can browse on your own hardware. The key point: This doesn't touch Reddit's servers. Ever. Download the Pushshift dataset, run my tool locally, get a fully browsable archive. Works on an air-gapped machine. Works on a Raspberry Pi serving your LAN. Works on a USB drive you hand to someone. What it does: Takes compressed data dumps from Reddit (.zst), Voat (SQL), and Ruqqus (.7z) and generates static HTML. No JavaScript, no external requests, no tracking. Open index.html and browse. Want search? Run the optional Docker stack with PostgreSQL – still entirely on your machine. API & AI Integration: Full REST API with 30+ endpoints – posts, comments, users, subreddits, full-text search, aggregations. Also ships with an MCP server (29 tools) so you can query your archive directly from AI tools. Self-hosting options: - USB drive / local folder (just open the HTML files) - Home server on your LAN - Tor hidden service (2 commands, no port forwarding needed) - VPS with HTTPS - GitHub Pages for small archives Why this matters: Once you have the data, you own it. No API keys, no rate limits, no ToS changes can take it away. Scale: Tens of millions of posts per instance. PostgreSQL backend keeps memory constant regardless of dataset size. For the full 2.38B post dataset, run multiple instances by topic. How I built it: Python, PostgreSQL, Jinja2 templates, Docker. Used Claude Code throughout as an experiment in AI-assisted development. Learned that the workflow is "trust but verify" – it accelerates the boring parts but you still own the architecture. Live demo: https://online-archives.github.io/redd-archiver-example/ GitHub: https://ift.tt/pnEYIHg (Public Domain) Pushshift torrent: https://ift.tt/glNjSo0... https://ift.tt/pnEYIHg January 13, 2026 at 09:05PM
Show HN: Data from a mixed-brand LiFePO₄ battery bank Hi HN — I’m sharing an empirical, long-term dataset from a DIY energy-storage project that ended up testing a common assumption in battery design. Conventional advice says never mix battery brands. That guidance is well-founded for series strings, but there’s surprisingly little data on purely parallel configurations. I built a 12 V, 500 Ah LiFePO₄ battery bank (1S5P) using mixed-brand cells and instrumented it for continuous monitoring over 73+ days, including high-frequency voltage sampling. The goal was to see whether cell-level differences actually manifest over time in a parallel topology. What the data shows No progressive voltage divergence across the observation period Voltage spread remained within ~10–15 mV Measured Peukert exponent ≈ 1.00 Thermal effects were small relative to instrumentation noise In practice, the parallel architecture appears to force electrical convergence when interconnect resistance is low. I’ve been referring to this as “architectural immunity” — the idea that topology can dominate cell-level mismatch under specific conditions. This is not a recommendation to mix batteries casually, and it’s not a safety guarantee. It’s an attempt to replace folklore with measurements and to define the boundary conditions where this does or does not hold. Everything is public: Raw CSV data Analysis scripts Full PDF report Replication protocol Repo: https://ift.tt/lqbOCVd I’m posting this to invite critique — especially around failure modes, instrumentation limits, or cases where this model would break down (e.g., higher C-rates, aging asymmetry, thermal gradients, different chemistries). Happy to answer technical questions. January 13, 2026 at 11:23PM
Show HN: Customizable OSINT dashboard to monitor the situation Monitor the situation to your own liking. Polymarket, Subway Surfers, Bluesky integration, Flight trackers. Runs all requests client side and doesn't store information. Open to feedback. https://ift.tt/8bDv5oQ January 12, 2026 at 11:02PM
Show HN: Sidecar – AI Social Manager (Analyzes past hits to write new posts) Hi HN, I built Sidecar ( https://sidecar.bz ) because I was having issues maintaining a social media presence for my last startup. I would spend a lot of time trying to create content, but I often froze up or burned out, and the marketing died. How it works: Instead of guessing what to write, Sidecar connects to your existing accounts (Threads, Bluesky, Mastodon, Facebook, Instagram) and analyzes your past posts to see what actually worked. It uses that data to generate weeks of new, text-based content that mimics your successful posts, which you can then bulk schedule in one go. I’d love to hear what you think of Sidecar. You can use code HNLAUNCH for a free month if you want to test the ai features. https://ift.tt/TzgyfVQ January 12, 2026 at 10:48PM
Show HN: Interactive California Budget (by Claude Code) There's been a lot of discussion around the california budget and some proposed tax policies, so I asked claude code to research the budget and turn it into an interactive dashboard. Using async subagents claude was able to research ~a dozen budget line items at once across multiple years, adding lots of helpful context and graphs to someone like me who was starting with little familiarity. It still struggles with frontend changes, but for research this probably 20-40x's my throughput. Let me know any additional data or visualizations that would be interesting to add! https://ift.tt/QrLVDKU January 11, 2026 at 10:43PM
Show HN: Persistent Memory for Claude Code (MCP) This is my attempt in building a memory that evolves and persist for claude code. My approach is inspired from Zettelkasten method, memories are atomic, connected and dynamic. Existing memories can evolve based on newer memories. In the background it uses LLM to handle linking and evolution. I have only used it with claude code so far, it works well with me but still early stage, so rough edges likely. I'm planning to extend it to other coding agents as I use several different agents during development. Looking for feedbacks! https://ift.tt/DmUepRh January 11, 2026 at 02:04AM
Show HN: I used Claude Code to discover connections between 100 books I think LLMs are overused to summarise and underused to help us read deeper. I built a system for Claude Code to browse 100 non-fiction books and find interesting connections between them. I started out with a pipeline in stages, chaining together LLM calls to build up a context of the library. I was mainly getting back the insight that I was baking into the prompts, and the results weren't particularly surprising. On a whim, I gave CC access to my debug CLI tools and found that it wiped the floor with that approach. It gave actually interesting results and required very little orchestration in comparison. One of my favourite trail of excerpts goes from Jobs’ reality distortion field to Theranos’ fake demos, to Thiel on startup cults, to Hoffer on mass movement charlatans ( https://ift.tt/b9y1ejK ). A fun tendency is that Claude kept getting distracted by topics of secrecy, conspiracy, and hidden systems - as if the task itself summoned a Foucault’s Pendulum mindset. Details: * The books are picked from HN’s favourites (which I collected before: https://ift.tt/f4U0qS6 ). * Chunks are indexed by topic using Gemini Flash Lite. The whole library cost about £10. * Topics are organised into a tree structure using recursive Leiden partitioning and LLM labels. This gives a high-level sense of the themes. * There are several ways to browse. The most useful are embedding similarity, topic tree siblings, and topics cooccurring within a chunk window. * Everything is stored in SQLite and manipulated using a set of CLI tools. I wrote more about the process here: https://ift.tt/YseuSCM I’m curious if this way of reading resonates for anyone else - LLM-mediated or not. https://ift.tt/EqUT1QA January 10, 2026 at 10:26PM
Show HN: 15 Years of StarCraft II Balance Changes Visualized Interactively Hi HN! "Never perfect. Perfection goal that changes. Never stops moving. Can chase, cannot catch." - Abathur ( https://www.youtube.com/watch?v=pw_GN3v-0Ls ) StarCraft 2 is one of the most balanced games ever - thanks to Blizzard’s pursuit of perfection. It has been over 15 years since the release of Wings of Liberty and over 10 years since the last installment, Legacy of the Void. Yet, balance updates continue to appear, changing how the game plays. Thanks to that, StarCraft is still alive and well! I decided to create an interactive visualization of all balance changes, both by patch and by unit, with smooth transitions. I had this idea quite a few years ago, yet LLMs made it possible - otherwise, I wouldn't have had the time to code or to collect all changes from hundreds of patches (not all have balance updates). It took way more time than expected - both dealing with parsing data and dealing with D3.js transitions. Pretty much pure vibe coding with Claude Code and Opus 4.5 - while constantly using Playwright skills and consulting Gemini 3 Pro ( https://ift.tt/kXHEGA2 ). While Opus 4.5 was much better at executing, it was often essential to use Gemini to get insights, to get cleaner code, or to inspect screenshots. The difference in quality was huge. Still, it was tricky, as LLMs do not know D3.js nearly as well as React. The D3.js transition part is a thing that sometimes I think would be better to do manually, and only use LLMs for details. But it was also a lesson. Enjoy! Source code is here: https://ift.tt/LVtRqWY https://ift.tt/nk8Xcy9 January 10, 2026 at 11:07PM
Show HN: Various shape regularization algorithms Shape regularization is a technique used in computational geometry to clean up noisy or imprecise geometric data by aligning segments to common orientations and adjusting their positions to create cleaner, more regular shapes. I needed a Python implementation so started with the examples implemented in CGAL then added a couple more for snap and joint regularization and metric regularization. https://ift.tt/WDCYiwQ January 9, 2026 at 07:43AM
Show HN: Similarity = cosine(your_GitHub_stars, Karpathy) Client-side GitHub profile analysis - Build your embedding from your Stars - Compare and discover popular people with similar interests and share yours - Generate a Skill Radar - Recommend repositories you might like https://puzer.github.io/github_recommender/ January 6, 2026 at 06:53PM
Show HN: macOS menu bar app to track Claude usage in real time I built a macOS menu bar app to track Claude usage in real time via API after hitting limits mid-flow too often. Signed and notarised by Apple. Open source. https://ift.tt/PxnrY7j https://ift.tt/WT92kom https://ift.tt/PxnrY7j January 8, 2026 at 11:54PM
Show HN: TierHive – Hourly-billed NAT VPS with private /24 subnets This idea has been floating in my head for about 10 years. Some of you might remember LowEndSpirit.com back before it became a forum, I started that. I've been obsessed with making tiny, cheap VPS actually useful ever since. TierHive is my attempt to make 128MB VPS great again :) It's a NAT VPS (KVM) platform with true hourly billing. Spin up a server, use it for 3 hours, delete it, pay for 3 hours. No monthly commitments, no minimums beyond a $5 top-up. The tradeoff is NAT (no dedicated IPv4), but I've tried to make that less painful: - Every account gets a /24 private subnet with full DHCP management. - Every server gets auto ssh port forwarding and a few TCP/UDP ports - Built-in HAProxy with Let's Encrypt SSL, load balancing, and auto-failover - WireGuard mesh between locations (Canada, Germany, UK currently) - PXE/iPXE boot support for custom installs - Email relay with DKIM/SPF - Recipe system for one-click deploys Still in alpha. Small team, rough edges, but I've been running my own stuff on it for months. Would love feedback — especially on whether the NAT tradeoff kills it for your use cases, or what's missing. (IPv6 is coming) https://tierhive.com https://tierhive.com/ January 8, 2026 at 11:14PM
Show HN: A to Z – A word game I built from a childhood road trip memory Long time lurker here. My family had a pen-and-paper game we'd play on long drives to visit my great-grandmother. After she passed, I spent the holidays recreating it: https://a26z.fun How it works: Find 15 words from a category (like "Stone Fruits," "US States," or "Dog Breeds") as fast as you can. Once you meet the 15 word minimum, you can play for as long as you want. Each letter shows how many target words start with it (A¹ = one word starts with A, N² = two words start with N) That small ² in the bottom-right? Multi-word answers allowed. For "US States" with N², both "NEW YORK" and "NORTH DAKOTA" count Unlimited guesses, 2 hints, and a shuffle button to reorder by frequency. Example: Category: US States | Letters: A¹ M¹ N² S² Answers: ALABAMA, MONTANA, NEW MEXICO, SOUTH DAKOTA If you're into Connections or Strands, this scratches a similar itch but with a deduction twist. https://a26z.fun/ January 8, 2026 at 01:01AM
Show HN: Unicode cursive font generator that checks cross-platform compatibility Hi HN, Unicode “cursive” and script-style fonts are widely used on social platforms, but many of them silently break depending on where they’re pasted — some render as tofu, some get filtered, and others display inconsistently across platforms. I built a small web tool that explores this problem from a compatibility-first angle: Instead of just converting text into cursive Unicode characters, the tool: • Generates multiple cursive / script variants based on Unicode blocks • Evaluates how safe each variant is across major platforms (Instagram, TikTok, Discord, etc.) • Explains why certain Unicode characters are flagged or unstable on specific platforms • Helps users avoid styles that look fine in one app but break in another Under the hood, it’s essentially mapping Unicode script characters and classifying them based on known platform filtering and rendering behaviors, rather than assuming “Unicode = universal.” This started as a side project after repeatedly seeing “fancy text” fail unpredictably in real usage. Feedback, edge cases, or Unicode quirks I may have missed are very welcome. https://ift.tt/N5iZ42a January 1, 2026 at 07:37PM
Show HN: Vibe Coding a static site on a $25 Walmart Phone Hi! I took a cheap $25 walmart phone and put a static server on it? Why? Just for a fun weekend project. I used Claude Code for most of the setup. I had a blast. It's running termux, andronix, nginx, cloudflared and even a prometheus node exporter. Here's the site: https://ift.tt/hqiFQVR https://ift.tt/plYiumM January 4, 2026 at 01:09AM
Show HN: A New Year gift for Python devs–My self-healing project's DNA analyzer I built a system that maps its own "DNA" using AST to enable self-healing capabilities. Instead of a standard release, I’ve hidden the core mapping engine inside a New Year gift file in the repo for those who like to explore code directly. It’s not just a script; it’s the architectural vision behind Ultra Meta. Check the HAPPY_NEW_YEAR.md file for the source https://ift.tt/S1cRjOz January 4, 2026 at 12:50AM
Show HN: Go-Highway – Portable SIMD for Go Go 1.26 adds native SIMD via GOEXPERIMENT=simd. This library provides a portability layer so the same code runs on AVX2, AVX-512, or falls back to scalar. Inspired by Google's Highway C++ library. Includes vectorized math (exp, log, sin, tanh, sigmoid, erf) since those come up a lot in ML/scientific code and the stdlib doesn't have SIMD versions. algo.SigmoidTransform(input, output) Requires go1.26rc1. Feedback welcome. https://ift.tt/TkhXcGD January 3, 2026 at 04:06AM
Show HN: Fluxer – open-source Discord-like chat Hey HN, and happy new year! I'm Hampus Kraft [1], a 22-year-old software developer nearing completion of my BSc in Computer Engineering at KTH Royal Institute of Technology in Sweden. I've been working on Fluxer on and off for about 5 years, but recently decided to work on it full-time and see how far it could take me. Fluxer is an open source [2] communication platform for friends, groups, and communities (text, voice, and video). It aims for "modern chat app" feature coverage with a familiar UX, while being developed in the open and staying FOSS (AGPLv3). The codebase is largely written in TypeScript and Erlang. Try it now (no email or password required): https://ift.tt/lAcNE1D – this creates an "unclaimed account" (date of birth only) so you can explore the platform. Unclaimed accounts can create/join communities but have some limitations. You can claim your account with email + password later if you want. I've developed this solo , with limited capital from some early supporters and testers. Please keep this in mind if you find what I offer today lacking; I know it is! I'm sharing this now to find contributors and early supporters who want to help shape this into the chat app you actually want. ~~~ Fluxer is not currently end-to-end encrypted, nor is it decentralised or federated. I'm open to implementing E2EE and federation in the future, but they're complex features, and I didn't want to end up like other community chat apps [3] that get criticised for broken core functionality and missing expected features while chasing those goals. I'm most confident on the backend and web app, so that's where I've focused. After some frustrating attempts with React Native, I'm sticking with a mobile PWA for now (including push notification support) while looking into Skip [4] for a true native app. If someone with more experience in native dev has any thoughts, let me know! Many tech-related communities that would benefit from not locking information into walled gardens still choose Discord or Slack over forum software because of the convenience these platforms bring, a choice that is often criticised [5][6][7]. I will not only work on adding forums and threads, but also enable opt-in publishing of forums to the open web, including RSS/Atom feeds, to give you the best of both worlds. ~~~ I don't intend to license any part of the software under anything but the AGPLv3, limit the number of messages [8], or have an SSO tax [9]. Business-oriented features like SSO will be prioritised on the roadmap with your support. You'd only pay for support and optionally for sponsored features or fixes you'd like prioritised. I don't currently plan on SaaS, but I'm open to support and maintenance contracts. ~~~ I want Fluxer to become an easy-to-deploy, fully FOSS Discord/Slack-like platform for companies, communities, and individuals who want to own their chat infrastructure, or who wish to support an independent and bootstrapped hosted alternative. But I need early adopters and financial support to keep working on it full-time. I'm also very interested in code contributors since this is a challenging project to work on solo. My email is hampus@fluxer.app. ~~~ There’s a lot more to be said; I’ll be around in the comments to answer questions and fix things quickly if you run into issues. Thank you, and wishing you all the best in the new year! [1] https://ift.tt/Njhlitv [2] https://ift.tt/TdGRuI1 [3] https://ift.tt/lWHBVD3 [4] https://skip.tools/ [5] https://ift.tt/FRsTGJV [6] https://ift.tt/SUpWtVA [7] https://ift.tt/ZSIaKxp [8] https://ift.tt/t2Qope0 [9] https://sso.tax/ https://fluxer.app January 3, 2026 at 01:30AM
Show HN: Feature detection exploration in Lidar DEMs via differential decomp I'm not a geospatial expert — I work in AI/ML. This started when I was exploring LiDAR data with agentic assitince and noticed that different signal decomposition methods revealed different terrain features. The core idea: if you systematically combine decomposition methods (Gaussian, bilateral, wavelet, morphological, etc.) with different upsampling techniques, each combination has characteristic "failure modes" that selectively preserve or eliminate certain features. The differences between outputs become feature-specific filters. The framework tests 25 decomposition × 19 upsampling methods across parameter ranges — about 40,000 combinations total. The visualization grid makes it easy to compare which methods work for what. Built in Cursor with Opus 4.5, NumPy, SciPy, scikit-image, PyWavelets, and OpenCV. Apache 2.0 licensed. I'd appreciate feedback from anyone who actually works with elevation data. What am I missing? What's obvious to practitioners that I wouldn't know? https://ift.tt/nm4BkPT January 1, 2026 at 05:59AM
Show HN: VectorDBZ, a desktop GUI for vector databases Hi HN, I built VectorDBZ, a cross-platform desktop app for exploring and analyzing vector databases like Qdrant, Weaviate, Milvus, and ChromaDB. It lets you browse collections, inspect vectors and metadata, run similarity searches, and visualize embeddings without writing custom scripts. GitHub (downloads and issues): https://ift.tt/jQ1dkRC Feedback welcome. If it’s useful, starring the repo helps keep me motivated. Thanks. https://ift.tt/jQ1dkRC January 1, 2026 at 08:55PM
Show HN: A Prompt-Injection Firewall for AI Agents and RAG Pipelines We built SafeBrowse — an open-source prompt-injection firewall for AI systems. Instead of relying on better prompts, SafeBrowse enforces a hard security boundary between untrusted web content and LLMs. It blocks hidden instructions, policy violations, and poisoned data before the AI ever sees it. Features: • Prompt injection detection (50+ patterns) • Policy engine (login/payment blocking) • Fail-closed by design • Audit logs & request IDs • Python SDK (sync + async) • RAG sanitization PyPI: pip install safebrowse Looking for feedback from AI infra, security, and agent builders. January 1, 2026 at 02:31AM
Show HN: A web-based lighting controller built because my old became a brick I’m a student and I built this because my old lightning controller (DMX) became a brick after the vendor’s control software was deprecated in 2025. My focus was entirely on developing a robust backend architecture to guarantee maximum performance. Everything is released under GPLv3. The current frontend is just a "vibecoded" dashboard made with plain HTML and JavaScript to keep rendering latency as low as possible. In earlier versions Svelte was used. Svelte added too much complexity for an initial mvp. Video: https://ift.tt/qxubY6O Repo: https://ift.tt/TlJUut0 Technical Details: The system uses a distributed architecture where a FastAPI server manages the state in a Redis. State changes are pushed via WebSockets to Raspberry Pi gateways, which then independently maintain the constant 44Hz binary stream to the lights. This "push model" saves massive amounts of bandwidth and ensures low latency. In a stress test, I processed 10 universes (5,120 channels) at 44Hz with zero packet loss (simulated). An OTP-based pairing makes the setup extremely simple (plug-and-play). I’m looking forward to your feedback on the architecture and the Redis approach! Happy New Year! https://ift.tt/TlJUut0 December 31, 2025 at 10:16PM
Show HN: Fleet / Event manager for Star Citizen MMO I built an open-source org management platform for Star Citizen, a space MMO where player orgs can have 50K+ members managing fleets worth millions. https://scorg.org The problem: SC's official tools won't launch until 2026, but players need to coordinate now - track 100+ ship fleets, schedule ops across timezones, manage alliances, and monitor voice activity during battles. Interesting challenges solved: 1. Multi-org data isolation - Users join multiple orgs, so every query needs scoping. 2. Canvas + Firebase Storage CORS - Couldn't export fleet layouts as PNG. Solution: fetch images as blobs, convert to base64 data URLs, then draw to canvas. No CORS config needed. 3. Discord bot - Built 4 microservices (VoiceActivityTracker, EventNotifier, ChannelManager, RoleSync) sharing Firebase state. Auto-creates channels for ops, cleans up when done. Features: role-based access, event calendar with RSVP, LFG matchmaking, drag-and-drop fleet builder, economy tools, alliance system, analytics dashboard, mobile-responsive. ~15 pages, fully functional. Custom military-inspired UI (monospace, gold accents). January 1, 2026 at 12:48AM
Show HN: I remade my website in the Sith Lord Theme and I hope it's fun I used the time over Christmas and in between the years to redesign my website. This time I decided to make it in the theme of an evil Sith Lord that commands the Galactic Cookie Empire, because I found my previous cookie consent game a bit boring after a while. Here's the website's welcome page and the cookie consent game: https://ift.tt/mZ1rvFy (the cookie consent game isn't started on any other page of my website, only on the welcome page) I also made a "making of" weblog article series, in case you're interested in the development process and how I implemented it and what kind of troubles I went through already: - Making of the Game: https://ift.tt/ojnhezX... - Making of the Avatar: https://ift.tt/ojnhezX... - Debuggers to toy around with: https://ift.tt/6HUhtQ0 It "should" work on modern browsers. I tested it on Firefox on Linuxes, Chrome/Chromium on Linuxes, and Safari on Macbook. Don't have an iPhone so I can't test that, but my two old Android phones were also working fine with the meta viewport hack (I can't believe this is still the "modern" way to do things after 15 years. Wtf). Best experience is of course with a bigger display. On smaller screen sizes, the game will use a camera to zoom around the game world and follow the player's spaceship. Minimum window width is 1280 pixels for no camera, and I think 800 pixels to be playable (otherwise the avatar gets in the way too much in the boss fights). Oh, there's also a secret boss fight that you can unlock when you toy around with the Dev Tools :) What's left to do on the avatar animation side: - I have to backport CMUdict to JavaScript / ECMAScript. That's what I'm working on right now, as I'm not yet satisfied with the timings of the phonemes. Existing tools and pipelines that do this in python aren't realtime, which leads to my next point. - I want to switch to using the "waveform energy detection" and a zero cross rate detector to time phonemes more correctly. I believe that changes in waveforms and their structures can detect differences in phonemes, but it's a gut feeling and not a scientific fact. Existing phoneme animation papers were kind of shit and broken (see my making of article 2). The phoneme boundary detector is highly experimental though and is gonna need a couple weeks more time until it's finished. That's it for now, hope you gonna enjoy your stay on my website and I hope you gonna have fun playing the Cookie Consent Game :) Oh, also, because it might not be obvious: No LLMs were used in the making of this website. Pretty much everything is hand-coded, and unbundled and unminified on purpose so visitors can learn from the code if they want to. ~Cookie https://ift.tt/mZ1rvFy December 30, 2025 at 11:38PM
Show HN: Brainrot Translator – Convert corporate speak to Gen Alpha and back Hey HN, I built this because the generational gap online is getting wider (and weirder). It’s an LLM-wrapper that translates "Boomer" (normal/corporate English) into "Brainrot" (Gen Alpha slang) and vice versa. It also has an "Image-to-Rot" feature that uses vision to describe uploaded images in slang. It’s mostly for fun, but actually kind of useful for deciphering what your younger cousins are saying. Would love to hear what you think! https://ift.tt/dws3lcb December 30, 2025 at 10:21PM
Show HN: Aroma: Every TCP Proxy Is Detectable with RTT Fingerprinting TL;DR explanation (go to https://ift.tt/SEYRdmh... if you want the formatted version) This is done by measuring the minimum TCP RTT (client.socket.tcpi_min_rtt) seen and the smoothed TCP RTT (client.socket.tcpi_rtt). I am getting this data by using Fastly Custom VCL, they get this data from the Linux kernel (struct tcp_info -> tcpi_min_rtt and tcpi_rtt). I am using Fastly for the Demo since they have PoPs all around the world and they expose TCP socket data to me. The score is calculated by doing tcpi_min_rtt/tcpi_rtt. It's simple but it's what worked best for this with the data Fastly gives me. Based on my testing, 1-0.7 is normal, 0.7-0.3 is normal if the connection is somewhat unstable (WiFi, mobile data, satellite...), 0.3-0.1 is low and may be a proxy, anything lower than 0.1 is flagged as TCP proxy by the current code. https://ift.tt/XDC2sZq December 26, 2025 at 02:04AM
Show HN: Neko.js, a recreation of the first virtual pet Hi HN, Here is a late Christmas present: I rebuilt Neko [1], the classic desktop cat that chases your mouse, as a tiny, dependency-free JavaScript library that runs directly on web pages. Live demo: https://louisabraham.github.io/nekojs/ GitHub: https://ift.tt/W4j5YXx Drop-in usage is a single script tag:
This is a fairly faithful recreation of Neko98: same state machine, same behaviors, same original 32×32 pixel sprites. It follows your cursor, falls asleep when idle, claws walls, and you can click it to cycle behavior modes. What made this project interesting to me is how I built it. I started by feeding the original C++ source (from the Wayback Machine) to Claude and let it "vibe code" a first JS implementation. That worked surprisingly well as a starting point, but getting it truly accurate required a lot of manual fixes: rewriting movement logic, fixing animation timing, handling edge cases the AI missed, etc. My takeaway: coding agents are very useful at resurrecting old codebases, and this is probably the best non-soulless use of AI for coding. It gets you 60–70% of the way there very fast, especially for legacy code that would otherwise rot unread. The last 30% still needs a human who cares about details. The final result is ~38KB uncompressed (~14KB brotli), zero dependencies, and can be dropped into a page with a single
Show HN: A solar system simulation in the browser I didn't realize Universe Sandbox ran on MacOS, and I was in the mood to play around a bit. Some functions it's got: - Random system generation - Sonification is super fun too - Habitability Simulation (Just for fun, don't cite this please) - Replacing, spawning, deleting objects I've had tons of fun building this, so I hope someone else can share the joy. It's free and runs in the browser. I'd love to hear any feedback. I think this is at a state where I might leave it as it is, but if people are interested in other features, maybe I'll keep working on it. I've kept saying I'll stop working on this for a while now though. https://ift.tt/H2f8dVr December 29, 2025 at 11:04PM
Show HN: AgentFuse – A local circuit breaker to prevent $500 OpenAI bills Hey HN, I’ve been building agents recently, and I hit a problem: I fell asleep while a script was running, and my agent got stuck in a loop. I woke up to a drained OpenAI credit balance. I looked for a tool to prevent this, but most solutions were heavy enterprise proxies or cloud dashboards. I just wanted a simple "fuse" that runs on my laptop and stops the bleeding before it hits the API. So I built AgentFuse. It is a lightweight, local library that acts as a circuit breaker for LLM calls. Drop-in Shim: It wraps the openai client (and supports LangChain) so you don't have to rewrite your agent logic. Local State: It uses SQLite in WAL mode to track spend across multiple concurrent agents/terminal tabs. Hard Limits: It enforces a daily budget (e.g., stops execution at $5.00). It’s open source and available on PyPI (pip install agent-fuse). I’d love feedback on the implementation, specifically the SQLite concurrency logic! I tried to make it as robust as possible without needing a separate server process. https://ift.tt/PSUaL5m December 28, 2025 at 12:46AM
Show HN: Buoy – A persistent, status-bar web server for local utilities I’m constantly building small web-based tools for my own use. Usually, my workflow ends with a dilemma: do I keep a terminal tab open forever running `npx http-server -p 8080`, or do I spend time configuring a Caddyfile for a 50-line HTML tool? Nothing felt right. I wanted something that felt like a native, always-on, utility that was easily accessible but invisible. I built Buoy. It’s a minimal server that: Lives in the status bar: I can see that it's running at a glance without hunting through ps aux. Is persistent by default: It starts with macOS and keeps my utilities alive in the background. Zero-config: It points at a XDG‑Standard www folder so I can create a symlink and be done. Small: I wanted to avoid the modern bloat. Buoy is a single, self-contained binary that's under 10MB. It’s a minimal tool that lets me build many small things and move on to the next. https://ift.tt/syPi9Eh December 25, 2025 at 09:51PM
Show HN: nunchux – A handy tmux launcher buddy thing Had some fun over the christmas holidays and nunchux is the output. A fun menu for tmux to reduce the number of apps I need to remember the name for. Also a nice quick way to browse hacker news via hackernews_tui :-) https://ift.tt/qM6Wc0t December 25, 2025 at 10:48PM
Show HN: "What Should I Build?" A directory of what people want Successful entrepreneurs always say that the most profitable tools are the ones that help you solve the issues you’re facing. The problem is, I apparently have no issues. So instead, I built a PoC of a minimalistic ideas directory focused on issues others are facing. Feedback is welcome https://ift.tt/Q0zcAmw December 23, 2025 at 10:27PM
Show HN: Openinary – Self-hosted image processing like Cloudinary Hi HN! I built Openinary because Cloudinary and Uploadcare lock your images and charge per request. Openinary lets you self-host a full image pipeline: transform, optimize, and cache images on your infra; S3, Cloudflare R2, or any S3-compatible storage. It’s the only self-hosted Cloudinary-like tool handling both transformations and delivery with a simple URL API (/t/w_800,h_800,f_avif/sample.jpg). Built with Node.js, Docker-ready. GitHub: https://ift.tt/EmG3dBo Feedback welcome; especially from Cloudinary users wanting the same UX but on their own infra! https://ift.tt/EmG3dBo December 23, 2025 at 09:31PM
Show HN: CleanCloud – Read-only cloud hygiene checks for AWS and Azure Hi HN, I’m a solo founder and SRE background engineer. I built CleanCloud to solve a problem I kept seeing on teams I worked with: cloud accounts slowly filling up with orphaned, unowned, or inactive resources created by elastic systems and IaC — but nobody wants tools that auto-delete things. CleanCloud is a small, open-source CLI that: - Scans AWS and Azure accounts in read-only mode - Identifies potential “hygiene” issues (unattached EBS volumes, old snapshots, inactive CloudWatch logs, untagged storage, unused Azure public IPs, etc.) - Uses conservative signals and confidence levels (HIGH / MEDIUM / LOW) - Never deletes or modifies resources - Is designed for review-only workflows (SRE-friendly, IaC-aware) What it intentionally does NOT do: - No auto-remediation - No cost optimization / FinOps dashboards - No agents, no SaaS, no ML - No recommendations based on a single risky signal This is early-stage and I’m explicitly looking for feedback from SREs / DevOps folks: - Are these the right problems to focus on? - Are the signals conservative enough to be trusted? - What rules would you actually want next? Repo (MIT licensed): https://ift.tt/SZJoyuH If this looks useful, a helps a lot. Brutally honest feedback welcome. Many Thanks Suresh December 22, 2025 at 11:15PM
Show HN: Explore ArXiv/HN/LessWrong with SQL and Vector Algebra I hardened up a Postgres database enough to give the public access to run arbitrary queries over an interesting and growing dataset. There's over 23M embeddings (2048*32 bit Voyage-3.5-Lite over ~300 token chunks, 20% overlapped), 600 GB of indexes for performance, a prompt for letting Claude Code w/ Opus 4.5 drive usage, and there's full support for vector algebra. One can actually embed arbitrary concepts like "FTX" and "guilt" vectors and store these embeddings on the server, and refer to them in queries like @FTX + @guilt. Because you and your agent can write and run 100k character SELECT statements for up to 60s, you can really answer some incredibly nuanced questions very easily now. I'm doing a startup building subjective judgement infrastructure, HMU if interested in angel investment. https://ift.tt/f5hV7i1 December 22, 2025 at 11:06PM
Show HN: Meds — High-performance firewall powered by NFQUEUE and Go Hi HN, I'm the author of Meds ( https://ift.tt/gdM37Qe ). Meds is a user-space firewall for Linux that uses NFQUEUE to inspect and filter traffic. In the latest v0.7.0 release, I’ve added ASN-based filtering using the Spamhaus DROP list (with IP-to-ASN mapping via IPLocate.io). Key highlights: Zero-lock core, ASN Filtering, Optimized Rate Limiting, TLS Inspection, Built-in Prometheus metrics and Swagger API. Any feedback is very welcome! https://ift.tt/gdM37Qe December 22, 2025 at 10:58PM
Show HN: Real-time SF 911 dispatch feed (open source) I built an open-source alternative to Citizen App's paid 911 feed for San Francisco. It streams live dispatch data from SF's official open data portal, uses an LLM to translate police codes into readable summaries, and auto-redacts sensitive locations (shelters, hospitals, etc.). Built it at a hack night after getting annoyed that Citizen is the only real-time option and they paywall it. Repo: https://ift.tt/JogGOzF Discord: https://ift.tt/fsbP7iR Happy to discuss the technical approach or take feedback. https://ift.tt/P0yEUAQ December 22, 2025 at 06:29AM
Show HN: I built a 1‑dollar feedback tool as a Sunday side project I’ve always found it funny how simple feedback widgets end up as $20–$30/month products. The tech is dead simple, infra is cheap, and most of us here could rebuild one in a weekend. So as a “principle experiment” I built my own today as a side project and priced it at 1 dollar. Just because if something is cheap to run and easy to replicate, it should be priced accordingly, and it’s also fun marketing. 1$ feedback tool. Shipped today, got the first users/moneys today, writing this post today. Side Sunday project, then back to the main product tomorrow. https://ift.tt/7x2o9Rh December 22, 2025 at 03:22AM
Show HN: HN Wrapped 2025 - an LLM reviews your year on HN I was looking for some fun project to play around with the latest Gemini models and ended up building this :) Enter your username and get: - Generated roasts and stats based on your HN activity 2025 - Your personalized HN front page from 2035 (inspired by a recent Show HN [0]) - An xkcd-style comic of your HN persona It uses the latest gemini-3-flash and gemini-3-pro-image (nano banana pro) models, which deliver pretty impressive and funny results. A few examples: - dang: https://ift.tt/rcJhCF8 - myself: https://ift.tt/iUrPlVK Give it a try and share yours :) Happy holidays! [0] https://ift.tt/L5aWkeB https://ift.tt/AVvaTeu December 20, 2025 at 07:09PM
Show HN: Automatic Riff Track Creator I'm a big fan of listening to humorous commentary tracks along with my favorite movies. Currently the only way to do this is to start one program with the commentary audio file, and another program with the video file, and use audio cues from both tracks to line them up using the seek bars of each interface while you're watching. It's... an experience. Especially frustrating if you need to pause for a second... This tool allows you to create a new audio track in the video file containing the commentary track merged with an existing audio track from the video. It also allows you to adjust the offset of the commentary track, so you can line it up with the audio in the movie at arbitrary points in case they are not already in sync. The script tries to do this automatically using subtitles and audio analysis, followed by an optional 'fine-tuning' step if you really want it dialed in to the millisecond. I hope this is useful to anyone else who wants to enjoy these "riff tracks" more easily :) https://ift.tt/DAPS7kN December 20, 2025 at 09:42PM
Show HN: MCPShark Viewer (VS Code/Cursor extension)- view MCP traffic in-editor A few days ago I posted MCPShark (a traffic inspector for the Model Context Protocol). I just shipped a VS Code / Cursor extension that lets you view MCP traffic directly in the editor, so you’re not jumping between terminals, logs, and "I think this is what got sent". VS Code Marketplace: https://marketplace.visualstudio.com/items?itemName=MCPShark... Main repo: https://ift.tt/6gHjylf Feature requests / issues: https://ift.tt/NGW40A8 Site: https://mcpshark.sh/ If you’re building MCP agents/tools: what would make MCP debugging actually easy—timeline view, session grouping, diffing tool args, exporting traces, something else? I’d be thankful if you could open a feature request here: https://ift.tt/NGW40A8 December 17, 2025 at 11:49PM
Show HN: Linggen – A local-first memory layer for your AI (Cursor, Zed, Claude) Hi HN, Working with multiple projects, I got tired of re-explaining our complex multi-node system to LLMs. Documentation helped, but plain text is hard to search without indexing and doesn't work across projects. I built Linggen to solve this. My Workflow: I use the Linggen VS Code extension to "init my day." It calls the Linggen MCP to load memory instantly. Linggen indexes all my docs like it’s remembering them—it is awesome. One click loads the full architectural context, removing the "cold start" problem. The Tech: Local-First: Rust + LanceDB. Code and embeddings stay on your machine. No accounts required. Team Memory: Index knowledge so teammates' LLMs get context automatically. Visual Map: See file dependencies and refactor "blast radius." MCP-Native: Supports Cursor, Zed, and Claude Desktop. Linggen saves me hours. I’d love to hear how you manage complex system context! Repo: https://ift.tt/NM6ogdw Website: https://linggen.dev https://ift.tt/NM6ogdw December 19, 2025 at 11:24PM
Show HN: Credible brings credibility scores directly on Hacker News Hi HN, This is Aki, a technical founder having previously shipped products to 1B+ people (I launched the heart button on twitter). I built Credible because I wanted a way to know whether something I'm about to read would be worth my time. I also got tired of context-switching to verify what I read. Credible is a Chrome extension that displays instant credibility scores directly into the pages you browse, including HN itself. ** How it works ** On HN Home: You see a credibility score next to each link. On HN Comments page: You see the full analysis of the linked article. This includes the linked article's key takeaways, credibility score, bias detection, and a breakdown of claims (facts vs opinions vs dubious) without leaving the page. They also show on our mobile-friendly feed here: https://ift.tt/xloHdyj Chrome Web Store: https://ift.tt/H0mGrJ6 We will have a major focus next year on shipping tools that utilize AI to make consumption a breeze. As we design that, would love to know: Is this scoring & UX useful for you? What would make it even better? https://ift.tt/sTcdCbf December 19, 2025 at 10:35PM