Feed aggregator

100 Jumps

Hacker News - Fri, 03/13/2026 - 4:12am

Article URL: https://100jumps.org/play/

Comments URL: https://news.ycombinator.com/item?id=47361797

Points: 2

# Comments: 1

Categories: Hacker News

Show HN: CacheLens – Local-first cost tracking proxy for LLM APIs

Hacker News - Fri, 03/13/2026 - 4:05am

I built CacheLens because I was burning through $200+/month on Claude API calls and had no idea where it was going.

It's a local HTTP proxy that sits between your app and the AI provider (Anthropic, OpenAI, Google). Every request flows through it, and it records token usage, cost, cache hit rates, latency — everything. Then there's a dashboard to visualize it all.

What makes it different from just checking your provider dashboard:

It's real-time (WebSocket live feed of every call as it happens) It works across all three major providers in one view It runs 100% locally — your prompts never leave your machine It has budget caps that actually block requests before you overspend It identifies optimization opportunities (cache misses, model downgrades, repeated prompts) Tech stack: Python, FastAPI, SQLite, vanilla JS. No React, no build step, no external dependencies beyond pip. The whole thing is ~3K lines of Python.

Interesting technical decisions:

The proxy captures streaming responses without buffering — it tees the byte stream so the client sees zero added latency Cost calculation uses a built-in pricing table with override support (providers change rates constantly) There's a Prometheus /metrics endpoint so you can plug it into existing monitoring Cacheability analysis uses diff-based detection across multiple API calls to identify what's actually static vs dynamic in your prompts Limitations I'm honest about:

The cacheability scorer is heuristic-based — solid for multi-call traces (~85% accurate), rougher for single prompts (~65%) Token counting uses cl100k_base for everything, which drifts ~10% for non-OpenAI models Three features (smart routing, scheduled reports, multi-user auth) are on the roadmap but not shipped yet Would love feedback, especially from anyone managing LLM costs at scale.

Comments URL: https://news.ycombinator.com/item?id=47361756

Points: 1

# Comments: 0

Categories: Hacker News

Chrome 146 Update Patches Two Exploited Zero-Days

Security Week - Fri, 03/13/2026 - 3:47am

The flaws can be exploited to manipulate data and bypass security restrictions, potentially leading to code execution.

The post Chrome 146 Update Patches Two Exploited Zero-Days appeared first on SecurityWeek.

Categories: SecurityWeek

Show HN: YAOS – A 1-click deploy, real-time sync engine for Obsidian

Hacker News - Fri, 03/13/2026 - 3:37am

Hey HN,

I'm a heavy Obsidian user.

I recently got tired of the two usual sync tradeoffs:

1. File-based sync (iCloud/Dropbox/Syncthing) that leaves you waiting for changes to propagate, or hands you a "conflicted copy." 2. Self-hosted setups (like CouchDB) that need touching VMs and dockerized databases to sync markdown.

So I built YAOS: a local-first, real-time sync engine for Obsidian.

Self-hosting OSS should have better UX.

You can deploy the backend to your own Cloudflare account in one click. It fits comfortably in Cloudflare's free tier (costing $0/month for normal personal use), and requires absolutely no terminal interaction, no SSH, and no env files.

You can try it out right now: https://github.com/kavinsood/yaos

How it works under the hood:

- Text sync uses Yjs CRDTs. It syncs real-time keystrokes and cursors rather than treating the vault as a pile of files to push around later. - Each vault maps to a Cloudflare Durable Object, giving you a low-latency, single-threaded coordinator at the edge. - The backend uses a chunked Checkpoint + Delta-Journal MVCC storage engine on top of the DO's SQLite storage. - Attachments sync separately via R2 (which is optional—text sync works fine without it).

The hardest part of this project was bridging Obsidian's synchronous UI and noisy OS file-watchers with the in-memory CRDT graph. I had to build a trailing-edge snapshot drain that coalesces rapid-fire IO bursts (like running a find-and-replace-all) into atomic CRDT transactions to prevent infinite write-loops.

The current design keeps a monolithic CRDT per vault. This is great for normal personal notes, but has a hard memory ceiling (~50MB of raw text). I chose this tradeoff because I cared more about fast, boringly reliable real-time ergonomics than unbounded enterprise scale.

I also wrote up engineering notes on the tricky parts (like handling offline folder rename collisions without resurrecting dead files) on GitHub.

I've spent the last three weeks doing brutal QA passes to harden mobile reconnects, IndexedDB quota failures, and offline split-brain directory renames.

I'd love feedback on the architecture, the code, or the trade-offs I made. I'll be hanging out in the thread to answer questions!

Comments URL: https://news.ycombinator.com/item?id=47361635

Points: 2

# Comments: 0

Categories: Hacker News

Shiny Object Syndrome

Hacker News - Fri, 03/13/2026 - 3:33am
Categories: Hacker News

Pages