🦞🌯 Lobster Roll

Stories by rrampage

Perplexity is using stealth, undeclared crawlers to evade no-crawl directives (blog.cloudflare.com)
Firefox-patch-bin, librewolf-fix-bin AUR packages contain malware (lists.archlinux.org)
The Real GenAI Issue (tbray.org)
SQL needed structure (scattered-thoughts.net)
How I Became the First Linux User in India (medium.com)
strace tips for better debugging (rrampage.github.io)
The Claude C Compiler: What It Reveals About the Future of Software (modular.com)
How fast can you open 1000 files? (lemire.me)
When Imperfect Systems are Good, Actually: Bluesky’s Lossy Timelines (jazco.dev)
The "high-level CPU" challenge (2008) (yosefk.com)
(Quite) A Few Words About Async (yoric.github.io)
An Implementation of J - Roger Hui (jsoftware.com)
Append-only programming (iafisher.com)
Databases are the endgame for data-oriented design (spacetimedb.com)
Andreessen’s Folly - The False Dichotomy of Software and Hardware (youtube.com)
APL since 1978 (dl.acm.org)
FAQ on Microsoft’s topological qubit thing (scottaaronson.blog)
microdot: The impossibly small web framework for Python and MicroPython (github.com)
A Mechanically Verified Garbage Collector for OCaml (kcsrk.info)
tinywasm: A tiny, interpreted WebAssembly Runtime written in Rust (github.com)
High-Performance DBMSs with io_uring: When and How to use it (arxiv.org)
Depth is all you need: how Antithesis crushes Gradius (antithesis.com)
Java at 30: The Genius Behind the Code That Changed Tech (thenewstack.io)
Hedy: Creating a Programming Language for Everyone (youtube.com)
Building a vector search engine that lets you choose precision at query time (clickhouse.com)
Programming in Assembly without an Operating System (youtube.com)
Inside NVIDIA GPUs: Anatomy of high performance matmul kernels - Aleksa Gordić (aleksagordic.com)
How to Grow an LSM-tree? Towards Bridging the Gap Between Theory and Practice (arxiv.org)
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling (arxiv.org)
Test-Time Scaling (TTS) is an important method for improving the performance of Large Language Models (LLMs) by using additional computation during the inference phase. However, current studies do not systematically analyze how policy models, Process Reward Models (PRMs), and problem difficulty infl...
Virtualizing NVidia HGX B200 GPUs with Open Source (ubicloud.com)