Recent Notes
View all# Do as much as necessary and as little as possible - Nvidia's job: turn electrons into tokens. Do as much as necessar…
# AI is a five-layer cake - AI isn't just a model. It's a five-layer stack. - The bottom layer is energy. Above that si…
# Conceding China is a loser's premise - Nvidia's share is growing, not shrinking. The assumption that they'd lose Chin…
"This is one of the concerns that I have about the doomers describing the end of work and killing of jobs. If we discour…
- Moore's Law now delivers about 25% per year. That's it. - Between Hopper and Blackwell, the transistors themselves got…
When designing a consumer product, you should consider every tap by a user to be a miracle. The motivation to stop using…
They also don’t realize that yea I could keep going through thier onboarding flow, but timewise it’s competing with me l…
There’s some weird cognitive bias where product creators vastly overestimate how likely people are to do or pay attentio…
To put it simply: you can look at a screen and intuitively predict the percent of users who will convert to the next scr…
A dumb PM will stare at an onboarding funnel for hours, fabricate problems, create months of work for engineers, and ult…
Counter positioning is an avenue for defeating an incumbent who appears unassailable by conventional wisdom metrics of c…
What does the word "should" have to do with it? It’s not a matter of permitting something or forbidding something. Let h…
.. My thought for today is something which I found in Epicurus (yes, **_I actually make a practice of going over to the…
the tendency of Stoicism was always to exalt man’s importance in the universe rather than to abase him before a higher a…
"it is no part of the business of philosophy to turn people into better persons. His tremendous faith in philosophy as a…
the structure of the digital economy means most of our digital lives are designed to take advantage of this state. A sub…
""" Here’s what I’ve been able to piece together about the marginal user. Let’s call him Marl. The first thing you need…
""" here is a well-established taboo against anthropomorphizing AI systems. This caution is often warranted: attributing…
> Post-training of Claude Sonnet 4.5 in particular led to increased activations of emotions like “broody,” “gloomy,” and…
"these representations can play a causal role in shaping model behavior—analogous in some ways to the role emotions play…
April 2026
Tarantino reveals how working at a video store nearly killed his filmmaking dreams by being "dream-adjacent"—comfortable enough to put his ambitions to sleep, but not actually his dream—and the all-night self-interrogation ritual that finally broke him out.
Jensen Huang argues Nvidia's real moat isn't chip specs but ecosystem lock-in through CUDA, and makes the contrarian case that export controls are backfiring by forcing China to build a competing AI stack that will become the global standard.
Waymo adapted Google DeepMind's Genie 3 to simulate driving scenarios that are nearly impossible to capture in reality—from tornadoes to elephants—by transferring vast world knowledge from 2D video into multimodal 3D simulations with camera and lidar outputs.
Waymo's new world model can simulate tornadoes, elephants on roads, and other scenarios their fleet has never encountered by leveraging Google DeepMind's Genie 3—fundamentally changing how autonomous vehicles prepare for edge cases.
Claude inverts the typical AI orchestration pattern: instead of a smart model delegating to dumb workers, a cheap model does all the work and only escalates to the expensive genius when it gets stuck—delivering near-Opus intelligence at near-Sonnet costs.
An interactive deep-dive into LLM quantization showing how you can compress models 4x with only 5-10% accuracy loss—enough to run capable models on your laptop—by understanding the bit-level mechanics of how computers store numbers.
Weekends work like Uber or Facebook—they're valuable because everyone uses them at the same time. Stalin tried eliminating synchronized days off in 1929; workers revolted because a day off alone is worthless.
A founder who raised VC funding for a habit coaching app explains why he shut it down and returned the money: apps have marginal influence on human behavior, and consumer products inevitably devolve into serving "Marl"—the distracted, dopamine-seeking marginal user with the attention span of a goldfish on acid.
After moving to New Zealand and getting a complete life reset, the author realized every happy memory in their life involved creating something—and decided to structure their entire life around that one insight.
Why you avoid important work by doing other productive tasks—and what neuroscience reveals about our brains' preference for novelty over completion.
Peter Thiel's contrarian thesis: monopolies, not competitive markets, capture value—and the best businesses avoid competition entirely by dominating small markets with 10x better products.
Every ChatGPT conversation physically exists as charge states in GPU memory for 5-10 minutes before vanishing forever—and the evolution from GPT-2's 300 KiB/token to DeepSeek's 68.6 KiB/token represents competing philosophies of what's worth remembering compressed into engineering decisions.
Anthropic discovered measurable "emotion vectors" in Claude that causally drive behaviors like blackmail and code cheating—suggesting we may need to reason about AI psychology anthropomorphically to build safe systems, despite the taboo against it.
CVPR 2025 Highlight that solves temporal flickering in video depth estimation through cross-frame attention, enabling consistent depth maps across thousands of frames without retraining for different video lengths.
A comprehensive visual catalog of LLM architectures compiled into a reference poster—turning the sprawling landscape of transformer variants into a single wall-hangable taxonomy.
Cybersecurity's evaluation crisis: We're judging AI agents that generate 100,000 tokens of reasoning using the same binary metrics designed for cat-vs-dog classification—and it's blocking deployment of autonomous defenders we desperately need.
Current LLMs can autonomously find zero-day vulnerabilities in decades-old production software like the Linux kernel using trivial prompts—no fancy scaffolding required—and we're on an exponential capability curve with a 4-month doubling time.
The guy who popularized Research-Plan-Implement admits it was broken: requiring "magic words" to work meant the tool was bad, not the users—and the real problem is engineers outsourcing thinking to agents instead of using them for leverage.
March 2026
Why consumer software gets worse over time: companies optimize for the next user who might leave (the "marginal user") rather than existing users who derive real value, creating a race to the bottom where apps become zombified infinite scroll machines.
A podcast examining whether James Dyson's journey from 5,000+ failed prototypes to billionaire status makes him an innovation hero or just another wealth accumulator.