LWiAI Podcast #238 - GPT 5.4 mini, OpenAI Pivot, Mamba 3, Attention Residuals
What Happened
OpenAI ships GPT-5.4 mini and nano, faster and more capable but up to 4x pricier, DLSS 5 looks like a real-time generative AI filter for video games | The Verge, and more!
Our Take
honestly? they're throwing around GPT-5.4 mini and Mamba architectures like it's a breakthrough. it's just more fine-tuning, and they're upcharging the whole damn thing by 4x. it's a classic revenue grab disguised as innovation. the attention residuals stuff isn't new, it's just how they're squeezing the cost out of the next generation.
look, the pivot to business and productivity only tells you everything you need to know. they're ditching the massive, unfocused research for actual money. it's smart, but it just means the bleeding edge is getting shelved for quarterly reports.
we're watching a lot of shiny architecture names—Mamba, MoE, etc.—but at the end of the day, it's just optimizing parameters for the next hook. don't get distracted by the bells and whistles; focus on the actual cost per token and deployment.
my take is that this is just optimized hype. it's not magic, it's just better plumbing, and they're selling the plumbing.
What To Do
Ignore the architectural buzzwords and focus solely on the cost per inference and deployment efficiency metrics.
Builder's Brief
What Skeptics Say
GPT-5.4 mini being up to 4x pricier than its predecessor while marketed as affordable signals OpenAI is repricing the floor upward, not lowering it; builders who structured cost models around mini-tier pricing need to renegotiate margins immediately.
Cited By
React
Get the weekly AI digest
The stories that matter, with a builder's perspective. Every Thursday.