Listen "The server-side rendering equivalent for LLM inference workloads"
Episode Synopsis
Ryan is joined by Tuhin Srivastava, CEO and co-founder of Baseten, to explore the evolving landscape of AI infrastructure and inference workloads, how the shift from traditional machine learning models to large-scale neural networks has made GPU usage challenging, and the potential future of hardware-specific optimizations in AI. Episode notes:Baseten is an AI infrastructure platform giving you the tooling, expertise, and hardware needed to bring AI products to market fast.Connect with Tuhin on LinkedIn or reach him at his email [email protected]. Shoutout to user Hitesh for winning a Populist badge for their answer to Cannot drop database because it is currently in use. See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.
More episodes of the podcast The Stack Overflow Podcast
How AWS re:Invented the cloud
16/01/2026
Vibe code anything in a Hanselminute
13/01/2026
Every ecommerce hero needs a Sidekick
09/01/2026
The most dangerous shortcuts in software
02/01/2026
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.