The server-side rendering equivalent for LLM inference workloads

The server-side rendering equivalent for LLM inference workloads

Ryan is joined by Tuhin Srivastava, CEO and co-founder of Baseten, to explore the evolving landscape of AI infrastructure and inference workloads, how the shift from traditional machine learning models to large-scale neural networks has made GPU usage challenging, and the potential future of hardware-specific optimizations in AI.

Episode notes:

Baseten is an AI infrastructure platform giving you the tooling, expertise, and hardware needed to bring AI products to market fast.

Connect with Tuhin on LinkedIn or reach him at his email tuhin@baseten.co.

Shoutout to user Hitesh for winning a Populist badge for their answer to Cannot drop database because it is currently in use.

See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

Populært innen Business og økonomi

stopp-verden
dine-penger-pengeradet
e24-podden
rss-penger-polser-og-politikk
rss-borsmorgen-okonominyhetene
finansredaksjonen
utbytte
tid-er-penger-en-podcast-med-peter-warren
pengesnakk
pengepodden-2
livet-pa-veien-med-jan-erik-larssen
rss-rettssikkerhet-bak-fasaden-pa-rettsstaten-norge
morgenkaffen-med-finansavisen
rss-sunn-okonomi
stormkast-med-valebrokk-stordalen
aksjepodden
lederpodden
rss-markedspuls-2
rss-fa-makro
okonomiamatorene