Mamba, Mamba-2 and Post-Transformer Architectures for Generative AI with Albert Gu - #693

Mamba, Mamba-2 and Post-Transformer Architectures for Generative AI with Albert Gu - #693

Today, we're joined by Albert Gu, assistant professor at Carnegie Mellon University, to discuss his research on post-transformer architectures for multi-modal foundation models, with a focus on state-space models in general and Albert’s recent Mamba and Mamba-2 papers in particular. We dig into the efficiency of the attention mechanism and its limitations in handling high-resolution perceptual modalities, and the strengths and weaknesses of transformer architectures relative to alternatives for various tasks. We dig into the role of tokenization and patching in transformer pipelines, emphasizing how abstraction and semantic relationships between tokens underpin the model's effectiveness, and explore how this relates to the debate between handcrafted pipelines versus end-to-end architectures in machine learning. Additionally, we touch on the evolving landscape of hybrid models which incorporate elements of attention and state, the significance of state update mechanisms in model adaptability and learning efficiency, and the contribution and adoption of state-space models like Mamba and Mamba-2 in academia and industry. Lastly, Albert shares his vision for advancing foundation models across diverse modalities and applications. The complete show notes for this episode can be found at https://twimlai.com/go/693.

Episoder(779)

Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski - #673

Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski - #673

Today we’re joined by Ben Prystawski, a PhD student in the Department of Psychology at Stanford University working at the intersection of cognitive science and machine learning. Our conversation cente...

26 Feb 202425min

Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh - #672

Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh - #672

Today we're joined by Armineh Nourbakhsh of JP Morgan AI Research to discuss the development and capabilities of DocLLM, a layout-aware large language model for multimodal document understanding. Armi...

19 Feb 202445min

Are Emergent Behaviors in LLMs an Illusion? with Sanmi Koyejo - #671

Are Emergent Behaviors in LLMs an Illusion? with Sanmi Koyejo - #671

Today we’re joined by Sanmi Koyejo, assistant professor at Stanford University, to continue our NeurIPS 2024 series. In our conversation, Sanmi discusses his two recent award-winning papers. First, we...

12 Feb 20241h 5min

AI Trends 2024: Reinforcement Learning in the Age of LLMs with Kamyar Azizzadenesheli - #670

AI Trends 2024: Reinforcement Learning in the Age of LLMs with Kamyar Azizzadenesheli - #670

Today we’re joined by Kamyar Azizzadenesheli, a staff researcher at Nvidia, to continue our AI Trends 2024 series. In our conversation, Kamyar updates us on the latest developments in reinforcement le...

5 Feb 20241h 10min

Building and Deploying Real-World RAG Applications with Ram Sriharsha - #669

Building and Deploying Real-World RAG Applications with Ram Sriharsha - #669

Today we’re joined by Ram Sriharsha, VP of engineering at Pinecone. In our conversation, we dive into the topic of vector databases and retrieval augmented generation (RAG). We explore the trade-offs ...

29 Jan 202435min

Nightshade: Data Poisoning to Fight Generative AI with Ben Zhao - #668

Nightshade: Data Poisoning to Fight Generative AI with Ben Zhao - #668

Today we’re joined by Ben Zhao, a Neubauer professor of computer science at the University of Chicago. In our conversation, we explore his research at the intersection of security and generative AI. W...

22 Jan 202439min

Learning Transformer Programs with Dan Friedman - #667

Learning Transformer Programs with Dan Friedman - #667

Today, we continue our NeurIPS series with Dan Friedman, a PhD student in the Princeton NLP group. In our conversation, we explore his research on mechanistic interpretability for transformer models, ...

15 Jan 202438min

AI Trends 2024: Machine Learning & Deep Learning with Thomas Dietterich - #666

AI Trends 2024: Machine Learning & Deep Learning with Thomas Dietterich - #666

Today we continue our AI Trends 2024 series with a conversation with Thomas Dietterich, distinguished professor emeritus at Oregon State University. As you might expect, Large Language Models figured ...

8 Jan 20241h 5min

Populært innen Politikk og nyheter

giver-og-gjengen-vg
aftenpodden
aftenpodden-usa
i-retten
forklart
stopp-verden
popradet
fotballpodden-2
det-store-bildet
dine-penger-pengeradet
rss-gukild-johaug
bt-dokumentar-2
hanna-de-heldige
aftenbla-bla
nokon-ma-ga
rss-ness
e24-podden
rss-penger-polser-og-politikk
frokostshowet-pa-p5
chit-chat-med-helle