
Full-Stack AI Systems Development with Murali Akula - #563
Today we’re joined by Murali Akula, a Sr. director of Software Engineering at Qualcomm. In our conversation with Murali, we explore his role at Qualcomm, where he leads the corporate research team focused on the development and deployment of AI onto Snapdragon chips, their unique definition of “full stack”, and how that philosophy permeates into every step of the software development process. We explore the complexities that are unique to doing machine learning on resource constrained devices, some of the techniques that are being applied to get complex models working on mobile devices, and the process for taking these models from research into real-world applications. We also discuss a few more tools and recent developments, including DONNA for neural architecture search, X-Distill, a method of improving the self-supervised training of monocular depth, and the AI Model Effeciency Toolkit, a library that provides advanced quantization and compression techniques for trained neural network models. The complete show notes for this episode can be found at twimlai.com/go/563
14 Maalis 202244min

100x Improvements in Deep Learning Performance with Sparsity, w/ Subutai Ahmad - #562
Today we’re joined by Subutai Ahmad, VP of research at Numenta. While we’ve had numerous conversations about the biological inspirations of deep learning models with folks working at the intersection of deep learning and neuroscience, we dig into uncharted territory with Subutai. We set the stage by digging into some of fundamental ideas behind Numenta’s research and the present landscape of neuroscience, before exploring our first big topic of the podcast: the cortical column. Cortical columns are a group of neurons in the cortex of the brain which have nearly identical receptive fields; we discuss the behavior of these columns, why they’re a structure worth mimicing computationally, how far along we are in understanding the cortical column, and how these columns relate to neurons. We also discuss what it means for a model to have inherent 3d understanding and for computational models to be inherently sensory motor, and where we are with these lines of research. Finally, we dig into our other big idea, sparsity. We explore the fundamental ideals of sparsity and the differences between sparse and dense networks, and applying sparsity and optimization to drive greater efficiency in current deep learning networks, including transformers and other large language models. The complete show notes for this episode can be found at twimlai.com/go/562
7 Maalis 202250min

Scaling BERT and GPT for Financial Services with Jennifer Glore - #561
Today we’re joined by Jennifer Glore, VP of customer engineering at SambaNova Systems. In our conversation with Jennifer, we discuss how, and why, Sambanova, who is primarily focused on building hardware to support machine learning applications, has built a GPT language model for the financial services industry. Jennifer shares her thoughts on the progress of industries like banking and finance, as well as other traditional organizations, in their attempts at using transformers and other models, and where they’ve begun to see success, as well as some of the hidden challenges that orgs run into that impede their progress. Finally, we explore their experience replicating the GPT-3 paper from a R&D perspective, how they’re addressing issues of predictability, controllability, governance, etc, and much more. The complete show notes for this episode can be found at twimlai.com/go/561
28 Helmi 202244min

Trends in Deep Reinforcement Learning with Kamyar Azizzadenesheli - #560
Today we’re joined by Kamyar Azizzadenesheli, an assistant professor at Purdue University, to close out our AI Rewind 2021 series! In this conversation, we focused on all things deep reinforcement learning, starting with a general overview of the direction of the field, and though it might seem to be slowing, thats just a product of the light being shined constantly on the CV and NLP spaces. We dig into themes like the convergence of RL methodology with both robotics and control theory, as well as a few trends that Kamyar sees over the horizon, such as self-supervised learning approaches in RL. We also talk through Kamyar’s predictions for RL in 2022 and beyond. This was a fun conversation, and I encourage you to look through all the great resources that Kamyar shared on the show notes page at twimlai.com/go/560!
21 Helmi 20221h 17min

Deep Reinforcement Learning at the Edge of the Statistical Precipice with Rishabh Agarwal - #559
Today we’re joined by Rishabh Agarwal, a research scientist at Google Brain in Montreal. In our conversation with Rishabh, we discuss his recent paper Deep Reinforcement Learning at the Edge of the Statistical Precipice, which won an outstanding paper award at the most recent NeurIPS conference. In this paper, Rishabh and his coauthors call for a change in how deep RL performance is reported on benchmarks when using only a few runs, acknowledging that typically, DeepRL algorithms are evaluated by the performance on a large suite of tasks. Using the Atari 100k benchmark, they found substantial disparities in the conclusions from point estimates alone versus statistical analysis. We explore the reception of this paper from the research community, some of the more surprising results, what incentives researchers have to implement these types of changes in self-reporting when publishing, and much more. The complete show notes for this episode can be found at twimlai.com/go/559
14 Helmi 202251min

Designing New Energy Materials with Machine Learning with Rafael Gomez-Bombarelli - #558
Today we’re joined by Rafael Gomez-Bombarelli, an assistant professor in the department of material science and engineering at MIT. In our conversation with Rafa, we explore his goal of fusing machine learning and atomistic simulations for designing materials, a topic he spoke about at the recent SigOpt AI & HPC Summit. We discuss the two ways in which he thinks of material design, virtual screening and inverse design, as well as the unique challenges each technique presents. We also talk through the use of generative models for simulation, the type of training data necessary for these tasks, and if he’s building hand-coded simulations vs existing packages or tools. Finally, we explore the dynamic relationship between simulation and modeling and how the results of one drive the others efforts, and how hyperparameter optimization gets incorporated into the various projects. The complete show notes for this episode can be found at twimlai.com/go/558
7 Helmi 202253min

Differentiable Programming for Oceanography with Patrick Heimbach - #557
Today we’re joined by Patrick Heimbach, a professor at the University of Texas working at the intersection of ML and oceanography. In our conversation with Patrick, we explore some of the challenges of computational oceanography, the potential use cases for machine learning in this field, as well as how it can be used to support scientists in solving simulation problems, and the role of differential programming and how it is expressed in his work. The complete show notes for this episode can be found at twimlai.com/go/557
31 Tammi 202234min

Trends in Machine Learning & Deep Learning with Zachary Lipton - #556
Today we continue our AI Rewind 2021 series joined by a friend of the show, assistant professor at Carnegie Mellon University, and AI Rewind veteran, Zack Lipton! In our conversation with Zack, we touch on recurring themes like “NLP Eating AI” and the recent slowdown in innovation in the field, the redistribution of resources across research problems, and where the opportunities for real breakthroughs lie. We also discuss problems facing the current peer-review system, notable research from last year like the introduction of the WILDS library, and the evolution of problems (and potential solutions) in fairness, bias, and equity. Of course, we explore some of the use cases and application areas that made notable progress in 2021, what Zack is looking forward to in 2022 and beyond, and much more! The complete show notes for this episode can be found at twimlai.com/go/556
27 Tammi 20221h 8min