Jeremy Howard — The Story of fast.ai and Why Python Is Not the Future of ML

Jeremy Howard — The Story of fast.ai and Why Python Is Not the Future of ML

Jeremy Howard is a founding researcher at fast.ai, a research institute dedicated to making Deep Learning more accessible. Previously, he was the CEO and Founder at Enlitic, an advanced machine learning company in San Francisco, California. Howard is a faculty member at Singularity University, where he teaches data science. He is also a Young Global Leader with the World Economic Forum, and spoke at the World Economic Forum Annual Meeting 2014 on "Jobs For The Machines." Howard advised Khosla Ventures as their Data Strategist, identifying the biggest opportunities for investing in data-driven startups and mentoring their portfolio companies to build data-driven businesses. Howard was the founding CEO of two successful Australian startups, FastMail and Optimal Decisions Group. Before that, he spent eight years in management consulting, at McKinsey & Company and AT Kearney. TOPICS COVERED: 0:00 Introduction 0:52 Dad things 2:40 The story of Fast.ai 4:57 How the courses have evolved over time 9:24 Jeremy’s top down approach to teaching 13:02 From Fast.ai the course to Fast.ai the library 15:08 Designing V2 of the library from the ground up 21:44 The ingenious type dispatch system that powers Fast.ai 25:52 Were you able to realize the vision behind v2 of the library 28:05 Is it important to you that Fast.ai is used by everyone in the world, beyond the context of learning 29:37 Real world applications of Fast.ai, including animal husbandry 35:08 Staying ahead of the new developments in the field 38:50 A bias towards learning by doing 40:02 What’s next for Fast.ai 40.35 Python is not the future of Machine Learning 43:58 One underrated aspect of machine learning 45:25 Biggest challenge of machine learning in the real world Follow Jeremy on Twitter: https://twitter.com/jeremyphoward Links: Deep learning R&D & education: http://fast.ai Software: http://docs.fast.ai Book: http://up.fm/book Course: http://course.fast.ai Papers: The business impact of deep learning https://dl.acm.org/doi/10.1145/2487575.2491127 De-identification Methods for Open Health Data https://www.jmir.org/2012/1/e33/ Visit our podcasts homepage for transcripts and more episodes! www.wandb.com/podcast 🔊 Get our podcast on Soundcloud, Apple, and Spotify! YouTube: https://www.youtube.com/c/WeightsBiases Apple Podcasts: https://bit.ly/2WdrUvI Spotify: https://bit.ly/2SqtadF We started Weights and Biases to build tools for Machine Learning practitioners because we care a lot about the impact that Machine Learning can have in the world and we love working in the trenches with the people building these models. One of the most fun things about these building tools has been the conversations with these ML practitioners and learning about the interesting things they’re working on. This process has been so fun that we wanted to open it up to the world in the form of our new podcast called Gradient Dissent. We hope you have as much fun listening to it as we had making it! 👩🏼‍🚀Weights and Biases: We’re always free for academics and open source projects. Email carey@wandb.com with any questions or feature suggestions. - Blog: https://www.wandb.com/articles - Gallery: See what you can create with W&B - https://app.wandb.ai/gallery - Continue the conversation on our slack community - http://bit.ly/wandb-forum 🎙Host: Lukas Biewald - https://twitter.com/l2k 👩🏼‍💻Producer: Lavanya Shukla - https://twitter.com/lavanyaai 📹Editor: Cayla Sharp - http://caylasharp.com/

Avsnitt(131)

Unlocking the Power of Language Models in Enterprise: A Deep Dive with Chris Van Pelt

Unlocking the Power of Language Models in Enterprise: A Deep Dive with Chris Van Pelt

In the premiere episode of Gradient Dissent Business, we're joined by Weights & Biases co-founder Chris Van Pelt for a deep dive into the world of large language models like GPT-3.5 and GPT-4. Chris bridges his expertise as both a tech founder and AI expert, offering key strategies for startups seeking to connect with early users, and for enterprises experimenting with AI. He highlights the melding of AI and traditional web development, sharing his insights on product evolution, leadership, and the power of customer conversations—even for the most introverted founders. He shares how personal development and authentic co-founder relationships enrich business dynamics. Join us for a compelling episode brimming with actionable advice for those looking to innovate with language models, all while managing the inherent complexities. Don't miss Chris Van Pelt's invaluable take on the future of AI in this thought-provoking installment of Gradient Dissent Business.We discuss:0:00 - Intro5:59 - Impactful relationships in Chris's life13:15 - Advice for finding co-founders16:25 - Chris's fascination with challenging problems22:30 - Tech stack for AI labs30:50 - Impactful capabilities of AI models36:24 - How this AI era is different47:36 - Advising large enterprises on language model integration51:18 - Using language models for business intelligence and automation52:13 - Closing thoughts and appreciationThanks for listening to the Gradient Dissent Business podcast, with hosts Lavanya Shukla and Caryn Marooney, brought to you by Weights & Biases. Be sure to click the subscribe button below, to keep your finger on the pulse of this fast-moving space and hear from other amazing guests#OCR #DeepLearning #AI #Modeling #ML

16 Nov 202352min

Providing Greater Access to LLMs with Brandon Duderstadt, Co-Founder and CEO of Nomic AI

Providing Greater Access to LLMs with Brandon Duderstadt, Co-Founder and CEO of Nomic AI

On this episode, we’re joined by Brandon Duderstadt, Co-Founder and CEO of Nomic AI. Both of Nomic AI’s products, Atlas and GPT4All, aim to improve the explainability and accessibility of AI.We discuss:- (0:55) What GPT4All is and its value proposition.- (6:56) The advantages of using smaller LLMs for specific tasks. - (9:42) Brandon’s thoughts on the cost of training LLMs. - (10:50) Details about the current state of fine-tuning LLMs. - (12:20) What quantization is and what it does. - (21:16) What Atlas is and what it allows you to do.- (27:30) Training code models versus language models.- (32:19) Details around evaluating different models.- (38:34) The opportunity for smaller companies to build open-source models. - (42:00) Prompt chaining versus fine-tuning models.Resources mentioned:Brandon Duderstadt - https://www.linkedin.com/in/brandon-duderstadt-a3269112a/Nomic AI - https://www.linkedin.com/company/nomic-ai/Nomic AI Website - https://home.nomic.ai/Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.#OCR #DeepLearning #AI #Modeling #ML

27 Juli 20231h 1min

Exploring PyTorch and Open-Source Communities with Soumith Chintala, VP/Fellow of Meta, Co-Creator of PyTorch

Exploring PyTorch and Open-Source Communities with Soumith Chintala, VP/Fellow of Meta, Co-Creator of PyTorch

On this episode, we’re joined by Soumith Chintala, VP/Fellow of Meta and Co-Creator of PyTorch. Soumith and his colleagues’ open-source framework impacted both the development process and the end-user experience of what would become PyTorch.We discuss:- The history of PyTorch’s development and TensorFlow’s impact on development decisions.- How a symbolic execution model affects the implementation speed of an ML compiler.- The strengths of different programming languages in various development stages.- The importance of customer engagement as a measure of success instead of hard metrics.- Why community-guided innovation offers an effective development roadmap.- How PyTorch’s open-source nature cultivates an efficient development ecosystem.- The role of community building in consolidating assets for more creative innovation.- How to protect community values in an open-source development environment.- The value of an intrinsic organizational motivation structure.- The ongoing debate between open-source and closed-source products, especially as it relates to AI and machine learning.Resources:- Soumith Chintalahttps://www.linkedin.com/in/soumith/- Meta | LinkedInhttps://www.linkedin.com/company/meta/- Meta | Websitehttps://about.meta.com/- Pytorchhttps://pytorch.org/Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.#OCR #DeepLearning #AI #Modeling #ML

13 Juli 20231h 8min

Advanced AI Accelerators and Processors with Andrew Feldman of Cerebras Systems

Advanced AI Accelerators and Processors with Andrew Feldman of Cerebras Systems

On this episode, we’re joined by Andrew Feldman, Founder and CEO of Cerebras Systems. Andrew and the Cerebras team are responsible for building the largest-ever computer chip and the fastest AI-specific processor in the industry.We discuss:- The advantages of using large chips for AI work.- Cerebras Systems’ process for building chips optimized for AI.- Why traditional GPUs aren’t the optimal machines for AI work.- Why efficiently distributing computing resources is a significant challenge for AI work.- How much faster Cerebras Systems’ machines are than other processors on the market.- Reasons why some ML-specific chip companies fail and what Cerebras does differently.- Unique challenges for chip makers and hardware companies.- Cooling and heat-transfer techniques for Cerebras machines.- How Cerebras approaches building chips that will fit the needs of customers for years to come.- Why the strategic vision for what data to collect for ML needs more discussion.Resources:Andrew Feldman - https://www.linkedin.com/in/andrewdfeldman/Cerebras Systems - https://www.linkedin.com/company/cerebras-systems/Cerebras Systems | Website - https://www.cerebras.net/Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.#OCR #DeepLearning #AI #Modeling #ML

22 Juni 20231h

Enabling LLM-Powered Applications with Harrison Chase of LangChain

Enabling LLM-Powered Applications with Harrison Chase of LangChain

On this episode, we’re joined by Harrison Chase, Co-Founder and CEO of LangChain. Harrison and his team at LangChain are on a mission to make the process of creating applications powered by LLMs as easy as possible.We discuss:- What LangChain is and examples of how it works. - Why LangChain has gained so much attention. - When LangChain started and what sparked its growth. - Harrison’s approach to community-building around LangChain. - Real-world use cases for LangChain.- What parts of LangChain Harrison is proud of and which parts can be improved.- Details around evaluating effectiveness in the ML space.- Harrison's opinion on fine-tuning LLMs.- The importance of detailed prompt engineering.- Predictions for the future of LLM providers.Resources:Harrison Chase - https://www.linkedin.com/in/harrison-chase-961287118/LangChain | LinkedIn - https://www.linkedin.com/company/langchain/LangChain | Website - https://docs.langchain.com/docs/Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.#OCR #DeepLearning #AI #Modeling #ML

1 Juni 202351min

Deploying Autonomous Mobile Robots with Jean Marc Alkazzi at idealworks

Deploying Autonomous Mobile Robots with Jean Marc Alkazzi at idealworks

On this episode, we’re joined by Jean Marc Alkazzi, Applied AI at idealworks. Jean focuses his attention on applied AI, leveraging the use of autonomous mobile robots (AMRs) to improve efficiency within factories and more.We discuss:- Use cases for autonomous mobile robots (AMRs) and how to manage a fleet of them. - How AMRs interact with humans working in warehouses.- The challenges of building and deploying autonomous robots.- Computer vision vs. other types of localization technology for robots.- The purpose and types of simulation environments for robotic testing.- The importance of aligning a robotic fleet’s workflow with concrete business objectives.- What the update process looks like for robots.- The importance of avoiding your own biases when developing and testing AMRs.- The challenges associated with troubleshooting ML systems.Resources: Jean Marc Alkazzi - https://www.linkedin.com/in/jeanmarcjeanazzi/idealworks |LinkedIn - https://www.linkedin.com/company/idealworks-gmbh/idealworks | Website - https://idealworks.com/Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.#OCR #DeepLearning #AI #Modeling #ML

18 Maj 202358min

How EleutherAI Trains and Releases LLMs: Interview with Stella Biderman

How EleutherAI Trains and Releases LLMs: Interview with Stella Biderman

On this episode, we’re joined by Stella Biderman, Executive Director at EleutherAI and Lead Scientist - Mathematician at Booz Allen Hamilton.EleutherAI is a grassroots collective that enables open-source AI research and focuses on the development and interpretability of large language models (LLMs).We discuss:- How EleutherAI got its start and where it's headed.- The similarities and differences between various LLMs.- How to decide which model to use for your desired outcome.- The benefits and challenges of reinforcement learning from human feedback.- Details around pre-training and fine-tuning LLMs.- Which types of GPUs are best when training LLMs.- What separates EleutherAI from other companies training LLMs.- Details around mechanistic interpretability.- Why understanding what and how LLMs memorize is important.- The importance of giving researchers and the public access to LLMs.Stella Biderman - https://www.linkedin.com/in/stellabiderman/EleutherAI - https://www.linkedin.com/company/eleutherai/Resources:- https://www.eleuther.ai/Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.#OCR #DeepLearning #AI #Modeling #ML

4 Maj 202357min

Scaling LLMs and Accelerating Adoption with Aidan Gomez at Cohere

Scaling LLMs and Accelerating Adoption with Aidan Gomez at Cohere

On this episode, we’re joined by Aidan Gomez, Co-Founder and CEO at Cohere. Cohere develops and releases a range of innovative AI-powered tools and solutions for a variety of NLP use cases.We discuss:- What “attention” means in the context of ML.- Aidan’s role in the “Attention Is All You Need” paper.- What state-space models (SSMs) are, and how they could be an alternative to transformers. - What it means for an ML architecture to saturate compute.- Details around data constraints for when LLMs scale.- Challenges of measuring LLM performance.- How Cohere is positioned within the LLM development space.- Insights around scaling down an LLM into a more domain-specific one.- Concerns around synthetic content and AI changing public discourse.- The importance of raising money at healthy milestones for AI development.Aidan Gomez - https://www.linkedin.com/in/aidangomez/Cohere - https://www.linkedin.com/company/cohere-ai/Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.Resources:- https://cohere.ai/- “Attention Is All You Need”#OCR #DeepLearning #AI #Modeling #ML

20 Apr 202351min

Populärt inom Business & ekonomi

badfluence
framgangspodden
varvet
rss-svart-marknad
uppgang-och-fall
rss-borsens-finest
lastbilspodden
rss-jossan-nina
rss-kort-lang-analyspodden-fran-di
affarsvarlden
24fragor
rss-inga-dumma-fragor-om-pengar
rss-en-rik-historia
rss-dagen-med-di
avanzapodden
borsmorgon
tabberaset
fill-or-kill
bathina-en-podcast
dynastin