2015 Holiday Special
Data Skeptic25 Des 2015

2015 Holiday Special

Today's episode is a reading of Isaac Asimov's The Machine that Won the War. I can't think of a story that's more appropriate for Data Skeptic.

Episoder(589)

Talking to GPT-2

Talking to GPT-2

GPT-2 is yet another in a succession of models like ELMo and BERT which adopt a similar deep learning architecture and train an unsupervised model on a massive text corpus. As we have been covering recently, these approaches are showing tremendous promise, but how close are they to an AGI?  Our guest today, Vazgen Davidyants wondered exactly that, and have conversations with a Chatbot running GPT-2.  We discuss his experiences as well as some novel thoughts on artificial intelligence.

31 Okt 201929min

Reproducing Deep Learning Models

Reproducing Deep Learning Models

Rajiv Shah attempted to reproduce an earthquake-predicting deep learning model.  His results exposed some issues with the model.  Kyle and Rajiv discuss the original paper and Rajiv's analysis.

23 Okt 201922min

What BERT is Not

What BERT is Not

Allyson Ettinger joins us to discuss her work in computational linguistics, specifically in exploring some of the ways in which the popular natural language processing approach BERT has limitations.

14 Okt 201927min

SpanBERT

SpanBERT

Omer Levy joins us to discuss "SpanBERT: Improving Pre-training by Representing and Predicting Spans". https://arxiv.org/abs/1907.10529

8 Okt 201924min

BERT is Shallow

BERT is Shallow

Tim Niven joins us this week to discuss his work exploring the limits of what BERT can do on certain natural language tasks such as adversarial attacks, compositional learning, and systematic learning.

23 Sep 201920min

BERT is Magic

BERT is Magic

Kyle pontificates on how impressed he is with BERT.

16 Sep 201918min

Applied Data Science in Industry

Applied Data Science in Industry

Kyle sits down with Jen Stirrup to inquire about her experiences helping companies deploy data science solutions in a variety of different settings.

6 Sep 201921min

Building the howto100m Video Corpus

Building the howto100m Video Corpus

Video annotation is an expensive and time-consuming process. As a consequence, the available video datasets are useful but small. The availability of machine transcribed explainer videos offers a unique opportunity to rapidly develop a useful, if dirty, corpus of videos that are "self annotating", as hosts explain the actions they are taking on the screen. This episode is a discussion of the HowTo100m dataset - a project which has assembled a video corpus of 136M video clips with captions covering 23k activities. Related Links The paper will be presented at ICCV 2019 @antoine77340 Antoine on Github Antoine's homepage

19 Aug 201922min

Populært innen Vitenskap

fastlegen
rekommandert
jss
tingenes-tilstand
rss-nysgjerrige-norge
sinnsyn
rss-rekommandert
rss-paradigmepodden
dekodet-2
villmarksliv
tomprat-med-gunnar-tjomlid
forskningno
pod-britannia
doktor-fives-podcast
fjellsportpodden
rss-overskuddsliv
vett-og-vitenskap-med-gaute-einevoll
abid-nadia-skyld-og-skam
nordnorsk-historie
rss-fibromyalgi