Rethinking Model Size: Train Large, Then Compress with Joseph Gonzalez - #378

Rethinking Model Size: Train Large, Then Compress with Joseph Gonzalez - #378

Today we’re joined by Joseph Gonzalez, Assistant Professor in the EECS department at UC Berkeley. In our conversation, we explore Joseph’s paper “Train Large, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers,” which looks at compute-efficient training strategies for models. We discuss the two main problems being solved; 1) How can we rapidly iterate on variations in architecture? And 2) If we make models bigger, is it really improving any efficiency?

Populært innen Politikk og nyheter

giver-og-gjengen-vg
aftenpodden
aftenpodden-usa
bt-dokumentar-2
forklart
popradet
stopp-verden
nokon-ma-ga
fotballpodden-2
dine-penger-pengeradet
det-store-bildet
aftenbla-bla
frokostshowet-pa-p5
e24-podden
rss-penger-polser-og-politikk
rss-dannet-uten-piano
rss-ness
rss-fredrik-og-zahid-loser-ingenting
rss-borsmorgen-okonominyhetene
unitedno