104 - Model Distillation, with Victor Sanh and Thomas Wolf
In this episode we talked with Victor Sanh and Th…
31 Minuten
Podcast
Podcaster
Beschreibung
vor 5 Jahren
In this episode we talked with Victor Sanh and Thomas Wolf from
HuggingFace about model distillation, and DistilBERT as one example
of distillation. The idea behind model distillation is compressing
a large model by building a smaller model, with much fewer
parameters, that approximates the output distribution of the
original model, typically for increased efficiency. We discussed
how model distillation was typically done previously, and then
focused on the specifics of DistilBERT, including training
objective, empirical results, ablations etc. We finally discussed
what kinds of information you might lose when doing model
distillation.
HuggingFace about model distillation, and DistilBERT as one example
of distillation. The idea behind model distillation is compressing
a large model by building a smaller model, with much fewer
parameters, that approximates the output distribution of the
original model, typically for increased efficiency. We discussed
how model distillation was typically done previously, and then
focused on the specifics of DistilBERT, including training
objective, empirical results, ablations etc. We finally discussed
what kinds of information you might lose when doing model
distillation.
Weitere Episoden
30 Minuten
vor 2 Jahren
51 Minuten
vor 2 Jahren
45 Minuten
vor 2 Jahren
48 Minuten
vor 2 Jahren
36 Minuten
vor 2 Jahren
In Podcasts werben
Kommentare (0)