Browsing by Subject "Multi Teacher Knowledge Distillation"
Now showing items 1-1 of 1
-
Multi-Teacher Knowledge Distillation Using Teacher's Domain Expertise
(North Dakota State University, 2022)Large BERT models cannot be used with low computing power and storage capacity. Knowledge Distillation solves this problem by distilling knowledge into a smaller BERT model while retaining much of the teacher’s accuracy ...