Now showing items 1-1 of 1

    • Multi-Teacher Knowledge Distillation Using Teacher's Domain Expertise 

      Hossain, Arafat Bin (North Dakota State University, 2022)
      Large BERT models cannot be used with low computing power and storage capacity. Knowledge Distillation solves this problem by distilling knowledge into a smaller BERT model while retaining much of the teacher’s accuracy ...