AI Model Distillation Technique

From GM-RKB
(Redirected from model distillation)
Jump to navigation Jump to search

An AI Model Distillation Technique is a machine learning technique that transfers model knowledge from a larger ML teacher model to a smaller ML student model (to create efficient ML models while preserving ML model performance).



References

2024

  • (Wikipedia, 2024) ⇒ https://en.wikipedia.org/wiki/Knowledge_distillation Retrieved:2024-9-6.
    • In machine learning, knowledge distillation or model distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized. It can be just as computationally expensive to evaluate a model even if it utilizes little of its knowledge capacity. Knowledge distillation transfers knowledge from a large model to a smaller model without loss of validity. As smaller models are less expensive to evaluate, they can be deployed on less powerful hardware (such as a mobile device).

      Knowledge distillation has been successfully used in several applications of machine learning such as object detection, acoustic models, and natural language processing. Recently, it has also been introduced to graph neural networks applicable to non-grid data.


2019