Neural Network Knowledge Distillation

Neural Network Knowledge Distillation

TensorFlow demonstration of the Knowledge Distillation framework to show how soft labels act as regularizers and a neural network’s knowledge can be transfered to a simpler model

comments powered by Disqus