-
Notifications
You must be signed in to change notification settings - Fork 51
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Add distill readme and adjust the documents. (#96)
- Loading branch information
1 parent
8bed522
commit c47e164
Showing
6 changed files
with
59 additions
and
24 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,24 @@ | ||
# Introduction | ||
Distilling the Knowledge in a Neural Network[<sup>1</sup>](#r_1) is a different type of training used to transfer the knowledge from the cumbersome models(teachers) to a small model(student) that is more suitable for deployment. | ||
|
||
EDL Distillation is a large scale and universal solution for knowledge distillation. | ||
|
||
- Decouple the teacher and student models | ||
- They can run in the same or different nodes and transfer knowledge via network even on heterogeneous machines. | ||
Use Distillation on resnet50 as an example: The teachers(Resnet101 for example) can be deployed on P4 GPU cards since they compute forward network generally and the student can be deployed on v100 GPU cards since they need more GPU memory. | ||
|
||
- It's flexible and efficient. | ||
- Teachers and students can be adjusted elastically in training by the resource utilization | ||
- Easier to use and deploy. | ||
- Few lines need to change. | ||
- End to end use. We release the Kubernetes' deployment solution for you. | ||
|
||
# Design | ||
## Architecture | ||
## Student | ||
## Teacher | ||
## Reader | ||
## Balancer | ||
|
||
## Reference | ||
1. <div id="r_1">[Distilling the Knowledge in a Neural Network](https://arxiv.org/pdf/1503.02531.pdf)</div> |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,2 @@ | ||
# Distill experiment on resnet50 | ||
TBD |
File renamed without changes.
File renamed without changes.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,12 @@ | ||
# Purpose | ||
This article illustrates how to run distill demo on Kubernetes cluster or one single machine. | ||
|
||
## On Kubernetes | ||
|
||
We have built the docker images for you and you can start a demo on Kubernetes immediately: | ||
|
||
1. Get the yaml files from: `edl/example/distill/k8s/` | ||
2. Use kubectl to create resources from them, such as `kubectl create -f student.yaml` | ||
|
||
## On a single node | ||
TBD |