-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Description
This issue investigates the training of a model using the mean-teacher algorithm. The method is inspired from https://github.com/perone/mean-teacher.
It uses the following steps (copied from the above repo):
- Take a supervised architecture and make a copy of it. Let's call the original model the student and the new one the teacher.
- At each training step, use the same minibatch as inputs to both the student and the teacher but add random augmentation or noise to the inputs separately.
- Add an additional consistency cost between the student and teacher outputs (after softmax).
- Let the optimizer update the student weights normally.
- Let the teacher weights be an exponential moving average (EMA) of the student weights. That is, after each training step, update the teacher weights a little bit toward the student weights.
Metadata
Metadata
Assignees
Labels
No labels