Some discussions about DivideMix implementation
See original GitHub issueHi, this is excellent work! I have read the paper and source code a few times in the past two weeks. They are inspiring, thanks for sharing them! I have two questions about your implementation, would you take a look when possible?
The first question is about co-guessing and label refinement in the train function. Is it safer to use net.eval() and net2.eval() in this block, then turn on net.train() before calculating the logits in line 101? I feel both net and net2 are used to prepare some labels in this block, which is just doing the evaluation.
https://github.com/LiJunnan1992/DivideMix/blob/d9d3058fa69a952463b896f84730378cdee6ec39/Train_cifar.py#L62-L67
The second question is about the linear_rampup function. I didn’t understand the reason for multiple lambda_u with the current epoch number current. Could you explain that?
https://github.com/LiJunnan1992/DivideMix/blob/d9d3058fa69a952463b896f84730378cdee6ec39/Train_cifar.py#L192-L194
Thank you very much!
Issue Analytics
- State:
- Created a year ago
- Comments:5 (2 by maintainers)

Top Related StackOverflow Question
The reason is that we observe gaussian distributions for the losses w.r.t correct&wrong labels.
Got it, thanks for taking the time to answer the questions! Thank you very much!