-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
coco-p1 training divergent #13
Comments
Training with such small amout of supervision is sensitive to hyper-parameters, please try batch 8 and logits weight 3 |
just corrected the config in latest commit |
Thanks for the reply, I will try the latest code |
I just tried the latest config, and I add IMS_PER_DEVICE=1 to avoid the below assert. But the training still diverged after about 40k steps. I got higher result 16% mAP, but it's still much lower than 19.64%. I notice that coco-p1 don't use multiple-scale training, will that influence the final result? |
Indeed... Thanks for correction, I'll fix it. The multi-scale training would affect performance a lot, please use |
Thanks for your reply. I get 18.49 mAP now, but it's still 1 point lower than the score presented in paper(19.64±0.34). Can this fluctuation in the result be considered normal? |
Thanks for your detailed reply! In my situation, the inference is carried out 4 times every 2k iter, not 2 times. Both the teacher and student models are evaluated twice which is bizarre. I didn't modify the code, could you reproduce this problem with the official code? |
I tried to reproduce results under coco-p1 configuration, but training divergent after 40k steps and I got only 14% mAP which is far lower than 19.64%. Could you help me, please
The text was updated successfully, but these errors were encountered: