troubles when reproducing #3
Replies: 6 comments 2 replies
-
跑的这么快,几张卡啊大哥。STEPS、 MAX_ITER增大了吗? |
Beta Was this translation helpful? Give feedback.
-
@NingYuanxiang I then doubled |
Beta Was this translation helpful? Give feedback.
-
Thanks for your attention! After seeing your issue, I conducted more experiments to provide some reference results, hoping it can help. Original: Setting1: Setting2: So, the reported results can be stably reproduced by default setting. |
Beta Was this translation helpful? Give feedback.
-
@wjf5203 Thanks for the result log. It seems that reducing batchsize from 2 image/GPU to 1 image/GPU will lose 1.0% mAP (48.62 v.s. 49.64), which is consistent with my results below. Is it has something to do with BN layer when using a smaller batchsize? Will SyncBN or other methods be help? Thanks in advance! |
Beta Was this translation helpful? Give feedback.
-
Hi, Dr.Wu, Does 46.4 in Table 3. means randomly initialize the model in cocopretrain and inference it in 360P. |
Beta Was this translation helpful? Give feedback.
-
How can we see the results of a training, they are not in log.txt? |
Beta Was this translation helpful? Give feedback.
-
Hi, thanks for the wonderful work.
But I had some troubles when trying to reproduce the results with command:
python3 projects/IDOL/train_net.py --config-file projects/IDOL/configs/ytvis19_r50.yaml --num-gpus 8 MODEL.WEIGHTS projects/IDOL/weights/cocopretrain_R50.pth SOLVER.IMS_PER_BATCH 16
The results is 46.96, which is lower than provided results 49.5. I'm using Torch 1.9.0, and batchsize was set to 16 instead of 32.
Is there something I missed? Looking forward to your reply.
Beta Was this translation helpful? Give feedback.
All reactions