Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

question about results #6

Open
Ha0Tang opened this issue May 3, 2020 · 3 comments
Open

question about results #6

Ha0Tang opened this issue May 3, 2020 · 3 comments

Comments

@Ha0Tang
Copy link

Ha0Tang commented May 3, 2020

Hi, why your results in Table 2 (cityspaces and ade) different from those from SPADE paper while you used the same dataset train/test splits?

@Seanseattle
Copy link
Owner

We used the pretrained models from SPADE Github and tested with the dataset's evaluation method. This means we ignored the "don't care label". For FID, we used the updated PyTorch-version FID, which got the same results as the official Tensorflow implementation. While the SPADE might use the older version. So some results may be different.

@Ha0Tang
Copy link
Author

Ha0Tang commented May 4, 2020

Results of SPADE paper on Cityscapes are 62.3 mIoU and 81.9 acc, but you reported 62.3 mIoU and 93.5 acc, why is mIoU the same, but acc is so different?

Results of SPADE paper on ADE20K are 38.5 mIoU and 79.9 acc, but you reported 42.0 mIoU and 81.4 acc, why both are so different?

@jessemelpolio
Copy link

jessemelpolio commented May 4, 2020

We cannot use the method mentioned in the SPADE paper to get the same results as the ones provided in their paper. It seems that we are not the only one who met this problem. Please see NVlabs/SPADE#39 and NVlabs/SPADE#100 and xh-liu/CC-FPSE#4.

It remains an unsolved problem and people are continuously getting different results from each other. In our case, we use the model (https://drive.google.com/file/d/12gvlTbMvUcJewQlSEaZdeb2CdOB-b8kQ/view) provided by SPADE to test the segmentation results on Cityscapes but we follow some users (xh-liu/CC-FPSE#4 (comment)) to ignore the ‘don’t care’ class. We believe ignoring such class is the main reason that our tested results are better, as also pointed in these issues. It could totally be a coincidence that the mIoU keeps the same to 62.3 as appeared in our testing and that in the SPADE paper. Our evaluation setting is applied to all models in Tab. 2, which at least makes the comparison rather fair in our end. If you could find how to solve this problem, we are very glad to hear it and give it a go and update our results if necessary.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants