Closed
Description
Hi,
Thank you for sharing this awesome code!
Base on this issue, I understand that you are not going to release the evaluation code, and I'm working on reimplementing them myself.
I have the following questions:
-
When computing the FID scores, do you compare to the generated images the original images or the cropped images (the same size as the generated ones)?
-
What are the image sizes you used for evaluation? Do you generate higher resolution ones for evaluation or just use the default size (512x256 for cityscape, and 256x256 for the others)?
-
What are the pre-trained segmentation models and code base you use for each datasets? Based on the paper, I assume these are the ones you use. Could you please confirm them?
- COCO stuff: code: kazuto1011/deeplab-pytorch model:
deeplabv2_resnet101_msc-cocostuff164k-100000.pth - ADE20K: code: CSAILVision/semantic-segmentation-pytorch model: baseline-resnet101-upernet
- Cityscapes: code: fyu/drn model: drn-d-105_ms_cityscapes.pth
- When you evaluate mIoUs and accuracies, do you upsample the images or downsample the labels? If so, how do you interpolate them?
Thanks in advance.
Best,
Godo
Metadata
Metadata
Assignees
Labels
No labels