Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to evaluate the Open-Vocabulary Segmentation results in Table 2? #5

Open
Glupayy opened this issue Jan 5, 2024 · 5 comments
Open

Comments

@Glupayy
Copy link

Glupayy commented Jan 5, 2024

Hi,

Thank you for sharing your impressive work!

I got confused about Table 2: How are the open vocabulary segmentation metrics calculated?
Also, could you please explain how Osprey outputs the mask to calculate these metrics?

Thanks for your help!

@CircleRadon
Copy link
Owner

Hi, @Glupapa
For open-vocabulary segmentation, all approaches employ ground truth boxes/masks as input to assess regional recognition capability. We leverage the semantic similarity as the matching measurement to calculate these metrics. We will release the codes for performance evaluation.
Actually, the current version of Osprey lacks the capability to generate output masks.

@Glupayy
Copy link
Author

Glupayy commented Jan 5, 2024

Thanks for your prompt response!
I noticed that the metrics used on Cityscapes and ADE20K-150 in Table2 are PQ, AP and mIoU, so I'm curious about how to calculate these metrics if Osprey may not output a mask. Could you please shed some light on this?
Thank you once again for your assistance.

@CircleRadon
Copy link
Owner

@Glupapa The groundtruth masks are used in calculating these metrics.

@yeliudev
Copy link

yeliudev commented Nov 5, 2024

Hi @CircleRadon! I'm having the same question as @Glupayy. How can we calculate the PQ, AP, and mIoU values since Osprey cannot output masks? My guess is that for each sample, the PQ, AP, and mIoU can either be 1 (when the predicted label is correct) or 0 (when the predicted label is wrong), and these 'binary' scores are averaged across all samples to obtain the values in Table2. I was wondering whether this is correct. Thanks!

@yeliudev
Copy link

yeliudev commented Nov 5, 2024

Hi @CircleRadon! I'm having the same question as @Glupayy. How can we calculate the PQ, AP, and mIoU values since Osprey cannot output masks? My guess is that for each sample, the PQ, AP, and mIoU can either be 1 (when the predicted label is correct) or 0 (when the predicted label is wrong), and these 'binary' scores are averaged across all samples to obtain the values in Table2. I was wondering whether this is correct. Thanks!

If this is the case, could you please share more insights about why leveraging such metrics rather than simply computing accuracies? Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants