Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Easy way to probe result examples? #11

Open
chancharikmitra opened this issue Sep 22, 2023 · 1 comment
Open

Easy way to probe result examples? #11

chancharikmitra opened this issue Sep 22, 2023 · 1 comment

Comments

@chancharikmitra
Copy link

chancharikmitra commented Sep 22, 2023

Hello! This is some really interesting work!

This is more of a question. Do you have any detailed analysis of the results just for InstructBLIP and InstructBLIP Vicuna? It would be great if you maybe had some of these results available for the models I mentioned. That way I could look at just the dataset rather than having to rerun the models on the benchmark. I just wanted to probe the success and failure cases in a bit more detail (seeing the example, model response, etc.).

Thanks!

@Bohao-Lee
Copy link
Collaborator

Thank you for your interest in our work, and we apologize for the delayed response. We have released the GPT-4V evaluation results for SEED-Bench-1 and SEED-Bench-2, which can be found at GPT-4V for SEED-Bench-1 and GPT-4V for SEED-Bench-2. If you're interested, please feel free to explore these results.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants