Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

I can not reproduce the result JMMMU evaluation result of SakanaAI/Llama-3-EvoVLM-JP-v2 #449

Open
CHENSSR opened this issue Dec 9, 2024 · 1 comment

Comments

@CHENSSR
Copy link

CHENSSR commented Dec 9, 2024

May I ask what generation parameters you used for JMMMU evaluation result of SakanaAI/Llama-3-EvoVLM-JP-v2 except from 'A maximum output length is set to 1,024 and a temperature is set to 0 for all models during inference.'? The best accuracy I can get so far has only 28.48%. Thank you!

Groups Version Filter n-shot Metric Value Stderr
jmmmu_all N/A none jmmmu_acc 0.2848 ± N/A
- culture_agnostic N/A none jmmmu_acc 0.2889 ± N/A
- culture_specific N/A none jmmmu_acc 0.2800 ± N/A
@kcz358
Copy link
Collaborator

kcz358 commented Dec 25, 2024

I think for this you might need to contact the author of the JMMMU or EvoVLM-JP-v2 and see if they have some advise on this

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants