You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
May I ask what generation parameters you used for JMMMU evaluation result of SakanaAI/Llama-3-EvoVLM-JP-v2 except from 'A maximum output length is set to 1,024 and a temperature is set to 0 for all models during inference.'? The best accuracy I can get so far has only 28.48%. Thank you!
Groups
Version
Filter
n-shot
Metric
Value
Stderr
jmmmu_all
N/A
none
jmmmu_acc
↑
0.2848
±
N/A
- culture_agnostic
N/A
none
jmmmu_acc
↑
0.2889
±
N/A
- culture_specific
N/A
none
jmmmu_acc
↑
0.2800
±
N/A
The text was updated successfully, but these errors were encountered:
May I ask what generation parameters you used for JMMMU evaluation result of SakanaAI/Llama-3-EvoVLM-JP-v2 except from 'A maximum output length is set to 1,024 and a temperature is set to 0 for all models during inference.'? The best accuracy I can get so far has only 28.48%. Thank you!
The text was updated successfully, but these errors were encountered: