Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MLVU aggregate function is incorrect #534

Open
SCZwangxiao opened this issue Feb 11, 2025 · 2 comments
Open

MLVU aggregate function is incorrect #534

SCZwangxiao opened this issue Feb 11, 2025 · 2 comments

Comments

@SCZwangxiao
Copy link

According to results in their leaderboard, the overall score is calculated by the average of each subtype accuracy, not the sample accuracy.

total_correct = 0
total_answered = 0
for k, v in category2score.items():
total_correct += v["correct"]
total_answered += v["answered"]
eval_logger.info(f"Overall Performance: {100 * total_correct / total_answered if total_answered > 0 else 0 : .1f}%")

@kcz358
Copy link
Collaborator

kcz358 commented Feb 13, 2025

Hi @shuyansy , do you mind take a look whether this logic is correct? Thanks!

@kcz358 kcz358 mentioned this issue Feb 27, 2025
2 tasks
@kcz358
Copy link
Collaborator

kcz358 commented Feb 27, 2025

Hi @SCZwangxiao , this issue should have been fixed in #555

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants