Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

A strange bug #27

Open
LiJunscs opened this issue Dec 30, 2024 · 1 comment
Open

A strange bug #27

LiJunscs opened this issue Dec 30, 2024 · 1 comment

Comments

@LiJunscs
Copy link

LiJunscs commented Dec 30, 2024

When I use the lmms-eval and videomme dataset to eval the videoxl, I want to implement an function that is interrupted to save the evaluation result. So I create a debug dataset which includes the first 12 row of videomme, including 4 videos and 12 questions.

The bug is as following,
1. When I use the whole 12 data, the first 3 questions' answer are C, A, C respectively.
2. When I use the first 3 questions, answers are A, A, A.
3. When I use the first question, the answer is A.
4. When I continue the eval with the result cached, the answers of question 0 and 2 are A, C

The result cached when interrupted is correct, and load successfully. Note that only the serializable parts of Instance(lmms-eval.api.instance) are cached, others are set to None. And if the eval finished successfully, the cached function will not be call, whole process is same as origin.

Is this my problem or videoxl or lmms-eval.

I eval the qwen2-vl of lmms-eval, answers of the first 3 questions are always A, A, A.

@shuyansy
Copy link
Collaborator

Sorry to give a late reply. I suppose this is because the random compression ratio is set as default settings. To this end, you can use self.model.config.beacon_ratio=[8] in the py file to achieve fix compression

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants