-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Could you early release the evaluation scripts with vicuna model. #1
Comments
+1 |
@KerolosAtef @avinash31d , Thank you for your interest in our work. Please find the details about the Vicuna-based quantitative evaluation benchmark here: https://github.com/mbzuai-oryx/Video-LLaVA/tree/main/quantitative_evaluation. |
thank you very much, but also the Vicuna model doesn't output the same results for each run. I have tried to reproduce some of the results of video chat GPT and this the results: |
@KerolosAtef We attribute this to the randomness introduced by the temperature parameter in both the tested model and the LLM used for evaluation. This will be addressed in our future work. |
okay good, |
We follow the same approach as Video-ChatGPT, i.e. using test splits. |
No description provided.
The text was updated successfully, but these errors were encountered: