-
Notifications
You must be signed in to change notification settings - Fork 37
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
test_on_tap.py results don't match expected results. #13
Comments
Attached below is the output:
|
In addition, I wanted to ask about the following concerns regarding this testing of TAP-Vid DAVIS:
|
Thanks for the messages. For d_avg: How did you compute 72.376? It looks like it's showing 70.5 in the snippet you posted. The d_x shown in each row is the running average, so 70.5 is the average across the 30 videos. For resolution: I think some papers use 256x256 at test time, but we find that higher-resolution input helps performance, if you can afford it. The stats are still computed at 256x256 though. |
@aharley I'm getting similar results (~70 d_x). The PointOdyssey paper reported about ~63 on this metric. I'm also getting ~7 on the MTE metric, while the paper reported ~4. My results for survival are in line with the paper. I was curious if you had changed or otherwise improved the reference model from the paper, or if there is a bug somewhere? |
Hello,
When running
test_on_tap.py
, I get different results than reported in the testing section.The mean
d_avg
of all 30 videos (output is added below) is72.376
, compared tod_avg 70.6; survival_16 89.3; median_l2 6.9
reported.I download the reference mode using
sh get_reference_model.sh
, and I test ontapvid_davis.pkl
which I downloaded and unzipped fromhttps://storage.googleapis.com/dm-tapnet/tapvid_davis.zip
.I would really appreciate any assistance and clarifications on the matter!
Assaf
The text was updated successfully, but these errors were encountered: