-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LTMU-H reproduction not performing as well as reported? #3
Comments
for context, i'm using the same checkpoints for HOI and STARK as listed in the readme, so I don't know if there's any additional training that needs to be done/if there's another checkpoint that gives the results in the paper |
Hi @vineetparikh, that's strange. I tested the repo multiple times and always had the correct results. No additional training or checkpoint rather than those posted in the README are needed. Maybe something is wrong with frames and annotations? Did you try to run the report method on the precomputed results we provide? |
Yup, I pre-extracted the frames with the same ffmpeg version and visualized them to make sure the annotations looked good (actually had opened another issue at matteo-dunnhofer/TREK-150-toolkit#5 before fixing it). Where could I find the precomputed results? I basically ran this from scratch and got results that way. |
For some reason I can't attach the YAML file for my conda env, so I'll post it as plaintext here but this should be import-able:
|
Any idea as to what's going on? |
I tried again but I still obtain the correct results. The yaml looks good. There might be something wrong with the annotation files. Send me an e-mail to [email protected] and I will share a different version. |
email sent! I'm additionally still confused on why my reproduced results are different from the ones in the link, but I guess we can take this discussion offline and update this thread with results |
i'm also willing to find time and hop on a call to debug! |
I replied to your e-mail. It's a quite busy period time for me, let's try so solve the issue offline first. |
This is the expected behaviour. Thanks for pointing out @relh! |
Thanks @relh for reproducing and confirming it's a setup issue on my end! Will follow up with you on fixing inconsistencies with my setup. (I'll leave this issue open until I figure this out and post the fix below, but will work on this offline: thanks to Matteo for all the help as well!) |
Hi there, thanks so much for the great work and toolkit for future benchmarks!
I'm running the LTMU-H baseline for TREK-150 under the OPE protocol to get some initial understanding of quantitative performance, and I'm finding that SS, NPS, and GSS are significantly lower than what's reported in the paper. I've posted my values below.
I followed the initial guidelines, so my initial thought is that there's something different between my setup and the setup used to run evaluation. Any idea as to what's going on?
The text was updated successfully, but these errors were encountered: