-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Great work! a few questions for the sake of reproducibility #6
Comments
Thanks for the compliments!
Actually, may I know if the released checkpoints work fine? I tidied up the code a bit which I think might break the checkpoint reading, but I never get a chance to try loading it.
|
If you want to use the BEV features to enhance other vehicles' predictions, you probably need to make some changes to the encoding mechanism. Currently it uses the center (where the ego vehicle is located at) as the query (IIRC, not quite remember anymore....), but if you want to use it for enhancements for other vehicle's predictions, you should use their corresponding BEV patches as the query. |
Thank you for your prompt response! The released checkpoints work fine if i set I was attempting to retrain your HiVT models to reproduce your results, using the pretrained map models to generate bev_features from scratch for both the training and val sets. Then, using the optimal hparams you specified in Appendix A, I was able to train the model to get a reasonable number but still pretty far from your reported results (0.391 vs 0.365 ADE you reported on Maptrv2_CL + bev).
finally, regarding your response about using the BEV features to enhance other vehicles' predictions: I believe the query is already set to use the local patches corresponding to each individual agent (here! :) ) Thank you so much! |
Hi Erica, Sorry for the late reply and thanks for checking the checkpoint for me. Was a bit busy these couple days.
Also one difference between AWS and HuggingFace version is that I tidied up the dimension of the bev features in huggingface ones. In AWS they are
Uhhh I see, thanks for checking. Yeah sorry it's been a while.... and I am not familiar with my code anymore :( |
Uhmm to try retraining it to get the checkpoint, maybe you can use the uploaded dataset and see try retraining from there? To check whether the BEV features make sense, you can try visualizing them (especially the streammapnet ones, their BEV features look pretty aligned with the road structure) using the vis script. The reason why this is crucial is that in MapTR lines of work, they just do simply reshaping to bring the BEV features from (height, width, feature_dim) to (height*width, feature_dim) and then do their map decoding. In StreamMapNet, they used a flipped view of the map BEV features and then do the map decoding (which is quite weird tbh, but in their repo issues I remember they say this is to be compatible with a framework used in their approach, maybe detr or something?), so they are more like an image where the (0,0) starts at the top left corner. So in my file here, I need to do some flipping of the BEV features. |
Let me know if you have any other questions and I am very happy to help. I think this line of work still has a lot of rooms for improvements but unfortunately I switched areas for my master research and do not have time to work on this area anymore... And hopefully you can build a cleaner codebase out of this haha. |
Thank you so much for your response, and for uploading some of the data to the huggingface repo! I was still getting AccessDenied error with aws, but it seems downloading from huggingface works fine. By any chance, are you able to also upload the other datasets (maptrv2_bev, maptrv2_cent_bev) to the huggingface repo as well, as those are the ones I'm interested in? |
Of course! I have just uploaded maptrv2_bev as well. maptrv2_cent_bev should also be done tmr. It takes a while to upload lol. Thanks for your patience. |
I have uploaded them all. Thanks for the support! |
Thank you for uploading the dataset! I just want to report on my attempt to reproduce your results on HiVT so far. I am using your HiVT_modified code (downloaded a fresh copy from scratch), and your provided huggingface dataset with 15113 train sequences and 4519 val sequences on nuscenes. the only difference between my setup and yours is my environment: because I had difficulty installing the environment according to the HiVT instuctions on my machine (equipped with nvidia A100, cuda 12.5), so I used torch 1.9.1 (instead of torch 1.8.0) but the same version of lightning and torch-geometric. however, I not able to reproduce your results after retraining HiVT from scratch with bev features (maptrv2_cent_bev). I am getting about the same performance as when I trained on my own generated data (using the pretrained mapping models). When using lr 3.5e-4 and weight_decay of 1e-2, which is the best hparams that you report in Appendix A of your paper, I am getting a minADE of 0.395 (in your paper you report 0.365; evaluating your provided pretrained model I am getting 0.369). Using lr 5e-4 and weight_decay of 1e-3, I am getting a little better performance @ 0.393. As this is almost a 10% worse performance than your reported results, I wanted to see where I may be going wrong (and to document reproducibility attempts to the community).
Thank you for your help. |
Hey Erica, thanks for letting me know your progress!
For StreamMapNet, because they use a custom split for training (to avoid overlapping areas between train and val), they have fewer samples than the MapTR series. But I do not recall what causes the difference between MapTRv2_cent and the others. I suspect one potential cause is that One suggestion I have is to maybe look at the frames that are skipped, specifically the places where the As for this question: "Did you make any changes to the indexing within the adaptor files?" I don't think I did. The indexing files I extracted from TrajData are consistent throughout—I only performed this extraction once at the start of the two projects. So, I believe it is more of an issue with the adaptor writing (maybe I changed some code between MapTR and MapTR_cent processing) rather than an issue with the indexing files.
Maybe I can also try reproducing the results on my side, but I need a bit of time to set up the environment and code again which might take some time... Thank you so much for your patience! |
-Thanks for the great work and the very thorough documentation. It has made using your code and building off your code very easy! I appreciate it a lot :)
I had a few questions, just to make sure I'm interpreting and reproducing your results accurately.
The text was updated successfully, but these errors were encountered: