-
-
Notifications
You must be signed in to change notification settings - Fork 391
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Potential misleading parts in the paper and code #161
Comments
Hello @PointsCoder , Thank you for raising this issue. In general, I personally believe that research is a mixup of successes and mistakes in the exploration of novel things, particularly for new areas lacking handy or mature tools and platforms. While we acknowledge the possibility of flaws in our projects, I assure you that we never intentionally make mistakes, and we prioritize academic integrity. That’s why we choose open-source almost all our projects. We greatly appreciate the efforts of researchers in identifying any shortcomings and working towards building more robust and comprehensive platforms for future use. For your questions,
Misc: We are not the first group to try open-loop planning on nuScenes. You can find pioneering results on FF [a], EE [b], etc. — |
Hi, Thanks for your response. I understand that researchers may make mistakes sometimes, and I believe your motivation behind those unintentional mistakes. I raised this issue not to criticize your work (your work is great in my opinion). My request is that you need to clarify the above-mentioned misleading parts by updating your paper. After reading the response, I believe that we all agree with the fact that the following misleading parts exist in the paper:
No matter the intentions behind (I generally believe your good intentions), I believe the above-mentioned misleading parts should be fixed by updating your paper (at least the arXiv version), for the following reasons:
Here are my suggestions for the paper revision:
Again, thanks for your contribution to the community. I am not judging or criticizing your work, but I believe these misleading parts should be clarified in the paper. |
For the use of historical ego-information in motion and planning head, from my understanding, you have still used sdc_embedding in your track query here, and this will still cause leakage of sdc_tracks (ego-history) to planning. Also, ego-forecasting in motion head is unnatural as you planned twice in your system. Let me know if my understanding is correct. |
The tracking status is estimated during testing. Using predicted history for planning is not a leakage IMO. Ego-forecasting is a common practice when employing motion forecasting methods for planning, as can be seen in works in motion fields and nuPlan challenges. As I said, I appreciate the following works that have initiated the discussions and presented additional results. For your requests and suggestions, I will discuss with the team.
|
Thanks for your reply.
Overall, I appreciate your feedback, and the discussion makes things clearer. Good luck to your team and hope to see more influential works from your group : ) |
Hello,
Thanks for your work and contribution to the community.
I noticed that your paper may have information that misleads the researchers who have followed your work.
As shown in [1][2][3], ego status and historical trajectory have short-cut effects on planning. From the method diagram, this paper doesn't show they have included such information. However, after carefully checking the provided code, this paper indeed includes both ego status and ego historical trajectory in the implementation:
Incorporating ego status and historical trajectory explicitly and implicitly will have huge effects on the motion planning performance. Removing 1 can cause a significant performance drop, and I believe 2 and 3 also play a significant role in boosting the planning performance. I am curious what will happen if removing all of them. I would suggest the authors should notify readers where they have included ego information into the system in this paper, and how it would affect the final performance.
This paper compared with ST-P3 in Table 7. However, after carefully checking the code of both ST-P3 and this repo, I found the evaluation metrics of ST-P3 and this repo are implemented differently, and this will make the comparison not fair (I am actually very curious why you chose different implementations since both works are from your group).
Specifically, I summarized the differences below:
Given the differences, the comparison in Table 7 looks very tricky and unfair.
Tables 2, 7, and 10 have different reported planning scores for the same model, but in which condition you are testing your model is not mentioned. The differences in experimental settings should be highlighted in the paper.
Overall, your group may be the first to try open-loop planning on nuScenes. It's possible that these unintentional mistakes and misleading parts will cause a lot of trouble for researchers to faithfully follow your work. I believe the use of ego and historical information in your system and the different implementations in metrics should be highlighted in the paper to avoid misleading other researchers.
Thank you again for your work and efforts. Looking forward to your response.
[1] Codevilla et al. Exploring the Limitations of Behavior Cloning for Autonomous Driving.
[2] Dauner et al. Parting with Misconceptions about Learning-based Vehicle Motion Planning.
[3] Zhai et al. Rethinking the Open-Loop Evaluation of End-to-End Autonomous Driving in nuScenes.
[4] Li et al. Is Ego Status All You Need for Open-Loop End-to-End Autonomous Driving?
The text was updated successfully, but these errors were encountered: