You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi!, thanks for your good project to the lidar motion. But I have a question about the data augmention. I notice that the "input_points" are augmented with global flip and scaling, but the "source_points" and "target_points" do not do this augmetation. Maybe the predicted motion can not transform "source_points" to "target_points" to compute precise chamfer loss. Is this a bug?
The text was updated successfully, but these errors were encountered:
Thanks for pointing it out. In the original implementation, all the points are concatenated together and sliced afterward. We are still verifying the preprocessing part after refactorizing.
Hi!, thanks for your good project to the lidar motion. But I have a question about the data augmention. I notice that the "input_points" are augmented with global flip and scaling, but the "source_points" and "target_points" do not do this augmetation. Maybe the predicted motion can not transform "source_points" to "target_points" to compute precise chamfer loss. Is this a bug?
The text was updated successfully, but these errors were encountered: