-
Notifications
You must be signed in to change notification settings - Fork 14
Open
Description
Dear all,
I am trying to replicate the results of VoxFormer on your KITTI360 benchmark by using your checkpoint. So far I have achieved sensible results, but with an occupancy IoU of about 34.5 instead of your 38.6. Here is how I proceeded to generate the needed query proposals for stage 2:
- Predict the depth from matched stereo images using the provided script with
--dataset kitti360 --baseline 331.53255659999996
.(the baseline number was indicated in the script. Outputs look sensible) - Estimate LIDAR from the depth maps using the provided script.
- Accumulate the LIDAR scans using your script a sequence length 10 (tried 1 alternatively with similar results). KITTI360 poses get inverted during loading (and matched to the correct image). Again the results look like a sensible accumulation of depth maps/LIDAR scans.
- Use Stage1 with the provided checkpoint to predict the query proposals.
- Use them with the images themselves as input for stage2 to get our final predictions (loading your stage2 checkpoint).
Note: As stereo images were not provided in SSCBench and the pose files in the dataset also do not necessarily match the frame Ids provided, I created a mapping of the SSCBench frame Ids to the KITTI360 frame ids.
Is my process correct? Why might my results be worse? Could you either provide detailed instructions on how to replicate your results or the predictions?
Kind regards,
Adrian
Metadata
Metadata
Assignees
Labels
No labels