-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Questions on Evaluations #3
Comments
Thanks for your good questions!
|
Thank you for the detailed responses! For the second question, does the provided dataset include the GT 3D edges? |
We have updated the dataset and evaluation code. You can check them. |
Hi! Thanks for the update. I have another few questions and I hope I am not acting silly :)
One hypothesis is that the 2D edges you used is not crisp. This can be observed from the 2D edge maps of PiDiNet and DexiNed. Suppose that PiDiNet and DexiNet are accurate, a question is then how do you know the obtained GT 3D edges are true 3D edges? If I create a crisp 3D edge reconstruction, I expect that my method would have a very low recall because of non-crisp 3D GT edges. |
Hey, for your first question, the CAD models are scaled by dividing each axis by the length of the largest side of the bounding box, resulting in a range within [0, 1]. Additionally, the camera poses during rendering are set in meters by default. Thus, if we generate a 3D CAD model based on the given camera poses and rendered images, it should be measured in meters. For your second question, it is difficult to obtain perfectly clean edge points. We obtained these edge points by projecting the ground-truth dense point clouds onto all 2D image planes and checking if they are located in the edge areas. Due to the blurred ground truth edges, we report recall and precision within a given threshold to reduce the effect caused by the blurred ground truth edge points, which is the issue you mentioned. |
Hi! Thanks for your kind response :) For the second answer, the recall needs false negatives, i.e., 3D GT edges that do not covered by the reconstructed edges. How does false negatives relate to the threshold? Assume that there are no reconstructed 3D edges on the cheek of the bear from the DTU dataset scan 105, but there are a lot of 3D GT edges on the cheek, do you count all the 3D GT edges on the cheek as false negatives? I apologize for asking so many questions. Thanks for your considerations. |
I have a few questions on your evaluations:
For the ABC_NEF dataset, the GT parametric curves have to be shifted and scaled so that they are located at the correct coordinate (see this issue). And since this dataset is a synthetic dataset, the 3D points are unit-less. How do you relate the threshold$\tau$ in millimeters mentioned in your paper to the threshold of the unit-less 3D GT? (In the NEF paper, they set the threshold 0.02, with no unit, as the L2 distance from the reconstructed edge point to the ground truth point.)
For generating the 3D GT edges from the DTU dataset, the supp says that the dense point cloud is projected to the images and then cross-comparing these projections with 2D edges observations to mark 3D edges. Are the "cross-comparing" done manually? If not, what strategies/algorithms did you use? In addition, the supp says, (Section B.3) "To ensure accuracy in the ground-truth edge points, we manually set thresholds for each scan and meticulously remove any floating points." What threshold and floating points are you referring to here?
For the metrics, what are the definitions of "Accuracy" and "Completeness"? The paper says that it follows the NEF and the LiMAP papers for the metrics but I do not find them from both papers. Maybe it's a typo as I find the definitions in the NEAT paper, Section F of supp. Also, do you have an explicit definition of what an "edge direction consistency" is?
In Table 1, are the reported numbers the average of all selected 82 CAD models? Do you use all 50 images per CAD model?
The text was updated successfully, but these errors were encountered: