You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi there, just arriving at kinfu_remake Party from learning (late) about Kinect Fusion-> PCL kinFu-> here
As I understand these processes rely on RGB feature detection and ICP for pose estimation. is that correct? I'm very interested in the overall approach used to reconstruct a fused meshes in realitime, but (have; and) would like to use an external tracking system to provide an absolute device pose to this process.
Is doing this feasible? Or is the code too tightly written to consider providing a (known) submillimeter accurate pose truth.
The text was updated successfully, but these errors were encountered:
Hi there, just arriving at kinfu_remake Party from learning (late) about Kinect Fusion-> PCL kinFu-> here
As I understand these processes rely on RGB feature detection and ICP for pose estimation. is that correct? I'm very interested in the overall approach used to reconstruct a fused meshes in realitime, but (have; and) would like to use an external tracking system to provide an absolute device pose to this process.
Is doing this feasible? Or is the code too tightly written to consider providing a (known) submillimeter accurate pose truth.
The text was updated successfully, but these errors were encountered: