TY - GEN
T1 - Efficient monocular pose estimation for complex 3D models
AU - Rubio, A.
AU - Villamizar, M.
AU - Ferraz, L.
AU - Penate-Sanchez, A.
AU - Ramisa, A.
AU - Simo-Serra, E.
AU - Sanfeliu, A.
AU - Moreno-Noguer, F.
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2015/6/29
Y1 - 2015/6/29
N2 - We propose a robust and efficient method to estimate the pose of a camera with respect to complex 3D textured models of the environment that can potentially contain more than 100; 000 points. To tackle this problem we follow a top down approach where we combine high-level deep network classifiers with low level geometric approaches to come up with a solution that is fast, robust and accurate. Given an input image, we initially use a pre-trained deep network to compute a rough estimation of the camera pose. This initial estimate constrains the number of 3D model points that can be seen from the camera viewpoint. We then establish 3D-to-2D correspondences between these potentially visible points of the model and the 2D detected image features. Accurate pose estimation is finally obtained from the 2D-to-3D correspondences using a novel PnP algorithm that rejects outliers without the need to use a RANSAC strategy, and which is between 10 and 100 times faster than other methods that use it. Two real experiments dealing with very large and complex 3D models demonstrate the effectiveness of the approach.
AB - We propose a robust and efficient method to estimate the pose of a camera with respect to complex 3D textured models of the environment that can potentially contain more than 100; 000 points. To tackle this problem we follow a top down approach where we combine high-level deep network classifiers with low level geometric approaches to come up with a solution that is fast, robust and accurate. Given an input image, we initially use a pre-trained deep network to compute a rough estimation of the camera pose. This initial estimate constrains the number of 3D model points that can be seen from the camera viewpoint. We then establish 3D-to-2D correspondences between these potentially visible points of the model and the 2D detected image features. Accurate pose estimation is finally obtained from the 2D-to-3D correspondences using a novel PnP algorithm that rejects outliers without the need to use a RANSAC strategy, and which is between 10 and 100 times faster than other methods that use it. Two real experiments dealing with very large and complex 3D models demonstrate the effectiveness of the approach.
UR - http://www.scopus.com/inward/record.url?scp=84938262734&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84938262734&partnerID=8YFLogxK
U2 - 10.1109/ICRA.2015.7139372
DO - 10.1109/ICRA.2015.7139372
M3 - Conference contribution
AN - SCOPUS:84938262734
T3 - Proceedings - IEEE International Conference on Robotics and Automation
SP - 1397
EP - 1402
BT - 2015 IEEE International Conference on Robotics and Automation, ICRA 2015
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2015 IEEE International Conference on Robotics and Automation, ICRA 2015
Y2 - 26 May 2015 through 30 May 2015
ER -