TY - GEN
T1 - Learning to reconstruct 3D manhattan wireframes from a single image
AU - Zhou, Yichao
AU - Qi, Haozhi
AU - Zhai, Yuexiang
AU - Sun, Qi
AU - Chen, Zhili
AU - Wei, Li Yi
AU - Ma, Yi
N1 - Funding Information:
This work is partially supported by Sony US Research Center, Adobe Research, Berkeley BAIR, and Bytedance Research Lab.
Publisher Copyright:
© 2019 IEEE.
PY - 2019/10
Y1 - 2019/10
N2 - From a single view of an urban environment, we propose a method to effectively exploit the global structural regularities for obtaining a compact, accurate, and intuitive 3D wireframe representation. Our method trains a single convolutional neural network to simultaneously detect salient junctions and straight lines, as well as predict their 3D depth and vanishing points. Compared with state-of-the-art learning-based wireframe detection methods, our network is much simpler and more unified, leading to better 2D wireframe detection. With a global structural prior (such as Manhattan assumption), our method further reconstructs a full 3D wireframe model, a compact vector representation suitable for a variety of high-level vision tasks such as AR and CAD. We conduct extensive evaluations of our method on a large new synthetic dataset of urban scenes as well as real images. Our code and datasets will be published along with the paper.
AB - From a single view of an urban environment, we propose a method to effectively exploit the global structural regularities for obtaining a compact, accurate, and intuitive 3D wireframe representation. Our method trains a single convolutional neural network to simultaneously detect salient junctions and straight lines, as well as predict their 3D depth and vanishing points. Compared with state-of-the-art learning-based wireframe detection methods, our network is much simpler and more unified, leading to better 2D wireframe detection. With a global structural prior (such as Manhattan assumption), our method further reconstructs a full 3D wireframe model, a compact vector representation suitable for a variety of high-level vision tasks such as AR and CAD. We conduct extensive evaluations of our method on a large new synthetic dataset of urban scenes as well as real images. Our code and datasets will be published along with the paper.
UR - http://www.scopus.com/inward/record.url?scp=85081915565&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85081915565&partnerID=8YFLogxK
U2 - 10.1109/ICCV.2019.00779
DO - 10.1109/ICCV.2019.00779
M3 - Conference contribution
AN - SCOPUS:85081915565
T3 - Proceedings of the IEEE International Conference on Computer Vision
SP - 7697
EP - 7706
BT - Proceedings - 2019 International Conference on Computer Vision, ICCV 2019
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 17th IEEE/CVF International Conference on Computer Vision, ICCV 2019
Y2 - 27 October 2019 through 2 November 2019
ER -