skip to main content
10.1145/3532719.3543235acmconferencesArticle/Chapter ViewAbstractPublication PagessiggraphConference Proceedingsconference-collections
poster

Interactive Editing of Monocular Depth

Published:25 July 2022Publication History

ABSTRACT

Recent advances in computer vision have made 3D structure-aware editing of still photographs a reality. Such computational photography applications use a depth map that is automatically generated by monocular depth estimation methods to represent the scene structure. In this work, we present a lightweight, web-based interactive depth editing and visualization tool that adapts low-level conventional image editing operations for geometric manipulation to enable artistic control in the 3D photography workflow. Our tool provides real-time feedback on the geometry through a 3D scene visualization to make the depth map editing process more intuitive for artists. Our web-based tool is open-source1 and platform-independent to support wider adoption of 3D photography techniques in everyday digital photography.

Skip Supplemental Material Section

Supplemental Material

siggraph_poster_video.mp4

mp4

70.8 MB

References

  1. S. Mahdi H. Miangoleh, Sebastian Dille, Long Mai, Sylvain Paris, and Yağız Aksoy. 2021. Boosting Monocular Depth Estimation Models to High-Resolution via Content-Adaptive Multi-Resolution Merging. In IEEE Conf. Comput. Vis. Pattern Recog.Google ScholarGoogle ScholarCross RefCross Ref
  2. Simon Niklaus, Long Mai, Jimei Yang, and Feng Liu. 2019. 3D ken burns effect from a single image. ACM Trans. Graph. (2019).Google ScholarGoogle Scholar
  3. René Ranftl, Alexey Bochkovskiy, and Vladlen Koltun. 2021. Vision transformers for dense prediction. In Int. Conf. Comput. Vis.Google ScholarGoogle ScholarCross RefCross Ref
  4. Meng-Li Shih, Shih-Yang Su, Johannes Kopf, and Jia-Bin Huang. 2020. 3D photography using context-aware layered depth inpainting. In IEEE Conf. Comput. Vis. Pattern Recog.Google ScholarGoogle ScholarCross RefCross Ref
  5. Neal Wadhwa, Rahul Garg, David E Jacobs, Bryan E Feldman, Nori Kanazawa, Robert Carroll, Yair Movshovitz-Attias, Jonathan T Barron, Yael Pritch, and Marc Levoy. 2018. Synthetic depth-of-field with a single-camera mobile phone. ACM Trans. Graph. (2018).Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    SIGGRAPH '22: ACM SIGGRAPH 2022 Posters
    July 2022
    132 pages
    ISBN:9781450393614
    DOI:10.1145/3532719

    Copyright © 2022 Owner/Author

    Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 25 July 2022

    Check for updates

    Qualifiers

    • poster
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate1,822of8,601submissions,21%

    Upcoming Conference

    SIGGRAPH '24
  • Article Metrics

    • Downloads (Last 12 months)54
    • Downloads (Last 6 weeks)2

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format