DepthScape: Authoring 2.5D Designs via Depth Estimation, Semantic Understanding, and Geometry Extraction
By: Xia Su , Cuong Nguyen , Matheus A. Gadelha and more
Potential Business Impact:
Makes flat pictures look 3D with depth.
2.5D effects, such as occlusion and perspective foreshortening, enhance visual dynamics and realism by incorporating 3D depth cues into 2D designs. However, creating such effects remains challenging and labor-intensive due to the complexity of depth perception. We introduce DepthScape, a human-AI collaborative system that facilitates 2.5D effect creation by directly placing design elements into 3D reconstructions. Using monocular depth reconstruction, DepthScape transforms images into 3D reconstructions where visual contents are placed to automatically achieve realistic occlusion and perspective foreshortening. To further simplify 3D placement through a 2D viewport, DepthScape uses a vision-language model to analyze source images and extract key visual components as content anchors for direct manipulation editing. We evaluate DepthScape with nine participants of varying design backgrounds, confirming the effectiveness of our creation pipeline. We also test on 100 professional stock images to assess robustness, and conduct an expert evaluation that confirms the quality of DepthScape's results.
Similar Papers
DepthFocus: Controllable Depth Estimation for See-Through Scenes
CV and Pattern Recognition
Lets computers see through glass like humans.
DASC: Depth-of-Field Aware Scene Complexity Metric for 3D Visualization on Light Field Display
Multimedia
Makes 3D pictures look good from any distance.
Viewpoint-Tolerant Depth Perception for Shared Extended Space Experience on Wall-Sized Display
Human-Computer Interaction
Makes big screens show 3D for everyone, no glasses.