CVPR 2014 Oral Talks
TechTalks from event: CVPR 2014 Oral Talks
Orals 4F : View Synthesis & Other Applications
-
Reconstructing Storyline Graphs for Image Recommendation from Web Community PhotosIn this paper, we investigate an approach for reconstructing storyline graphs from large-scale collections of Internet images, and optionally other side information such as friendship graphs. The storyline graphs can be an effective summary that visualizes various branching narrative structure of events or activities recurring across the input photo sets of a topic class. In order to explore further the usefulness of the storyline graphs, we leverage them to perform the image sequential prediction tasks, from which photo recommendation applications can benefit. We formulate the storyline reconstruction problem as an inference of sparse time-varying directed graphs, and develop an optimization algorithm that successfully addresses a number of key challenges of Web-scale problems, including global optimality, linear complexity, and easy parallelization. With experiments on more than 3.3 millions of images of 24 classes and user studies via Amazon Mechanical Turk, we show that the proposed algorithm improves other candidate methods for both storyline reconstruction and image prediction tasks.
-
Active Flattening of Curved Document Images via Two Structured BeamsDocument images captured by a digital camera often suffer from serious geometric distortions. In this paper,we propose an active method to correct geometric distortions in a camera-captured document image. Unlike many passive rectification methods that rely on text-lines or features extracted from images, our method uses two structured beams illuminating upon the document page to recover two spatial curves. A developable surface is then interpolated to the curves by finding the correspondence between them. The developable surface is finally flattened onto a plane by solving a system of ordinary differential equations. Our method is a content independent approach and can restore a corrected document image of high accuracy with undistorted contents. Experimental results on a variety of real-captured document images demonstrate the effectiveness and efficiency of the proposed method.
-
Image-based Synthesis and Re-Synthesis of Viewpoints Guided by 3D ModelsWe propose a technique to use the structural informa- tion extracted from a set of 3D models of an object class to improve novel-view synthesis for images showing unknown instances of this class. These novel views can be used to "amplify" training image collections that typically contain only a low number of views or lack certain classes of views entirely (e. g. top views). We extract the correlation of position, normal, re- flectance and appearance from computer-generated images of a few exemplars and use this information to infer new appearance for new instances. We show that our approach can improve performance of state-of-the-art detectors using real-world training data. Additional applications include guided versions of inpainting, 2D-to-3D conversion, super- resolution and non-local smoothing.
-
Bayesian View Synthesis and Image-Based Rendering PrinciplesIn this paper, we address the problem of synthesizing novel views from a set of input images. State of the art methods, such as the Unstructured Lumigraph, have been using heuristics to combine information from the original views, often using an explicit or implicit approximation of the scene geometry. While the proposed heuristics have been largely explored and proven to work effectively, a Bayesian formulation was recently introduced, formalizing some of the previously proposed heuristics, pointing out which physical phenomena could lie behind each. However, some important heuristics were still not taken into account and lack proper formalization. We contribute a new physics-based generative model and the corresponding Maximum a Posteriori estimate, providing the desired unification between heuristics-based methods and a Bayesian formulation. The key point is to systematically consider the error induced by the uncertainty in the geometric proxy. We provide an extensive discussion, analyzing how the obtained equations explain the heuristics developed in previous methods. Furthermore, we show that our novel Bayesian model significantly improves the quality of novel views, in particular if the scene geometry estimate is inaccurate.
- All Sessions
- Orals 1A : Matching & Reconstruction
- Orals 1B : Segmentation & Grouping
- Orals 1C : Statistical Methods & Learning I
- Orals 1D : Action Recognition
- Orals 2A : Motion & Tracking
- Orals 2B : Discrete Optimization
- Orals 2D : Attribute-Based Recognition & Human Pose Estimation
- Orals 2F : Convolutional Neural Networks
- Orals 3A : Physics-Based Vision & Shape-from-X
- Orals 3B : Video: Events, Activities & Surveillance
- Orals 3C : Medical & Biological Image Analysis
- Orals 3D : Low-Level Vision & Image Processing
- Orals 4A : Computational Photography: Sensing and Display
- Orals 4B : Recognition: Detection, Categorization, Classification
- Orals 4C : 3D Geometry & Shape
- Orals 4D : Statistical Methods and Learning II
- Orals 2E : Face & Gesture
- Orals 4E : Optimization Methods
- Orals 4F : View Synthesis & Other Applications
- Special 1 : Plenary Session
- Special 2 : Special Journal Session
- Special 3 : Plenary Session