0
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      UASOL, a large-scale high-resolution outdoor stereo dataset

      data-paper

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          In this paper, we propose a new dataset for outdoor depth estimation from single and stereo RGB images. The dataset was acquired from the point of view of a pedestrian. Currently, the most novel approaches take advantage of deep learning-based techniques, which have proven to outperform traditional state-of-the-art computer vision methods. Nonetheless, these methods require large amounts of reliable ground-truth data. Despite there already existing several datasets that could be used for depth estimation, almost none of them are outdoor-oriented from an egocentric point of view. Our dataset introduces a large number of high-definition pairs of color frames and corresponding depth maps from a human perspective. In addition, the proposed dataset also features human interaction and great variability of data, as shown in this work.

          Abstract

          Design Type(s) modeling and simulation objective • image creation and editing objective • database creation objective
          Measurement Type(s) image
          Technology Type(s) digital camera
          Factor Type(s) geographic location
          Sample Characteristic(s) Alicante • anthropogenic habitat

          Machine-accessible metadata file describing the reported data (ISA-Tab format)

          Related collections

          Most cited references8

          • Record: found
          • Abstract: not found
          • Article: not found

          Vision meets robotics: The KITTI dataset

            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Stereo processing by semiglobal matching and mutual information.

            This paper describes the Semi-Global Matching (SGM) stereo method. It uses a pixelwise, Mutual Information based matching cost for compensating radiometric differences of input images. Pixelwise matching is supported by a smoothness constraint that is usually expressed as a global cost function. SGM performs a fast approximation by pathwise optimizations from all directions. The discussion also addresses occlusion detection, subpixel refinement and multi-baseline matching. Additionally, postprocessing steps for removing outliers, recovering from specific problems of structured environments and the interpolation of gaps are presented. Finally, strategies for processing almost arbitrarily large images and fusion of disparity images using orthographic projection are proposed.A comparison on standard stereo images shows that SGM is among the currently top-ranked algorithms and is best, if subpixel accuracy is considered. The complexity is linear to the number of pixels and disparity range, which results in a runtime of just 1-2s on typical test images. An in depth evaluation of the Mutual Information based matching cost demonstrates a tolerance against a wide range of radiometric transformations. Finally, examples of reconstructions from huge aerial frame and pushbroom images demonstrate that the presented ideas are working well on practical problems.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Make3D: learning 3D scene structure from a single still image.

              We consider the problem of estimating detailed 3D structure from a single still image of an unstructured environment. Our goal is to create 3D models that are both quantitatively accurate as well as visually pleasing. For each small homogeneous patch in the image, we use a Markov Random Field (MRF) to infer a set of "plane parameters" that capture both the 3D location and 3D orientation of the patch. The MRF, trained via supervised learning, models both image depth cues as well as the relationships between different parts of the image. Other than assuming that the environment is made up of a number of small planes, our model makes no explicit assumptions about the structure of the scene; this enables the algorithm to capture much more detailed 3D structure than does prior art and also give a much richer experience in the 3D flythroughs created using image-based rendering, even for scenes with significant nonvertical structure. Using this approach, we have created qualitatively correct 3D models for 64.9 percent of 588 images downloaded from the Internet. We have also extended our model to produce large-scale 3D models from a few images.
                Bookmark

                Author and article information

                Contributors
                zbauer@dccia.ua.es
                Journal
                Sci Data
                Sci Data
                Scientific Data
                Nature Publishing Group UK (London )
                2052-4463
                29 August 2019
                29 August 2019
                2019
                : 6
                : 162
                Affiliations
                ISNI 0000 0001 2168 1800, GRID grid.5268.9, Institute for Computer Research, , University of Alicante, ; P.O. Box 99, 03080 Alicante, Spain
                Author information
                http://orcid.org/0000-0001-8447-2344
                Article
                168
                10.1038/s41597-019-0168-5
                6715739
                31467361
                06b39bc8-a3b2-4cab-adad-49772ec3394f
                © The Author(s) 2019

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

                The Creative Commons Public Domain Dedication waiver http://creativecommons.org/publicdomain/zero/1.0/ applies to the metadata files associated with this article.

                History
                : 29 January 2019
                : 29 July 2019
                Categories
                Data Descriptor
                Custom metadata
                © The Author(s) 2019

                databases,electrical and electronic engineering
                databases, electrical and electronic engineering

                Comments

                Comment on this article