3
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: not found
      • Article: not found

      An integration of bottom-up and top-down salient cues on RGB-D data: saliency from objectness versus non-objectness

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Related collections

          Most cited references24

          • Record: found
          • Abstract: found
          • Article: not found

          Guided Search 2.0 A revised model of visual search.

          An important component of routine visual behavior is the ability to find one item in a visual world filled with other, distracting items. This ability to performvisual search has been the subject of a large body of research in the past 15 years. This paper reviews the visual search literature and presents a model of human search behavior. Built upon the work of Neisser, Treisman, Julesz, and others, the model distinguishes between a preattentive, massively parallel stage that processes information about basic visual features (color, motion, various depth cues, etc.) across large portions of the visual field and a subsequent limited-capacity stage that performs other, more complex operations (e.g., face recognition, reading, object identification) over a limited portion of the visual field. The spatial deployment of the limited-capacity process is under attentional control. The heart of the guided search model is the idea that attentional deployment of limited resources isguided by the output of the earlier parallel processes. Guided Search 2.0 (GS2) is a revision of the model in which virtually all aspects of the model have been made more explicit and/or revised in light of new data. The paper is organized into four parts: Part 1 presents the model and the details of its computer simulation. Part 2 reviews the visual search literature on preattentive processing of basic features and shows how the GS2 simulation reproduces those results. Part 3 reviews the literature on the attentional deployment of limited-capacity processes in conjunction and serial searches and shows how the simulation handles those conditions. Finally, Part 4 deals with shortcomings of the model and unresolved issues.
            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            Learning Deconvolution Network for Semantic Segmentation

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Contextual cueing: implicit learning and memory of visual context guides spatial attention.

              Global context plays an important, but poorly understood, role in visual tasks. This study demonstrates that a robust memory for visual context exists to guide spatial attention. Global context was operationalized as the spatial layout of objects in visual search displays. Half of the configurations were repeated across blocks throughout the entire session, and targets appeared within consistent locations in these arrays. Targets appearing in learned configurations were detected more quickly. This newly discovered form of search facilitation is termed contextual cueing. Contextual cueing is driven by incidentally learned associations between spatial configurations (context) and target locations. This benefit was obtained despite chance performance for recognizing the configurations, suggesting that the memory for context was implicit. The results show how implicit learning and memory of visual context can guide spatial attention towards task-relevant aspects of a scene.
                Bookmark

                Author and article information

                Journal
                Signal, Image and Video Processing
                SIViP
                Springer Nature
                1863-1703
                1863-1711
                February 2018
                August 20 2017
                February 2018
                : 12
                : 2
                : 307-314
                Article
                10.1007/s11760-017-1159-7
                5e6d8ebb-6559-4222-82bd-a3777aa8eb2c
                © 2018

                http://www.springer.com/tdm

                History

                Comments

                Comment on this article