2
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Transformer-based land use and land cover classification with explainability using satellite imagery

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Transformer-based models have greatly improved Land Use and Land Cover (LULC) applications. Their revolutionary ability to analyze and extract key information has greatly advanced the field. However, the high computational cost of these models presents a considerable obstacle to their practical implementation. Therefore, this study aims to strike a balance between computational cost and accuracy when employing transformer-based models for LULC analysis. We exploit transfer learning and fine-tuning strategies to optimize the resource utilization of transformer-based models. Furthermore, transparency is the core principle of our methodology to promote fairness and trust in applying LULC models across various domains, including forestry, environmental studies, and urban or rural planning. To ensure transparency, we have employed Captum, which enables us to uncover and mitigate potential biases and interpret AI-driven decisions. Our results indicate that transfer learning can potentially improve transformer-based models in satellite image classification, and strategic fine-tuning can maintain efficiency with minimal accuracy trade-offs. This research highlights the potential of Explainable AI (XAI) in Transformer-based models for achieving more efficient and transparent LULC analysis, thereby encouraging continued innovation in the field.

          Related collections

          Most cited references9

          • Record: found
          • Abstract: not found
          • Article: not found

          A Survey on Transfer Learning

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Deep Learning in Remote Sensing: A Comprehensive Review and List of Resources

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Transformers in Vision: A Survey

              Astounding results from Transformer models on natural language tasks have intrigued the vision community to study their application to computer vision problems. Among their salient benefits, Transformers enable modeling long dependencies between input sequence elements and support parallel processing of sequence as compared to recurrent networks e.g. , Long short-term memory (LSTM). Different from convolutional networks, Transformers require minimal inductive biases for their design and are naturally suited as set-functions. Furthermore, the straightforward design of Transformers allows processing multiple modalities ( e.g. , images, videos, text and speech) using similar processing blocks and demonstrates excellent scalability to very large capacity networks and huge datasets. These strengths have led to exciting progress on a number of vision tasks using Transformer networks. This survey aims to provide a comprehensive overview of the Transformer models in the computer vision discipline. We start with an introduction to fundamental concepts behind the success of Transformers i.e., self-attention, large-scale pre-training, and bidirectional feature encoding. We then cover extensive applications of transformers in vision including popular recognition tasks ( e.g. , image classification, object detection, action recognition, and segmentation), generative modeling, multi-modal tasks ( e.g. , visual-question answering, visual reasoning, and visual grounding), video processing ( e.g. , activity recognition, video forecasting), low-level vision ( e.g. , image super-resolution, image enhancement, and colorization) and 3D analysis ( e.g. , point cloud classification and segmentation). We compare the respective advantages and limitations of popular techniques both in terms of architectural design and their experimental value. Finally, we provide an analysis on open research directions and possible future works. We hope this effort will ignite further interest in the community to solve current challenges towards the application of transformer models in computer vision.
                Bookmark

                Author and article information

                Contributors
                mehakkhan3@hotmail.com
                reza.arghandeh@hvl.no
                Journal
                Sci Rep
                Sci Rep
                Scientific Reports
                Nature Publishing Group UK (London )
                2045-2322
                20 July 2024
                20 July 2024
                2024
                : 14
                : 16744
                Affiliations
                Department of Computer Science, Electrical Engineering and Mathematical Sciences, Western Norway University of Applied Sciences, ( https://ror.org/05phns765) Bergen, Norway
                Article
                67186
                10.1038/s41598-024-67186-4
                11271450
                39033183
                75488992-a48f-4cd7-975b-e7ab65e8db3e
                © The Author(s) 2024

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

                History
                : 26 January 2024
                : 9 July 2024
                Funding
                Funded by: Europeon Space Agency
                Funded by: Western Norway University Of Applied Sciences
                Categories
                Article
                Custom metadata
                © Springer Nature Limited 2024

                Uncategorized
                computer science,ecology
                Uncategorized
                computer science, ecology

                Comments

                Comment on this article