50
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: not found

      Time Series Classification with HIVE-COTE : The Hierarchical Vote Collective of Transformation-Based Ensembles

      1 , 1 , 1
      ACM Transactions on Knowledge Discovery from Data
      Association for Computing Machinery (ACM)

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          A recent experimental evaluation assessed 19 time series classification (TSC) algorithms and found that one was significantly more accurate than all others: the Flat Collective of Transformation-based Ensembles (Flat-COTE). Flat-COTE is an ensemble that combines 35 classifiers over four data representations. However, while comprehensive, the evaluation did not consider deep learning approaches. Convolutional neural networks (CNN) have seen a surge in popularity and are now state of the art in many fields and raises the question of whether CNNs could be equally transformative for TSC.

          We implement a benchmark CNN for TSC using a common structure and use results from a TSC-specific CNN from the literature. We compare both to Flat-COTE and find that the collective is significantly more accurate than both CNNs. These results are impressive, but Flat-COTE is not without deficiencies. We significantly improve the collective by proposing a new hierarchical structure with probabilistic voting, defining and including two novel ensemble classifiers built in existing feature spaces, and adding further modules to represent two additional transformation domains. The resulting classifier, the Hierarchical Vote Collective of Transformation-based Ensembles (HIVE-COTE), encapsulates classifiers built on five data representations. We demonstrate that HIVE-COTE is significantly more accurate than Flat-COTE (and all other TSC algorithms that we are aware of) over 100 resamples of 85 TSC problems and is the new state of the art for TSC. Further analysis is included through the introduction and evaluation of 3 new case studies and extensive experimentation on 1,000 simulated datasets of 5 different types.

          Related collections

          Most cited references36

          • Record: found
          • Abstract: not found
          • Article: not found

          The WEKA data mining software

            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Rotation forest: A new classifier ensemble method.

            We propose a method for generating classifier ensembles based on feature extraction. To create the training data for a base classifier, the feature set is randomly split into K subsets (K is a parameter of the algorithm) and Principal Component Analysis (PCA) is applied to each subset. All principal components are retained in order to preserve the variability information in the data. Thus, K axis rotations take place to form the new features for a base classifier. The idea of the rotation approach is to encourage simultaneously individual accuracy and diversity within the ensemble. Diversity is promoted through the feature extraction for each base classifier. Decision trees were chosen here because they are sensitive to rotation of the feature axes, hence the name "forest." Accuracy is sought by keeping all principal components and also using the whole data set to train each base classifier. Using WEKA, we examined the Rotation Forest ensemble on a random selection of 33 benchmark data sets from the UCI repository and compared it with Bagging, AdaBoost, and Random Forest. The results were favorable to Rotation Forest and prompted an investigation into diversity-accuracy landscape of the ensemble models. Diversity-error diagrams revealed that Rotation Forest ensembles construct individual classifiers which are more accurate than these in AdaBoost and Random Forest, and more diverse than these in Bagging, sometimes more accurate as well.
              Bookmark
              • Record: found
              • Abstract: not found
              • Conference Proceedings: not found

              A Convolutional Neural Network for Modelling Sentences

                Bookmark

                Author and article information

                Journal
                ACM Transactions on Knowledge Discovery from Data
                ACM Trans. Knowl. Discov. Data
                Association for Computing Machinery (ACM)
                1556-4681
                1556-472X
                July 20 2018
                July 20 2018
                : 12
                : 5
                : 1-35
                Affiliations
                [1 ]University of East Anglia, Norwich, United Kingdom
                Article
                10.1145/3182382
                2ef8bbc4-a0f0-41b8-a93e-3a588057c4ab
                © 2018

                http://www.acm.org/publications/policies/copyright_policy#Background

                History

                Comments

                Comment on this article