16
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Improved Transition-Based Parsing by Modeling Characters instead of Words with LSTMs

      Preprint
      , ,

      Read this article at

          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          We present extensions to a continuous-state dependency parsing method that makes it applicable to morphologically rich languages. Starting with a high-performance transition-based parser that uses long short-term memory (LSTM) recurrent neural networks to learn representations of the parser state, we replace lookup-based word representations with representations constructed from the orthographic representations of the words, also using LSTMs. This allows statistical sharing across word forms that are similar on the surface. Experiments for morphologically rich languages show that the parsing model benefits from incorporating the character-based encodings of words.

          Related collections

          Most cited references3

          • Record: found
          • Abstract: found
          • Article: not found

          Framewise phoneme classification with bidirectional LSTM and other neural network architectures.

          In this paper, we present bidirectional Long Short Term Memory (LSTM) networks, and a modified, full gradient version of the LSTM learning algorithm. We evaluate Bidirectional LSTM (BLSTM) and several other network architectures on the benchmark task of framewise phoneme classification, using the TIMIT database. Our main findings are that bidirectional networks outperform unidirectional ones, and Long Short Term Memory (LSTM) is much faster and also more accurate than both standard Recurrent Neural Nets (RNNs) and time-windowed Multilayer Perceptrons (MLPs). Our results support the view that contextual information is crucial to speech processing, and suggest that BLSTM is an effective architecture with which to exploit it.
            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            Incrementality in deterministic dependency parsing

              Bookmark
              • Record: found
              • Abstract: not found
              • Book Chapter: not found

              Building a Turkish Treebank

                Bookmark

                Author and article information

                Journal
                2015-08-04
                2015-08-11
                Article
                1508.00657
                151558f8-1ac0-481b-85fe-71c07fa21168

                http://arxiv.org/licenses/nonexclusive-distrib/1.0/

                History
                Custom metadata
                In Proceedings of EMNLP 2015
                cs.CL

                Theoretical computer science
                Theoretical computer science

                Comments

                Comment on this article