1
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: not found

      A hierarchical framework for improving ride comfort of autonomous vehicles via deep reinforcement learning with external knowledge

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Ride comfort plays an important role in determining the public acceptance of autonomous vehicles (AVs). Many factors, such as road profile, driving speed, and suspension system, influence the ride comfort of AVs. This study proposes a hierarchical framework for improving ride comfort by integrating speed planning and suspension control in a vehicle‐to‐everything environment. Based on safe, comfortable, and efficient speed planning via dynamic programming, a deep reinforcement learning‐based suspension control is proposed to adapt to the changing pavement conditions. Specifically, a deep deterministic policy gradient with external knowledge (EK‐DDPG) algorithm is designed for the efficient self‐adaptation of suspension control strategies. The external knowledge of action selection and value estimation from other AVs are combined into the loss functions of the DDPG algorithm. In numerical experiments, real‐world pavements detected in 11 districts of Shanghai, China, are applied to verify the proposed method. Experimental results demonstrate that the EK‐DDPG‐based suspension control improves ride comfort on untrained rough pavements by 27.95% and 3.32%, compared to a model predictive control (MPC) baseline and a DDPG baseline, respectively. Meanwhile, the EK‐DDPG‐based suspension control improves computational efficiency by 22.97%, compared to the MPC baseline, and performs at the same level as the DDPD baseline. This study provides a generalized and computationally efficient approach for improving the ride comfort of AVs.

          Related collections

          Most cited references62

          • Record: found
          • Abstract: found
          • Article: not found

          Mastering the game of Go with deep neural networks and tree search.

          The game of Go has long been viewed as the most challenging of classic games for artificial intelligence owing to its enormous search space and the difficulty of evaluating board positions and moves. Here we introduce a new approach to computer Go that uses 'value networks' to evaluate board positions and 'policy networks' to select moves. These deep neural networks are trained by a novel combination of supervised learning from human expert games, and reinforcement learning from games of self-play. Without any lookahead search, the neural networks play Go at the level of state-of-the-art Monte Carlo tree search programs that simulate thousands of random games of self-play. We also introduce a new search algorithm that combines Monte Carlo simulation with value and policy networks. Using this search algorithm, our program AlphaGo achieved a 99.8% winning rate against other Go programs, and defeated the human European Go champion by 5 games to 0. This is the first time that a computer program has defeated a human professional player in the full-sized game of Go, a feat previously thought to be at least a decade away.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Continuous control with deep reinforcement learning

            We adapt the ideas underlying the success of Deep Q-Learning to the continuous action domain. We present an actor-critic, model-free algorithm based on the deterministic policy gradient that can operate over continuous action spaces. Using the same learning algorithm, network architecture and hyper-parameters, our algorithm robustly solves more than 20 simulated physics tasks, including classic problems such as cartpole swing-up, dexterous manipulation, legged locomotion and car driving. Our algorithm is able to find policies whose performance is competitive with those found by a planning algorithm with full access to the dynamics of the domain and its derivatives. We further demonstrate that for many of the tasks the algorithm can learn policies end-to-end: directly from raw pixel inputs. 10 pages + supplementary
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Playing Atari with Deep Reinforcement Learning

              We present the first deep learning model to successfully learn control policies directly from high-dimensional sensory input using reinforcement learning. The model is a convolutional neural network, trained with a variant of Q-learning, whose input is raw pixels and whose output is a value function estimating future rewards. We apply our method to seven Atari 2600 games from the Arcade Learning Environment, with no adjustment of the architecture or learning algorithm. We find that it outperforms all previous approaches on six of the games and surpasses a human expert on three of them. NIPS Deep Learning Workshop 2013
                Bookmark

                Author and article information

                Journal
                Computer-Aided Civil and Infrastructure Engineering
                Computer aided Civil Eng
                Wiley
                1093-9687
                1467-8667
                May 2023
                November 21 2022
                May 2023
                : 38
                : 8
                : 1059-1078
                Affiliations
                [1 ] Key Laboratory of Road and Traffic Engineering of the Ministry of Education Tongji University Shanghai 201804 China
                [2 ] Urban Planning and Transportation Group Eindhoven University of Technology Eindhoven 98195 The Netherlands
                [3 ] Department of Civil and Environmental Engineering University of Washington Seattle Washington USA
                Article
                10.1111/mice.12934
                a26de577-6422-4dc1-8b51-11757a5620e4
                © 2023

                http://onlinelibrary.wiley.com/termsAndConditions#vor

                History

                Comments

                Comment on this article