Please log in

Paper / Information search system

日本語

ENGLISH

Help

Please log in

  • Summary & Details

Research on Autonomous Driving Decision Based on Improved Deep Deterministic Policy Algorithm

Detailed Information

Author(E)1) Shi YK, 2) Jian Wu, 3) Shiping Song
Affiliation(E)1) Jilin University, 2) Jilin University, 3) Jilin University
Abstract(E)Autonomous driving technology, as the product of the fifth stage of the information technology revolution, is of great significance for improving urban traffic and environmentally friendly sustainable development. Autonomous driving can be divided into three main modules. The input of the decision module is the perception information from the perception module and the output of the control strategy to the control module. The deep reinforcement learning method proposes an end-to-end decision-making system design scheme. This paper adopts the Deep Deterministic Policy Gradient Algorithm (DDPG) that incorporates the Priority Experience Playback (PER) method. The framework of the algorithm is based on the actor-critic network structure model. The model takes the continuously acquired perception information as input and the continuous control of the vehicle as output. Combined with the CARLA simulation environment, the state space of the CNN network based on the input of the car's front view image is designed, and the action space design takes into account the actual situation of the accelerator brake being used at different times. In the design of the reward function, the reward function based on the car state information and the reward function based on the artificial potential field method (APF) are designed respectively. After that, based on the CARLA virtual urban driving environment, the DDPG algorithm and the PER-DDPG algorithm with different reward functions were simulated and verified, and tested in different scenarios. The final experimental results show that the APF-PER-DDPG algorithm performs best. Compared with the DDPG algorithm, the average reward has increased by about 27.7%, and the proportion of dangerous actions is the smallest, which has dropped by about 24.8%. The test results show that improving the sampling method and the reward function based on the artificial potential field can improve the performance of the algorithm.

About search

close

How to use the search box

You can enter up to 5 search conditions. The number of search boxes can be increased or decreased with the "+" and "-" buttons on the right.
If you enter multiple words separated by spaces in one search box, the data that "contains all" of the entered words will be searched (AND search).
Example) X (space) Y → "X and Y (including)"

How to use "AND" and "OR" pull-down

If "AND" is specified, the "contains both" data of the phrase entered in the previous and next search boxes will be searched. If you specify "OR", the data that "contains" any of the words entered in the search boxes before and after is searched.
Example) X AND Y → "X and Y (including)"  X OR Z → "X or Z (including)"
If AND and OR searches are mixed, OR search has priority.
Example) X AND Y OR Z → X AND (Y OR Z)
If AND search and multiple OR search are mixed, OR search has priority.
Example) W AND X OR Y OR Z → W AND (X OR Y OR Z)

How to use the search filters

Use the "search filters" when you want to narrow down the search results, such as when there are too many search results. If you check each item, the search results will be narrowed down to only the data that includes that item.
The number in "()" after each item is the number of data that includes that item.

Search tips

When searching by author name, enter the first and last name separated by a space, such as "Taro Jidosha".