Makantasis et al., 2020 - Google Patents
Deep reinforcement‐learning‐based driving policy for autonomous road vehiclesMakantasis et al., 2020
View PDF- Document ID
- 13647309572985860824
- Author
- Makantasis K
- Kontorinaki M
- Nikolos I
- Publication year
- Publication venue
- IET Intelligent Transport Systems
External Links
Snippet
In this work, the problem of path planning for an autonomous vehicle that moves on a freeway is considered. The most common approaches that are used to address this problem are based on optimal control methods, which make assumptions about the model of the …
- 230000006399 behavior 0 abstract description 37
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/04—Inference methods or devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/02—Knowledge representation
- G06N5/022—Knowledge engineering, knowledge acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06Q—DATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/04—Forecasting or optimisation, e.g. linear programming, "travelling salesman problem" or "cutting stock problem"
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computer systems based on specific mathematical models
- G06N7/005—Probabilistic networks
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Makantasis et al. | Deep reinforcement‐learning‐based driving policy for autonomous road vehicles | |
| Duan et al. | Hierarchical reinforcement learning for self‐driving decision‐making without reliance on labelled driving data | |
| Li et al. | Deep reinforcement learning enabled decision-making for autonomous driving at intersections | |
| Li et al. | Combining decision making and trajectory planning for lane changing using deep reinforcement learning | |
| Zheng et al. | Behavioral decision‐making model of the intelligent vehicle based on driving risk assessment | |
| Luo et al. | Real-time cooperative vehicle coordination at unsignalized road intersections | |
| Hua et al. | Multi-agent reinforcement learning for connected and automated vehicles control: Recent advancements and future prospects | |
| Deng et al. | Deep reinforcement learning based decision-making strategy of autonomous vehicle in highway uncertain driving environments | |
| Ashwin et al. | Deep reinforcement learning for autonomous vehicles: lane keep and overtaking scenarios with collision avoidance | |
| Liu et al. | Heuristics‐oriented overtaking decision making for autonomous vehicles using reinforcement learning | |
| Wang et al. | Vehicle Trajectory Prediction by Knowledge‐Driven LSTM Network in Urban Environments | |
| Zhou et al. | Autonomous vehicles’ intended cooperative motion planning for unprotected turning at intersections | |
| Li et al. | Personalized lane change decision algorithm using deep reinforcement learning approach | |
| Liu et al. | Graph reinforcement learning application to co-operative decision-making in mixed autonomy traffic: Framework, survey, and challenges | |
| Liu et al. | Comprehensive predictive control method for automated vehicles in dynamic traffic circumstances | |
| Cai et al. | Rule‐constrained reinforcement learning control for autonomous vehicle left turn at unsignalized intersection | |
| Hu et al. | Manoeuvre prediction and planning for automated and connected vehicles based on interaction and gaming awareness under uncertainty | |
| Fu et al. | Framework and operation of digital twin smart freeway | |
| Zhang et al. | Intelligent driving model considering vehicular dynamics and heterogeneous road environments | |
| Zhai et al. | Model for the cooperative obstacle‐avoidance of the automated vehicle swarm in a connected vehicles environment | |
| Liu et al. | Dynamic motion planner with trajectory optimisation for automated highway lane‐changing driving | |
| Esaid et al. | Machine learning-based eco-approach and departure: Real-time trajectory optimization at connected signalized intersections | |
| Hu et al. | Toward multi-task generalization in autonomous navigation: A human-in-the-loop adversarial reinforcement learning with diffusion policy | |
| Lee et al. | Automated driving control in mixed traffic flow using V2V communication | |
| Aslam et al. | An AIoT-enabled digital twin CAVs with a DRL-based framework for trajectory planning |