Self-Driving Cars Using Genetic Algorithm
Self-Driving Cars Using Genetic Algorithm
Self-Driving Cars Using Genetic Algorithm
https://doi.org/10.22214/ijraset.2020.32200
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue XI Nov 2020- Available at www.ijraset.com
Abstract: This paper deals with a 2D unity simulation of an autonomous car learning to drive in a simplified environment
containing only lane and static obstacles. Learning is performed using the neural network and evolutionary algorithm. The cars
navigate by evolving themselves from the previous generation. The neural network used in this paper is standard, fully
connected, feed forward neural network. The cars are steered by a feed forward Neural Network. The weights of the network are
trained using a modified genetic algorithm.
Keywords: Autonomous Car, Self Driving cars, Deep Learning Cars, Genetic Cars, Self driving cars using EANN.
I. INTRODUCTION
Autonomous vehicle was an imaginary idea which came to reality with advent of deep learning based machine learning in AI. AV’s
main motive is to bring a safety to driving. Human driving is an accident prone. We are drunk, tired and irrational decision maker.
2019’s annual report of the International Traffic Safety Data and Analysis Group (IRTAD) over 1.3 million people lost their lives
annually and 10 million people are seriously injured [10]. More than 50% of those injured are pedestrians. AV can bring possibility
to replace accident prone human driving by providing safety and comfort.
Fayjie et al has applied an approach for obstacle avoidance and navigation in urban environment using Deep Reinforcement learning
applied Deep Q network. Camera sensors and laser sensors are used to collect the input data. This approach uses Hokugo Lidar and
a camera. For running a deep learning algorithm it uses an embedded GPU.
GA is good at searching large-scale, complex, non-differentiable and multimodal spaces; it doesn’t need the gradient information of
the error function. On the other hand, it doesn’t need to consider whether the error function is differentiable, so some punishments
may be added into the error function, so as to improve the network commonality, reduce the complexity of the network.
This paper is organized as follows. Section II presents summary of challenges for autonomous vehicles and problems that have been
overcome by different methodology. Applying evolutionary algorithm on N cars to navigate themselves through different courses
and avoid obstacles is given in Section III. Section IV presents conclusion.
Perception of the environment remains the biggest challenge to reliable and safe driving. [7] Authors presented control strategies,
robust control strategies and parameter observation based control strategies on path tracking for autonomous vehicle. Autonomous
Vehicle technology’s main aim was to improve driving comfort, safety, economy as well as reduce road accident rate. Path tracking
is a basic part of autonomous vehicle aims to follow the route accurately, ensure vehicle stability and satisfies the robust
performance of the control system. [9] Author has applied an approach for obstacle avoidance and navigation in urban environment
using Deep Reinforcement learning applied Deep Q network. Camera sensors and laser sensors are used to collect the input data.
This approach uses Hokugo Lidar and a camera. For running a deep learning algorithm it uses an embedded GPU.
III. METHODOLOGY
The first generation starts with 20 cars each of them having its own neural network, makes its intelligence. They have five front
facing sensors covering approximately 90 degrees. The maximum range of sensor is 10 unity units. The neural network comprise of
4 layers, the first layer is an input layer with 5 neurons, two hidden layers having 4 and 3 neurons and an output layer with 2
neurons.
The weights of neural network are trained using genetic algorithm. At first there are 20 randomly initialized cars spawned. The best
cars are selected and recombined and create new offspring cars. These offspring cars create new population of 20 cars which are
slightly mutated in order to inject more diversity into population. The newly created cars then try to navigate the course again and
the process of evaluation, selection and recombination and mutation starts again one complete cycle from evaluation of one
population to the evaluation of the next is called a generation.
The first five points you are seeing in fig a are just the current reading of the five distance sensors of the car. These sensors measure
the distance to the nearest wall. The blue crosses are simply there to visualize where the sensors are currently pointing. They all
have their own weights. However they share their learning in the sense that best individual of one generation are used to create
individual of next generation. In the order to create new individual two other individual are crossed and the offspring is then
mutated. This mutation makes them better than previous generation.
Figure b Figure c
Figure e
IV. CONCLUSION
In this paper, an approach of deep learning in autonomous car learning to navigate in a simplified environment and static obstacles
is proposed. This approach used a Neural Network and evolutionary algorithm to train the car in the 2D unity simulated
environment. The verification results are expected to demonstrate that learning autonomous driving in a simulated environment is a
step towards driving on real streets.
V. ACKNOWLEDGEMENT
The author would like to acknowledge the support and guidance of Prof. Swapna Augustine Nikale, Department of Information
Technology, B. K. Birla College (Autonomous), Kalyan.
REFERENCES
[1] Hafeez, F., Sheikh, U. U., Alkhaldi, N., Garni, H. Z. A., Arfeen, Z. A., & Khalid, S. A. (2020). Insights and Strategies for an Autonomous Vehicle With a
Sensor Fusion Innovation: A Fictional Outlook. IEEE Access, 8, 135162–135175. https://doi.org/10.1109/access.2020.3010940
[2] Fridman, L., Brown, D. E., Glazer, M., Angell, W., Dodd, S., Jenik, B., Terwilliger, J., Patsekin, A., Kindelsberger, J., Ding, L., Seaman, S., Mehler, A.,
Sipperley, A., Pettinato, A., Seppelt, B. D., Angell, L., Mehler, B., & Reimer, B. (2019). MIT Advanced Vehicle Technology Study: Large-Scale Naturalistic
Driving Study of Driver Behavior and Interaction with Automation. IEEE Access, 7, 102021–102038. https://doi.org/10.1109/access.2019.2926040
[3] Luettel, T., Himmelsbach, M., & Wuensche, H.-J. (2012). Autonomous Ground Vehicles—Concepts and a Path to the Future. Proceedings of the IEEE,
100(Special Centennial Issue), 1831–1839. https://doi.org/10.1109/jproc.2012.2189803
[4] Memon, Q., Ahmed, M., Ali, S., Memon, A. R., & Shah, W. (2016). Self-driving and driver relaxing vehicle. 2016 2nd International Conference on Robotics
and Artificial Intelligence (ICRAI), 170–174. https://doi.org/10.1109/icrai.2016.7791248
[5] James, S. S., Anderson, S. R., & Lio, M. D. (2020). Longitudinal Vehicle Dynamics: A Comparison of Physical and Data-Driven Models Under Large-Scale
Real-World Driving Conditions. IEEE Access, 8, 73714–73729. https://doi.org/10.1109/access.2020.2988592
[6] Bagloee, S. A., Tavana, M., Asadi, M., & Oliver, T. (2016). Autonomous vehicles: challenges, opportunities, and future implications for transportation policies.
Journal of Modern Transportation, 24(4), 284–303. https://doi.org/10.1007/s40534-016-0117
[7] Yao, Q., Tian, Y., Wang, Q., & Wang, S. (2020). Control Strategies on Path Tracking for Autonomous Vehicle: state of the art and future challenges. IEEE
Access. https://doi.org/10.1109/access.2020.3020075
[8] Sophia H. Duffy et al., Sit, Stay, Drive: The Future of Autonomous Car Liability, 16SMU Science and Technology Law Review. 453 (2016)
https://scholar.smu.edu/scitech/vol16/iss3/4
[9] Fayjie, A. R., Hossain, S., Oualid, D., & Lee, D.-J. (2018). Driverless Car: Autonomous Driving Using Deep Reinforcement Learning in Urban Environment.
2018 15th International Conference on Ubiquitous Robots (UR), 896–901. https://doi.org/10.1109/urai.2018.844179
[10] Rasouli, A., & Tsotsos, J. K. (2020). Autonomous Vehicles That Interact With Pedestrians: A Survey of Theory and Practice. IEEE Transactions on Intelligent
Transportation Systems, 21(3), 900–918. https://doi.org/10.1109/tits.2019.2901817
[11] SAE J3016 automated-driving graphic. (2020, May 15). SAE International. https://www.sae.org/news/2019/01/sae-updates-j3016-automated-driving-graphic
[12] Yurtsever, E., Lambert, J., Carballo, A., & Takeda, K. (2020). A Survey of Autonomous Driving: Common Practices and Emerging Technologies. IEEE
Access, 8, 58443–58469. https://doi.org/10.1109/access.2020.2983149
[13] Dinesh Kumar, A., Karthika, R., & Soman, K. P. (2020). Stereo Camera and LIDAR Sensor Fusion-Based Collision Warning System for Autonomous
Vehicles. Advances in Computational Intelligence Techniques, 239–252. https://doi.org/10.1007/978-981-15-2620-6_17
[14] Geiger, A., Lauer, M., Moosmann, F., Ranft, B., Rapp, H., Stiller, C., & Ziegler, J. (2012). Team AnnieWAY’s Entry to the 2011 Grand Cooperative Driving
Challenge. IEEE Transactions on Intelligent Transportation Systems, 13(3), 1008–1017. https://doi.org/10.1109/tits.2012.2189882
[15] Patel, N., Choromanska, A., Krishnamurthy, P., & Khorrami, F. (2019). A deep learning gated architecture for UGV navigation robust to sensor failures.
Robotics and Autonomous Systems, 116, 80–97. https://doi.org/10.1016/j.robot.2019.03.001
[16] Li, Q., Chen, L., Li, M., Shaw, S.-L., & Nuchter, A. (2014). A Sensor-Fusion Drivable-Region and Lane-Detection System for Autonomous Vehicle
Navigation in Challenging Road Scenarios. IEEE Transactions on Vehicular Technology, 63(2), 540–555. https://doi.org/10.1109/tvt.2013.2281199
[17] Chavez-Garcia, R. O., & Aycard, O. (2016). Multiple Sensor Fusion and Classification for Moving Object Detection and Tracking. IEEE Transactions on
Intelligent Transportation Systems, 17(2), 525–534. https://doi.org/10.1109/tits.2015.2479925