<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.2 20190208//EN" "http://jats.nlm.nih.gov/publishing/1.2/JATS-journalpublishing1.dtd">
<article article-type="research-article" dtd-version="1.2" xml:lang="ru" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"><front><journal-meta><journal-id journal-id-type="issn">2518-1092</journal-id><journal-title-group><journal-title>Research result. Information technologies</journal-title></journal-title-group><issn pub-type="epub">2518-1092</issn></journal-meta><article-meta><article-id pub-id-type="doi">10.18413/2518-1092-2024-9-2-0-8</article-id><article-id pub-id-type="publisher-id">3495</article-id><article-categories><subj-group subj-group-type="heading"><subject>ARTIFICIAL INTELLIGENCE AND DECISION MAKING</subject></subj-group></article-categories><title-group><article-title>&lt;strong&gt;COMPARATIVE ANALYSIS OF DEEP LEARNING ALGORITHMS WITH REINFORCEMENT DDPG, PPO AND SAC FOR UNMANNED CAR CONTROL IN CARLA SIMULATOR&lt;/strong&gt;</article-title><trans-title-group xml:lang="en"><trans-title>&lt;strong&gt;COMPARATIVE ANALYSIS OF DEEP LEARNING ALGORITHMS WITH REINFORCEMENT DDPG, PPO AND SAC FOR UNMANNED CAR CONTROL IN CARLA SIMULATOR&lt;/strong&gt;</trans-title></trans-title-group></title-group><contrib-group><contrib contrib-type="author"><name-alternatives><name xml:lang="ru"><surname>Tikhonov</surname><given-names>Maksim Konstantinovich</given-names></name><name xml:lang="en"><surname>Tikhonov</surname><given-names>Maksim Konstantinovich</given-names></name></name-alternatives><email>samualgame@gmail.com</email></contrib></contrib-group><pub-date pub-type="epub"><year>2024</year></pub-date><volume>9</volume><issue>2</issue><fpage>0</fpage><lpage>0</lpage><self-uri content-type="pdf" xlink:href="/media/information/2024/2/ИТ_НР_9_2_8.pdf" /><abstract xml:lang="ru"><p>This paper presents a comparative analysis of three advanced deep reinforcement learning algorithms: Deep Deterministic Policy Gradient (DDPG), Proximal Policy Optimization (PPO) and Soft Actor-Critic (SAC) implemented in the Stable Baselines 3 library. The aim of the study is to evaluate the performance and applicability of each of the algorithms for the task of driving an unmanned vehicle in the complex and dynamic environment provided by the CARLA simulator, focusing on key metrics such as total distance, total reward, average speed, deviation from the center of the roadway, and success rate of episodes. The authors describe the experimental testing methodology in detail, including the tuning of training parameters and performance evaluation criteria. Experimental results demonstrate differences in the performance of the algorithms, revealing their strengths and weaknesses in the context of autonomous driving. The paper contributes to the understanding of the advantages and limitations of each algorithm in the context of autonomous driving and offers recommendations for their practical application.</p></abstract><trans-abstract xml:lang="en"><p>This paper presents a comparative analysis of three advanced deep reinforcement learning algorithms: Deep Deterministic Policy Gradient (DDPG), Proximal Policy Optimization (PPO) and Soft Actor-Critic (SAC) implemented in the Stable Baselines 3 library. The aim of the study is to evaluate the performance and applicability of each of the algorithms for the task of driving an unmanned vehicle in the complex and dynamic environment provided by the CARLA simulator, focusing on key metrics such as total distance, total reward, average speed, deviation from the center of the roadway, and success rate of episodes. The authors describe the experimental testing methodology in detail, including the tuning of training parameters and performance evaluation criteria. Experimental results demonstrate differences in the performance of the algorithms, revealing their strengths and weaknesses in the context of autonomous driving. The paper contributes to the understanding of the advantages and limitations of each algorithm in the context of autonomous driving and offers recommendations for their practical application.</p></trans-abstract><kwd-group xml:lang="ru"><kwd>deep reinforcement learning</kwd><kwd>autonomous driving</kwd><kwd>DDPG</kwd><kwd>PPO</kwd><kwd>SAC</kwd><kwd>Stable Baselines 3</kwd><kwd>CARLA</kwd></kwd-group><kwd-group xml:lang="en"><kwd>deep reinforcement learning</kwd><kwd>autonomous driving</kwd><kwd>DDPG</kwd><kwd>PPO</kwd><kwd>SAC</kwd><kwd>Stable Baselines 3</kwd><kwd>CARLA</kwd></kwd-group></article-meta></front><back><ref-list><title>Список литературы</title><ref id="B1"><mixed-citation>Lillicrap T.P. et al. Continuous control with deep reinforcement learning // arXiv preprint arXiv:1509.02971. &amp;ndash; 2015.</mixed-citation></ref><ref id="B2"><mixed-citation>Chang C.C. et al. Autonomous driving control using the ddpg and rdpg algorithms // Applied Sciences. &amp;ndash; 2021. &amp;ndash; Т. 11. &amp;ndash; №. 22. &amp;ndash; С. 10659.</mixed-citation></ref><ref id="B3"><mixed-citation>Schulman J. et al. Proximal policy optimization algorithms // arXiv preprint arXiv:1707.06347. &amp;ndash; 2017.</mixed-citation></ref><ref id="B4"><mixed-citation>Emuna R., Borowsky A., Biess A. Deep reinforcement learning for human-like driving policies in collision avoidance tasks of self-driving cars // arXiv preprint arXiv:2006.04218. &amp;ndash; 2020.</mixed-citation></ref><ref id="B5"><mixed-citation>Haarnoja T. et al. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor // International conference on machine learning. &amp;ndash; PMLR, 2018. &amp;ndash; P. 1861-1870.</mixed-citation></ref><ref id="B6"><mixed-citation>Ke P., Yanxin Z., Chenkun Y. A decision-making method for Self-driving based on deep reinforcement learning // Journal of Physics: Conference Series. &amp;ndash; IOP Publishing, 2020. &amp;ndash; Т. 1576. &amp;ndash; №. 1. &amp;ndash; P. 012025.</mixed-citation></ref><ref id="B7"><mixed-citation>Youssef F., Houda B. Comparative study of end-to-end deep learning methods for self-driving car // International Journal of Intelligent Systems and Applications. &amp;ndash; 2020. &amp;ndash; Т. 12. &amp;ndash; P. 15-27.</mixed-citation></ref><ref id="B8"><mixed-citation>Li D., Okhrin O. Modified DDPG car-following model with a real-world human driving experience with CARLA simulator // Transportation research part C: emerging technologies. &amp;ndash; 2023. &amp;ndash; Т. 147. &amp;ndash; P. 103987.</mixed-citation></ref></ref-list></back></article>