An Intelligent Congestion Control Strategy in Heterogeneous V2X Based on Deep Reinforcement Learning
Abstract
:1. Introduction
- Firstly, through an analysis of the heterogeneous V2X network architecture, a congestion control model based on deep reinforcement learning (DRL) is established using Markov’s random memoryless property;
- Secondly, we obtain the QoS parameters of different services in vehicle communications, calculate the minimum cost according to the importance of the service, and define the overhead weights and congestion sensitivity factor according to the different importance of services;
- In addition, by observing the current network state information, using the deep reinforcement learning PPO2 algorithm to learn from historical experience, and with the help of congestion-sensitive factors, a large amount of historical QoS data is used as a training set to optimize actions by combining the current state information of the network and selecting the congestion window size at the next moment. Thus, an intelligent congestion control strategy driven by QoS on-demand is formed;
- Finally, we build the ns-3 simulation platform to verify the performance of the ICCDRL proposed in the paper.
2. Research Background and Related Works
3. Intelligent Congestion Control Model Based on DRL in Heterogeneous V2X
3.1. Basic Model
3.2. Design of State Space
3.2.1. Size of The Congestion Window
3.2.2. Number of ACK Packets Fed Back
3.2.3. Round-Trip Time
3.2.4. Throughput
3.2.5. Packet Loss Rate
3.3. Transition Probability Matrix
3.3.1. Probability Distribution
3.3.2. On-Demand-Driven Congestion Sensitivity Factor Based on QoS
3.4. Design of Action Space
3.5. Reward Function
3.6. Policy Function
3.7. Description of the Algorithm ICCDRL
- Step 1:
- Input the initial state of the network and initialize the parameter of the policy function ;
- Step 2:
- Collect the actions corresponding to the state needed to run the policy function ;
- Step 3:
- At state , perform action to obtain reward ;
- Step 4:
- Obtain the expectations of the reward according to ;
- Step 5:
- Obtain the optimization objective function according to ;
- Step 6:
- Obtain the optimal value of the optimized objective function with the help of the gradient ascent method;
- Step 7:
- Execute action and update according to ;
- Step 8:
- Repeat step 2 through step 7.
4. Simulation Experiments and Result Analysis
4.1. Simulation Environment
4.2. Analysis of Simulation Experiments
4.2.1. Static Scenario
4.2.2. High-Speed Mobile Scenario
- (1)
- Comparison of Congestion window
- (2)
- Comparison of Bottleneck Link Utilization
- (3)
- Comparison of Throughput
- (4)
- Comparison of RTT
- (5)
- Comparison of Packet Loss
- (6)
- Evaluation of Fairness and Friendliness
- (7)
- Comparison of Convergence Speed
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
V2X | Vehicle to Everything |
V2V | Vehicle to Vehicle |
V2I | Vehicle to Infrastructure |
V2P | Vehicle to Pedestrian |
V2N | Vehicle to Network |
DRL | Deep Reinforcement Learning |
QoS | Quality of Service |
DSRC | Dedicated Short Range Communication |
C-V2X | Cellular Vehicle to Everything |
3GPP | 3rd Generation Partnership Project |
C-V2X | Cellular Vehicle-to-Everything |
CSMA | Carrier Sense Multiple Access |
TCP | Transmission Control Protocol |
ICCDRL | Intelligent Congestion Control Strategy Based on Deep Reinforcement Learning |
VANET | Vehicular Ad-Hoc Network |
IoV | Internet of Vehicles |
UBRCC | Utility-Based Rate Congestion Control |
HSR | Hierarchical State Routing |
UAVs | Unmanned Aerial Vehicle |
NDNs | Named Data Networking |
DRL-CCP | Deep Reinforcement Learning Congestion Control Protocol |
BDP | Bandwidth-delay product |
DCC | Dial Control Center |
SUMO | Simulation of Urban Mobility |
FTP | File Transfer Protocol |
BER | Bit Error Rate |
PPO2 | Proximal Policy Optimization |
DQN | Deep Q-Network |
SDN | Software Defined Network |
Hd-TCP | High-Speed TCP |
DL-TCP | Deep-Learning-Based TCP |
References
- Wang, R.; Deng, X.; Xu, Z. Survey on simulation testing and evaluation of Internet of vehicles. Appl. Res. Comput. 2019, 36, 1921–1926+1939. [Google Scholar] [CrossRef]
- Zhang, D.; Zhang, M.; Ding, F. Survey of connectivity for 5G-vehicular Ad Hoc networks. J. Nanjing Univ. Posts Telecommun. (Nat. Sci. Ed.) 2018, 38, 27–36. [Google Scholar] [CrossRef]
- Kenney, J.B. Dedicated Short-Range Communications (DSRC) Standards in the United States. Proc. IEEE 2011, 99, 1162–1182. [Google Scholar] [CrossRef]
- Rahim, N.-A.; Liu, Z.; Lee, H.; Ali, G.; Pesch, D.; Xiao, P. A Survey on Resource Allocation in Vehicular Networks. IEEE Trans. Intell. Transp. Syst. 2020, 23, 701–721. [Google Scholar] [CrossRef]
- Yousefi, S.; Mousavi, M.; Fathy, M. Vehicular ad hoc networks (VANETS): Challenges and Perspectives. In Proceedings of the 2006 6th International Conference on ITS Telecommunications, Chengdu, China, 21–23 June 2006; pp. 761–766. [Google Scholar]
- Huang, X.; Zhao, D.; Peng, H. Empirical Study of DSRC Performance based on Safety Pilot Model Deployment Data. IEEE Trans. Intell. Transp. Syst. 2017, 18, 2619–2628. [Google Scholar] [CrossRef] [Green Version]
- Xiong, K.; Leng, S.; Huang, C.; Yuen, C.; Guan, Y. Intelligent Task Offloading for Heterogeneous V2X Communications. IEEE Trans. Intell. Transp. Syst. 2021, 22, 2226–2238. [Google Scholar] [CrossRef]
- Shen, X.; Li, J.; Chen, L.; Chen, J.; He, S. Heterogeneous LTE/DSRC Approach to Support Real-time Vehicular Communications. In Proceedings of the 2018 10th International Conference on Advanced Infocomm Technology (ICAIT), Stockholm, Sweden, 12–15 August 2018; pp. 122–127. [Google Scholar]
- Mir, Z.H.; Toutouh, J.; Filali, F.; Ko, Y.B. Enabling DSRC and C-V2X Integrated Hybrid Vehicular Networks: Architecture and Pro-tocol. IEEE Access 2020, 8, 180909–180927. [Google Scholar] [CrossRef]
- Huang, W.; Ding, L.; Meng, D.; Wang, J.H.; Xu, Y.; Zhang, W. QoE-Based Resource Allocation for Heterogeneous Multi-Radio Communication in Software-Defined Vehicle Networks. IEEE Access 2018, 6, 3387–3399. [Google Scholar] [CrossRef]
- Zhang, H.; Li, T.; Li, D. Research on Vehicle Behavior Analysis Based Technologies for Intelligent Vehicular Networks. J. Electron. Inf. Technol. 2020, 42, 36–49. [Google Scholar]
- Andrews, J.G.; Buzzi, S.; Choi, W.; Hanly, S.; Lozano, A.; Soong, A.C.K.; Zhang, J.C. What Will 5G Be? IEEE J. Sel. Areas Commun. 2014, 32, 1065–1082. [Google Scholar] [CrossRef]
- Qiu, T.; Chen, N.; Li, K.; Qiao, D.; Fu, Z. Heterogeneous ad hoc networks: Architectures, advances and challenges. Ad Hoc Netw. 2017, 55, 143–152. [Google Scholar] [CrossRef]
- Li, R.; Zhao, Z.; Zhou, X.; Ding, G.; Chen, Y.; Wang, Z.; Zhang, H. Intelligent 5G: When Cellular Networks Meet Artificial Intelligence. IEEE Wirel. Commun. 2017, 24, 175–183. [Google Scholar] [CrossRef]
- Duan, X.Y.; Liu, Y.N.; Wang, X.B. SDN Enabled 5G-VANET: Adaptive Vehicle Clustering and Beamformed Transmission for Aggregated Traffic. IEEE Commun. Mag. 2017, 55, 120–127. [Google Scholar] [CrossRef]
- Roshdi, M.; Bhadauria, S.; Hassan, K.; Fischer, G. Deep Reinforcement Learning based Congestion Control for V2X Communication. In Proceedings of the 2021 IEEE 32nd Annual International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC), Virtual, 13–16 September 2021; pp. 1–6. [Google Scholar] [CrossRef]
- Choi, J.-Y.; Jo, H.-S.; Mun, C.; Yook, J.-G. Deep Reinforcement Learning-Based Distributed Congestion Control in Cellular V2X Networks. IEEE Wirel. Commun. Lett. 2021, 10, 2582–2586. [Google Scholar] [CrossRef]
- Yoon, Y.; Kim, H. Balancing Power and Rate Control for Improved Congestion Control in Cellular V2X Communication Environments. IEEE Access 2020, 8, 105071–105081. [Google Scholar] [CrossRef]
- Floyd, S.; Henderson, T. The NewReno Modification to TCP’s Fast Recovery Algorithm: RFC2582[A/OL]. Available online: https://dl.acm.org/doi/pdf/10.17487/RFC2582 (accessed on 1 March 2022).
- Ha, S.; Rhee, I.; Xu, L.S. Cubic: A new TCP-friendly high-speed TCP variant. ACM SIGOPS Oper. Syst. Rev. 2008, 42, 64–74. [Google Scholar] [CrossRef]
- Brakmo, L.S.; O’malley, S.W.; Peterson, L.L. TCP Vegas: New techniques for congestion detection and avoidance. In Proceedings of the Conference on Communication Architectures, Protocols and Application, New York, NY, USA, 31 August–2 September 1994; ACM: New York, NY, USA, 1994; pp. 24–35. [Google Scholar]
- Mascolo, S.; Casetti, C.; Gerla, M. TCP westwood: Bandwidth estimation for enhanced transport over wireless links. In Proceedings of the 7th Annual International Conference on Mobile Computing and Networking (MOBICOM), Rome, Italy, 16–21 July 2001; ACM: New York, NY, USA, 2001; pp. 287–297. [Google Scholar]
- Liu, B.; Hu, Z.; Wang, H. Distributed Control Strategy for Vehicular Networking Channel Congestion. J. Harbin Univ. Sci. Technol. 2020, 25, 12–18. [Google Scholar] [CrossRef]
- Sun, Y.; Xing, A. Congestion control strategy based on tabu search algorithm for vehicle ad hoc network MAC layer. J. Xi’an Univ. Posts Telecommun. 2017, 22, 15–21. [Google Scholar] [CrossRef]
- Tan, G.; Han, G.; Zhang, F. Distributed congestion control strategy using network utility maximization theory in VANET. J. Commun. 2019, 40, 82–91. [Google Scholar]
- Stoma, M.; Dudziak, A. The future of autonomous vehic1es in the opinion of automotive market users. Energies 2021, 14, 4777. [Google Scholar] [CrossRef]
- Csiszár, C.; Fldes, D. System Model for Autonomous Road Freight Transportation. Promet (Zagreb) 2018, 30, 93–103. [Google Scholar] [CrossRef] [Green Version]
- Martinez, A.; Caibano, E.; Romo, J. Analysis of Low Cost Communication Technologies for V2I Applications. Appl. Sci. 2020, 10, 1249. [Google Scholar] [CrossRef] [Green Version]
- Ye, H.; Li, G.Y.; Juang, B.H.F. Deep reinforcement learning based resource allocation for V2V communications. IEEE Trans. Veh. Technol. 2019, 68, 3163–3173. [Google Scholar] [CrossRef] [Green Version]
- Zhang, X.; Peng, M.; Yan, S. Deep reinforcement learning based mode selection and resource allocation for cellular V2X communi-cations. IEEE Internet Things J. 2019, 23, 2372–2385. [Google Scholar]
- Cui, L.; Yuan, Z.; Ming, Z. Improving the Congestion Control Performance for Mobile Networks in High-Speed Railway via Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2020, 69, 5864–5875. [Google Scholar] [CrossRef]
- Zhao, N.; Cheng, Y.; Liu, Z. Deep Reinforcement Learning-Based Channel Intelligent Access Method and NS3 Simulation. Comput. Simul. 2021, 38, 292–296. [Google Scholar]
- Xie, R.; Jia, X.; Wu, K. Adaptive Online Decision Method for Initial Congestion Window in 5G Mobile Edge Computing Using Deep Reinforcement Learning. IEEE J. Sel. Areas Commun. 2020, 38, 389–403. [Google Scholar] [CrossRef]
- Na, W.; Bae, B.; Cho, S. DL-TCP: Deep Learning-Based Transmission Control Protocol for Disaster 5G mmWave Networks. IEEE Access 2019, 7, 145134–145144. [Google Scholar] [CrossRef]
- Lan, D.; Tan, X.; Lv, J. A Deep Reinforcement Learning Based Congestion Control Mechanism for NDN. In Proceedings of the ICC 2019—2019 IEEE International Conference on Communications (ICC), Shanghai, China, 20–24 May 2019. [Google Scholar]
- Xiao, K.; Mao, S.; Tugnait, J.K. TCP-Drinc: Smart Congestion Control Based on Deep Reinforcement Learning. IEEE Access 2019, 7, 11892–11904. [Google Scholar] [CrossRef]
- Gawłowicz, P.; Zubow, A. ns3-gym: Extending openai gym for networking research. arXiv 2018, arXiv:1810.03943. [Google Scholar]
Parameter | Value | Parameter | Value |
---|---|---|---|
Scene size | 0.4 km2 | Modulation Technology | OFDM |
Scene Type | Two-way single lane | Packet Size(packet) | 50–100 MB |
Number of vehicles | 0–150 | Number of data flows | 5–20 |
Movement speed of nodes | 40–60 km/h | One-way time delay | 60 ms |
Channel Type | Wireless Channels | Wireless Random Error | 0.0001 |
Frequency | 5.9 GHz | Simulation time | 800 s |
Bottleneck Bandwidth | 200 Mbps | Data transfer rate | 60 Mbps |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, H.; Li, H.; Zhao, Y. An Intelligent Congestion Control Strategy in Heterogeneous V2X Based on Deep Reinforcement Learning. Symmetry 2022, 14, 947. https://doi.org/10.3390/sym14050947
Wang H, Li H, Zhao Y. An Intelligent Congestion Control Strategy in Heterogeneous V2X Based on Deep Reinforcement Learning. Symmetry. 2022; 14(5):947. https://doi.org/10.3390/sym14050947
Chicago/Turabian StyleWang, Hui, Haoyu Li, and Yuan Zhao. 2022. "An Intelligent Congestion Control Strategy in Heterogeneous V2X Based on Deep Reinforcement Learning" Symmetry 14, no. 5: 947. https://doi.org/10.3390/sym14050947