A Software-Defined Directional Q-Learning Grid-Based Routing Platform and Its Two-Hop Trajectory-Based Routing Algorithm for Vehicular Ad Hoc Networks
Abstract
:1. Introduction
1.1. Problem Background and Motivations
- Grid directionality and source-to-destination directionality for intergrid routing: By analyzing vehicle trajectories in the overall network environment, we define the number of outgoing vehicles from the current grid to the adjacent grid as the grid directionality of this adjacent grid in this paper. We believe that the grid directionality has an important impact on increasing the successful transmission rate and reducing the delay of the packet delivery. If you select a neighbor grid that is more in line with the vehicle’s driving direction, i.e., the grid with higher grid directionality, as the next-hop grid to forward the packet to, the packet has a higher probability of reaching the destination. In addition, the direction from the source node to the destination node, i.e., the source-to-destination directionality in this paper, also guides the best selection of the next-hop neighbor grid. If Q-learning selects a neighbor grid that has the higher grid and the source-to-destination directionalities, this grid will obtain a higher reward and Q-value, which are used as the decision basis of the intergrid routing. As shown in Figure 1, the starting position of the packet is at , and it will be sent to the destination at . Because the Q-value of is the highest one among all neighbor grids of , the intergrid routing for selects as the best next-hop grid. This means that the intergrid routing can use the grid directionality to train the Q-model and select the most plausible neighbor grid to forward the packets to.
- Two-hop directionality for intragrid routing: By analyzing the historical vehicle trajectories to determine the grid directionality of each grid, we can select the neighbor vehicle whose future driving trajectory goes to the best next-hop grid selected by the intergrid routing. However, there may be more than one possible neighbor vehicle that meets the above conditions; thus, how to make an appropriate choice is the important issue to address. As shown in Figure 2, vehicle node located in has two neighbor nodes, and , located in the best next-hop grid . Which of the two is the best relay node? If a neighbor node whose future driving trajectory moves to the best next-hop grid first, and then, to the best neighbor grid next, which is called the best two-hop next grid in this article, this node will be a better relay node than any node whose trajectory does not follow this two-hop path. This is because this relay node would have a higher probability to meet the destination node than others, even though it does not meet any neighbor node on this two-hop path to forward the packet to, but has to carry the packet by itself to the destination using SCF. Hence, this paper proposes two-hop trajectory-based routing (THTR) as follows. THTR first performs the intergrid routing on the best next-hop grid, i.e., , of to select the next best neighbor grid, i.e., . It then analyzes the historical trajectories of and located in the best next-hop grid, , to find their future positions. If will enter the best two-hop next grid at position , it would be a better strategy for to forward the packet to node instead of , because the future position of , that is, , enters later.
1.2. Contributions
- We design the Software-Defined Directional QGrid (SD-QGrid) network architecture, which combines the centralized control concept of the SDN Control Node (CN), and proposes the offline Q-learning training process and the online routing decision process in a V2X network architecture with reference to vehicle trajectories and three types of directionalities.
- Using the number of vehicles and the trajectory and directionality of vehicles between grids as parameters, the offline Q-learning training process of this paper modifies the Q-learning of reinforcement learning by increasing the moving directions between neighboring grids from the original four to eight and using the corresponding Q-tables for different time periods [22] as the macroreference for intergrid routing packet delivery.
- In the online routing decision process, a two-hop trajectory-based routing (THTR) algorithm is proposed, referring to the historical trajectory of the vehicle and the two-hop transmission path for calculating the future trajectory. This algorithm selects the neighbor node which is most suitable for the direction along the best next-hop grid and the best two-hop next grid, and then forwards packets to it, increasing the probability to meet the destination node and resulting in a shorter packet transmission path and lower packet transmission delay.
- We conduct simulation experiments to analyze and find the routing paths in V2X, and then obtain real and meaningful performance data of the routing algorithm using the real vehicle movement trajectories in Taipei City.
2. Related Work
- Buses: the trajectory is very accurate, with a relatively fixed arrival time and travel path.
- General private cars: have regular trajectories and obvious regularity in time and space.
- Taxis: have a flexible and variable travel route, and often have a credible destination [23], but the travel path is not fixed.
2.1. Position-Based Routing Algorithms without Adopting Reinforcement Learning
2.2. Position-Based Routing Algorithms with Reinforcement Learning
- (1)
- The ITAR-FQ [25] algorithm architecture consists of two main parts: the real-time traffic aware process and road evaluation (RTAP-RE) and the routing decision process with fuzzy Q-learning (RDP-FQ). The RTAP-RE designs a road evaluation method to process traffic information and estimate road quality (RQ), with reference to the number of vehicles moving in the same/a different direction, the length of the road, packet generation time, current time, etc. RDP-FQ divides the entire routing process into multiple routing processes. As the intersection is divided into multiple routing processes, the packet continues to look for the next intersection and selects a new road segment until the packet reaches the destination. Facing the intersection, it calculates the road score (RC) by combining RQ and the Manhattan distance (), and selects the road with the highest RC as the next forward road segment. The reward function is modified by referring to the benchmark reward and the corresponding value given by Fuzzy Logic. The advantage of this paper is that the Fuzzy Logic indicator is added to Q-learning for routing decisions in road segments. The disadvantage is that the directionality of the relative position of the transmitter and the destination are not considered. If the road segment with the destination end is the next hop at the intersection, it may cause the packet to be transmitted over a long distance or repeated on the road segment, and the historical movement trajectory of the vehicle is not used for training to obtain better routing decisions.
- (2)
- ADOPEL [14] assumes that two kinds of packets are exchanged between vehicle nodes: beacons and event-driven messages. The former is for the exchange of information such as the position, speed, and direction of travel between vehicles. The latter is used for vehicle nodes to collect traffic data and transmit it to the traffic control center (TCC), so that the TCC can gain a macroscopic understanding of the entire vehicle network environment, analyze the current number of neighbor vehicle nodes and the transmission delay of each node, and substitute this decision data into Q-learning, before using the Q-value to select better next-hop neighbor nodes for each vehicle node. The advantage of this approach is that it can better adapt to the high mobility and topology variability of the VANET network environment, whereas the disadvantage is that it does not refer to the future trajectory of the vehicle.
- (3)
- In the QTAR [15] network architecture, Q-learning is used for vehicle-to-vehicle (V2V) and RSU-to-RSU (R2R) networks. Decision elements, such as connection reliability and EED, are added to the Q-learning operation, and then, next-hop neighbor nodes are selected and packets are forwarded. The advantage of this approach is that the Q-learning planning method can improve throughput and PDR, but the disadvantage is that the directionality of the nodes is not considered, resulting in an unsatisfactory overall transmission performance when used in a real-world environment.
- (4)
- The QGrid [16] divides the entire map environment into equal grids, and uses Q-learning in advance. It uses the number of vehicles in the area as a parameter, and calculates the Q-table in advance to determine the transfer direction between grids, which is a macrotransfer consideration. QGrid_G uses the greedy method of packet transmission, which preferentially selects the neighbor node closest to the destination. QGrid_M uses the two-hop Markov prediction [26,27,28] method to predict the grid that the packet will pass through in the future, and preferentially selects the neighbor node with a higher probability of entering the next-hop grid as the next-hop node. The above two methods are microscopic transfer considerations. The advantage of these is that the transmission of packets is determined by considering the different macro- and microlevel aspects at the same time. Considering the number of vehicle nodes in each grid, the packets are preferentially transmitted to the grid with more vehicle nodes so that the packet loss rate is reduced. The disadvantage is that the vehicle density is only considered, and the influence of the overall movement direction between grids on the overall environment is not considered. The advanced QGrid is a routing protocol improvement made by QGrid, which is for vehicles with relatively regular and fixed trajectories, such as bus nodes, in the network environment. The vehicle node carries the packet to the destination grid and then continues the next-hop packet forwarding. The advantage of this is that it considers the reference vehicle trajectory, as well as the next-hop node selection, based on the trajectory information. The disadvantage is that even if the vehicle node has neighbor nodes, it still needs to continue to carry the packet until it reaches the grid destination where the node is located, resulting in a longer packet transmission delay time.
3. SD-QGrid Routing Platform and Algorithms
3.1. SD-QGrid Network Architecture
- The offline Q-learning training process: Using the vehicle trajectory and grid-related information, the offline Q-learning training process is responsible for training the Q-learning model to generate Q-tables of each grid at the SDN CN when the whole SD-QGrid system initializes or a new vehicle enters this V2X environment. The SD-QGrid may periodically perform this process to update Q-tables.
- The online routing decision process: Whenever a vehicle node carrying a packet intends to transmit this packet to the destination node, the SD-QGrid executes the online routing decision process to select the best neighbor node as the relay. Consequently, the packet will efficiently reach its destination at the end of the routing process.
3.2. Experimental Map
3.3. SD-QGRID Offline Q-Learning Training Process
- After the initialization of the SD-QGrid, all vehicles on a road segment issue HELLO messages to the corresponding RSUs.
- The RSU collects vehicle trajectory information from connected road segments and analyzes it to generate the historical trajectory tables.
- The RSU periodically transmits the analyzed tables and vehicle trajectories of the connected road segments to the SDN CN.
- The SDN CN aggregates the historical vehicle information of each grid from the information sent by its RSUs in the network environment.
- The SDN CN extracts important Q-learning parameters from the aggregated information of each grid. It finally uses Q-learning to train the Q-model for calculating the Q-table and Q-values of each grid.
3.3.1. RSU Analysis Unit
3.3.2. Aggregation Unit of SDN CN
3.3.3. Learning Unit of the SDN CN
- Step 1:
- Step 2:
- According to Table 7, the SDN CN calculates the ratio of the average number of outgoing nodes from the current node to each neighbor grid to the sum of the average number of outgoing nodes from the current node to all eight neighbor grids. Then, it proceeds to Step 3.
- Step 3:
- From Equation (1), Q-learning can use the discount rate to judge the packet transfer between grids. When the discount rate is higher, it means that the future reward from the next state will be higher. Relatively speaking, when the discount rate is higher, the future reward will be higher. Therefore, the discount rate is set that will be obtained by sending packets to neighboring grids in different directions according to the ratio of the average number of nodes per grid, and the ratio of the average directionality of each grid calculated above. We define the discount rate for selecting neighbor at time period as follows:
- Step 4:
- The SDN CN performs the Q-learning calculation with the calculated discount rate to generate the corresponding Q-table of each grid for different time periods, and records the Q-value of each grid for the eight neighbor grids in the Q-table database.
3.4. SD-QGrid Online Routing Decision Process
- The packet-carrying node issues a Neighbor Query packet to its RSU. This Neighbor Query packet contains vehicle IDs of the packet-carrying node, its neighbor nodes, and the destination node.
- As the RSU receives the Neighbor Query packet, it stores those vehicle IDs first. Then, it uses them to check whether historical trajectories of neighbor nodes of the packet-carrying node can be found on its trajectory cache and whether Q-tables of the grid where this RSU belongs to and the eight neighbor grids have been cached in it. If yes, it proceeds to Step 3. If not, it will notify the SDN CN to sends back only the missing historical trajectories of neighbor nodes and Q-tables of its associated grid and eight neighbor grids from the trajectory database and Q-table database, respectively. Then, these missing data are cached on the trajectory cache and Q-table cache of the RSU.
- From the Q-table of this grid, the routing decision unit of the RSU first selects the grid with the largest Q-value among the eight neighbor grids as the best next-hop grid.
- According to the best next-hop grid, Q-tables of the eight neighbor grids, and historical trajectories of the neighbor nodes of the packet-carrying node, the RSU then executes the two-hop trajectory-based routing (THTR) algorithm to select as the relay the neighbor node whose future driving trajectory continues along the best next-hop grid and the best two-hop next grid.
- Finally, the RSU issues the Neighbor Response packet containing the vehicle ID of the selected neighbor node to inform the packet-carrying node to forward the packet to this neighbor node.
3.4.1. The Link Expiration Time (LET) between Two Nodes
3.4.2. Routing Decision Unit of RSU
- Step 1:
- Determine whether the packet-carrying node is the destination node. If so, go to Step 9; if not, go to Step 2.
- Step 2:
- Determine whether has any neighbor node. If so, go to Step 3; if not, go to Step 8.
- Step 3:
- For some neighbor node of located in the neighbor grid , test whether the Q-value of the neighbor grid is higher than the Q-value of grid , where the vehicle is located. If so, go to Step 4; if not, go to Step 6.
- Step 4:
- If the moving direction of the neighbor node is known and it has a fixed route and schedule (such as a bus), go to Step 5; if not, go to Step 6.
- Step 5:
- Execute Algorithm 1. If the best next-hop node exists, go to Step 7; if not, go to Step 6.
- Step 6:
- Among all neighbor nodes in the neighbor grid with the higher Q-value than that of the current grid, select the neighbor node that is closest to the destination. If exists, go to Step 7; if not, go to Step 8.
- Step 7:
- Notify the packet-carrying node to use this node as the next-hop node and forward the packet to it. Go to Step 1.
- Step 8:
- Vehicle continues to hold the packet and waits for the next transmission opportunity. Go to Step 1.
- Step 9:
- If the current node is the destination node, the entire routing decision process ends.
Algorithm 1. Next-hop node selecting algorithm. |
Input:
|
Algorithm 2. One-Two-Hop Value (Node A, Node X) |
Input:
|
4. Performance Evaluation
4.1. Simulation Environment and Parameters
- Advanced QGrid [16]: QGrid improves the routing protocol for vehicle nodes with relatively fixed trajectories, such as buses. If the vehicle node passes through the grid where the packet destination is located in the future, it will continue to carry the packet until it enters the grid, and then proceed with next-hop routing selection to transfer the packet.
- GPSR [33]: A position-based routing protocol that continuously forwards packets from the nearest neighbor to the destination until it reaches the destination.
- (a)
- Delivery ratio: the ratio of the successful arrival of message packets to the total number of message packets generated.
- (b)
- Average end-to-end delay: how long it takes, on average, for a message packet to travel from the source to the destination.
- (c)
- Overhead: the ratio of the total number of forwarded message packets to the total number of originally sent message packets.
4.2. Simulation Results
5. Conclusions and Future Directions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Chen, S.; Hu, J.; Shi, Y.; Peng, Y.; Fang, J.; Zhao, R.; Zhao, L. Vehicle-to-Everything (v2x) Services Supported by LTE-Based Systems and 5G. IEEE Commun. Stand. Mag. 2017, 1, 70–76. [Google Scholar] [CrossRef]
- Zhang, D.; Zhang, T.; Liu, X. Novel Self-Adaptive Routing Service Algorithm for Application in VANET. Appl. Intell. 2019, 49, 1866–1879. [Google Scholar] [CrossRef]
- Mehta, K.; Malik, L.G.; Bajaj, P. VANET: Challenges, Issues and Solutions. In Proceedings of the International Conference on Emerging Trends in Engineering and Technology, ICETET, Nagpur, India, 16–18 December 2013; pp. 78–79. [Google Scholar]
- Cooper, C.; Franklin, D.; Ros, M.; Safaei, F.; Abolhasan, M. A Comparative Survey of VANET Clustering Techniques. IEEE Commun. Surv. Tutor. 2017, 19, 657–681. [Google Scholar] [CrossRef]
- Nazib, R.A.; Moh, S. Routing Protocols for Unmanned Aerial Vehicle-Aided Vehicular Ad Hoc Networks: A Survey. IEEE Access 2020, 8, 77535–77560. [Google Scholar] [CrossRef]
- Kevin, F. A Delay-Tolerant Network Architecture for Challenged Internets. In Proceedings of the 2003 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications, Karlsruhe, Germany, 25 August 2003. [Google Scholar]
- Ohta, Y.; Ohta, T.; Kohno, E.; Kakuda, Y. A Store-Carry-Forward-Based Data Transfer Scheme Using Positions and Moving Direction of Vehicles for VANETs. In Proceedings of the 2011 Tenth International Symposium on Autonomous Decentralized Systems, Tokyo, Japan, 23–27 March 2011. [Google Scholar]
- Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
- Dowling, J.; Curran, E.; Cunningham, R.; Cahill, V. Using Feedback in Collaborative Reinforcement Learning to Adaptively Optimize MANET Routing. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2005, 35, 360–372. [Google Scholar] [CrossRef] [Green Version]
- Doddalinganavar, S.S.; Tergundi, P.V.; Patil, R.S. Survey on Deep Reinforcement Learning Protocol in VANET. In Proceedings of the 1st IEEE International Conference on Advances in Information Technology (ICAIT 2019), Chikmagalur, India, 25–27 July 2019; pp. 81–86. [Google Scholar]
- Liu, J.; Wang, Q.; He, C.; Jaffrès-Runser, K.; Xu, Y.; Li, Z.; Xu, Y. QMR:Q-learning based Multi-objective optimization Routing protocol for Flying Ad Hoc Networks. Comput. Commun. 2019, 150, 304–316. [Google Scholar] [CrossRef]
- Coutinho, W.P.; Battarra, M.; Fliege, J. The Unmanned Aerial Vehicle Routing and Trajectory Optimization Problem, a Taxonomic Review. Comput. Ind. Eng. 2018, 120, 116–128. [Google Scholar] [CrossRef] [Green Version]
- Nazib, R.A.; Moh, S. Reinforcement Learning-Based Routing Protocols for Vehicular Ad Hoc Networks: A Comparative Survey. IEEE Access 2021, 9, 27552–27587. [Google Scholar] [CrossRef]
- Soua, A.; Afifi, H. Adaptive Data Collection Protocol using Reinforcement Learning for VANETs. In Proceedings of the 2013 9th International Wireless Communications and Mobile Computing Conference (IWCMC 2013), Sardinia, Italy, 1–5 July 2013; pp. 1040–1045. [Google Scholar]
- Wu, J.; Fang, M.; Li, H.; Li, X. RSU-Assisted Traffic-Aware Routing based on Reinforcement Learning for Urban VANETs. IEEE Access 2020, 8, 5733–5748. [Google Scholar] [CrossRef]
- Li, F.; Song, X.; Chen, H.; Li, X.; Wang, Y. Hierarchical Routing for Vehicular Ad Hoc Networks via Reinforcement Learning. IEEE Trans. Veh. Technol. 2019, 68, 1852–1865. [Google Scholar] [CrossRef]
- Watkins, C.J.C.H.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
- Kim, H.; Feamster, N. Improving network management with software defined networking. IEEE Commun. Mag. 2013, 51, 114–119. [Google Scholar] [CrossRef]
- Wickboldt, J.A.; de Jesus, W.P.; Isolani, P.H.; Both, C.B.; Rochol, J.; Granville, L.Z. Software-Defined Networking: Management Requirements and Challenges. IEEE Commun. Mag. 2015, 53, 278–285. [Google Scholar] [CrossRef]
- Hans, C.Y.; Quer, G.; Rao, R.R. Wireless SDN Mobile Ad Hoc Network: From Theory to Practice. In Proceedings of the 2017 IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017; pp. 1–7. [Google Scholar]
- Zhao, Z.; Cumino, P.; Souza, A.; Rosário, D.; Braun, T.; Cerqueira, E.; Gerla, M. Software-defined unmanned aerial vehicles networking for video dissemination services. Ad Hoc Netw. 2018, 83, 68–77. [Google Scholar] [CrossRef]
- Chang, I.-C.; Hung, M.-H.; Chang, C.-R.; Yen, C.-E. NTR: An Efficient Trajectory-based Routing Protocol for the Vehicular Delay Tolerant Networks. In Proceedings of the 2017 IEEE International Conference on Systems, Man, and Cybernetics, Banff, AB, Canada, 5–8 October 2017. [Google Scholar]
- Feng, F.W.Y.; Liao, J.; Qian, Q. Driving Path Prediction Based Routing Protocol in Vehicular Ad Hoc Networks. Int. J. Distrib. Sens. Netw. 2013, 9, 837381. [Google Scholar] [CrossRef]
- Dijkstra, E.W. A Note on Two Problems in Connection with Graphs. Numer. Math. 1959, 1, 269–271. [Google Scholar] [CrossRef] [Green Version]
- He, Y.; He, R.; Yu, C. Intersection-based Traffic-Aware Routing with Fuzzy Q-learning for Urban VANETs. In Proceedings of the 2021 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC), Chengdu, China, 18–20 June 2021. [Google Scholar]
- Wang, E.; Yang, Y.; Jia, B.; Guo, T. The DTN Routing Algorithm based on Markov Meeting Time Span Prediction Model. Int. J. Distrib. Sens. Netw. 2013, 9, 736796. [Google Scholar] [CrossRef] [Green Version]
- Ip, Y.-K.; Lau, W.-C.; Yue, O.-C. Forwarding and Replication Strategies for DTN with Resource Constraints. In Proceedings of the 2007 IEEE 65th Vehicular Technology Conference, Dublin, Ireland, 22–25 April 2007; pp. 1260–1264. [Google Scholar]
- Liu, S.; Li, F.; Zhang, Q.; Shen, M. A Markov Chain Prediction Model for Routing in Delay Tolerant Networks. In International Conference on Big Data Computing and Communications; Springer: Cham, Switzerland, 2015; Volume 9196, pp. 479–490. [Google Scholar]
- Luo, L.; Sheng, L.; Yu, H.; Sun, G. Intersection-Based V2X Routing via Reinforcement Learning in Vehicular Ad Hoc Networks. IEEE Trans. Intell. Transp. Syst. 2022, 23, 5446–5459. [Google Scholar] [CrossRef]
- Available online: https://ptx.transportdata.tw/PTX/ (accessed on 20 October 2022).
- Taleb, T.; Sakhaee, E.; Jamalipour, A.; Hashimoto, K.; Kato, N.; Nemoto, Y. A Stable Routing Protocol to Support ITS Services in VANET Networks. IEEE Trans. Veh. Technol. 2007, 56, 3337–3347. [Google Scholar] [CrossRef]
- Riley, G.F.; Henderson, T.R. The NS-3 Network Simulator. In Modeling and Tools for Network Simulation; Springer: Berlin/Heidelberg, Germany, 2010; pp. 15–34. [Google Scholar]
- Karp, B.; Kung, H.-T. GPSR: Greedy Perimeter Stateless Routing for Wireless Networks. In Proceedings of the 6th Annual International Conference on Mobile Computing and Networking, Boston, MA, USA, 1 August 2000; ACM: New York, NY, USA, 2000; pp. 243–254. [Google Scholar]
Protocol | Analyzes Real Vehicle Trajectory Data | Considers the Directionality of the Vehicle Trajectory | Considers the Directionality of Source to Destination | Considers the Number of Vehicles in Each Area | Uses Reinforcement Learning for Routing Decisions | Simulation Experiments Using Real Vehicle Trajectories | Considers Two-Hop Next-Grid Routing | Designs Software-Defined Routing Platform |
---|---|---|---|---|---|---|---|---|
ITAR-FQ | No | No | No | Yes | Yes | No | No | No |
ADOPEL | No | Yes | No | Yes | Yes | No | No | No |
QTAR | No | No | No | Yes | Yes | No | No | No |
QGrid_G | No | No | No | Yes | Yes (only one Q-table, four neighbor grids) | Yes | No | No |
AdvQGrid | Yes | No | No | Yes | Yes (only one Q-table, four neighbor grids) | Yes | No | No |
SD-QGrid | Yes | Yes | Yes | Yes | Yes (different Q-tables corresponding to different time periods, eight neighbor grids) | Yes | Yes | Yes |
Time Period | Congestion Level |
---|---|
23:00–07:00 | normal |
07:00–09:00 | congested |
09:00–12:00 | normal |
12:00–13:00 | congested |
13:00–17:00 | normal |
17:00–19:00 | congested |
19:00–22:00 | normal |
22:00–23:00 | congested |
Vehicle ID | Time Stamp | Location Lon | Location Lat |
---|---|---|---|
0 | weekday_07:00:00 + 08:00, | 121.50853 | 25.04262 |
0 | weekday_07:00:20 + 08:00, | 121.5328 | 25.034475 |
0 | weekday_07:00:40 + 08:00 | 121.533667 | 25.044598 |
… | … | … | … |
0 | weekday_08:20:20 + 08:00 | 121.525173 | 25.052125 |
Grid Index | The Start Intersection Coordinate (x, y) of Road Segment i | The End Intersection Coordinate (x, y) of Road Segment i | ||
---|---|---|---|---|
(0, 2) | (xs, ys) | (xe, ye) | 314 | 2021-10-13–10-12 |
Current Grid Index | Neighbor Grid Index | The Start Intersection Coordinate (x, y) of Road Segment i | The End Intersection Coordinate (x, y) of Road Segment i | Average Number of Outgoing Nodes | |
---|---|---|---|---|---|
(0, 2) | (0, 3) | (xs, ys) | (xe, ye) | 110 | 2021-10-13–10-12 |
Grid Index | ||
---|---|---|
(0, 2) | 314 | 2021-10-13–10-12 |
Average Grid Directionality for Eight Neighbor Grids | ||
---|---|---|
(0, 2) | 134/143/506/197/10/5/53/5 | 2021-10-13–10-12 |
Parameters | Parameter Value or Range |
---|---|
0.8 | |
[0.3, 0.9] | |
0.1 | |
3 | |
1 | |
Reward | 0, 100 |
Experimental map range | 4000 m × 6000 m |
Experiment time | 3000 s |
MAC protocol | IEEE 802.11 p |
Radio propagation model | Log Distance Propagation Loss Model |
Buffer size | 10 MB |
Bandwidth | 11 Mbps |
Transmission range (m) | 400, 450, 500, 550, 600 |
Grid size | 1000 m |
TTL (s) | 10, 20, 30, 40, 50 |
numPair | 50, 100, 150, 200, 250 |
Beacon time (s) | 1 |
Message time (s) | 1, 5, 10, 15, 20 |
Average Delivery Ratio | Average End-to-End Delay | Average Transmission Overhead | |
---|---|---|---|
QGrid | +17.00% | −27.09% | +1.04% |
advQGrid | +10.22% | −31.93% | +1.24% |
Average Delivery Ratio | Average End-to-End Delay | Average Transmission Overhead | |
---|---|---|---|
QGrid | +21.01% | −24.70% | +1.09% |
advQGrid | +13.66% | −34.00% | +1.53% |
Average Delivery Ratio | Average End-to-End Delay | Average Transmission Overhead | |
---|---|---|---|
QGrid | +13.66% | −27.67% | +0.51% |
advQGrid | +9.27% | −35.26% | +1.01% |
Average Delivery Ratio | Average End-to-End Delay | Average Transmission Overhead | |
---|---|---|---|
QGrid | +16.47% | −28.65% | +1.42% |
advQGrid | +10.77% | −35.08% | +2.31% |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yang, C.-P.; Yen, C.-E.; Chang, I.-C. A Software-Defined Directional Q-Learning Grid-Based Routing Platform and Its Two-Hop Trajectory-Based Routing Algorithm for Vehicular Ad Hoc Networks. Sensors 2022, 22, 8222. https://doi.org/10.3390/s22218222
Yang C-P, Yen C-E, Chang I-C. A Software-Defined Directional Q-Learning Grid-Based Routing Platform and Its Two-Hop Trajectory-Based Routing Algorithm for Vehicular Ad Hoc Networks. Sensors. 2022; 22(21):8222. https://doi.org/10.3390/s22218222
Chicago/Turabian StyleYang, Chen-Pin, Chin-En Yen, and Ing-Chau Chang. 2022. "A Software-Defined Directional Q-Learning Grid-Based Routing Platform and Its Two-Hop Trajectory-Based Routing Algorithm for Vehicular Ad Hoc Networks" Sensors 22, no. 21: 8222. https://doi.org/10.3390/s22218222
APA StyleYang, C. -P., Yen, C. -E., & Chang, I. -C. (2022). A Software-Defined Directional Q-Learning Grid-Based Routing Platform and Its Two-Hop Trajectory-Based Routing Algorithm for Vehicular Ad Hoc Networks. Sensors, 22(21), 8222. https://doi.org/10.3390/s22218222