Next Article in Journal
A Multi-Rate Simulation Strategy Based on the Modified Time-Domain Simulation Method and Multi-Area Data Exchange Method of Power Systems
Next Article in Special Issue
On the Detection of Vulnerable Road Users—Simulating Limits of an FMCW Radar
Previous Article in Journal
Improvement in Laptop Heat Dissipation with Taguchi Method
Previous Article in Special Issue
Single-Timestamp Skew Correction (STSC) in V2X Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient V2V Communications by Clustering-Based Collaborative Caching †

Graduate School of Informatics and Engineering, The University of Electro-Communications, Chofu, Tokyo 182-8585, Japan
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in Tokunaga, H.; Tang, S. Efficient V2V communications by clustering-based collaborative caching. In Proceedings of the 2022 IEEE International Conference on Vehicular Electronics and Safety (ICVES), Bogota, Colombia, 14–16 November 2022; pp. 1–6, doi: 0.1109/ICVES56941.2022.9986927.
Electronics 2024, 13(5), 883; https://doi.org/10.3390/electronics13050883
Submission received: 14 December 2023 / Revised: 31 January 2024 / Accepted: 22 February 2024 / Published: 25 February 2024
(This article belongs to the Special Issue Advances in Vehicular Ad Hoc Networks (VANETs))

Abstract

:
Vehicle-to-vehicle (V2V) communication plays an important role in enabling autonomous driving. However, when multiple vehicles request the same content, like road conditions, delivering it individually by V2V communication can significantly increase traffic volume, potentially causing congestion in the wireless channel. To address this issue, Content-Centric Network (CCN) technology is applied to V2V communication, which improves communication efficiency by exploiting content cached at vehicles. However, previous methods faced the following challenges: (i) vehicles could not use content stored in nearby vehicles outside the communication path, and (ii) redundant caching of the same content occurred at nearby vehicles. To tackle these challenges, this paper proposes a collaborative caching method in which vehicles are grouped into clusters and each cluster has a designated head responsible for managing caches across all vehicles within the cluster. In this way, this method enables vehicles to use the content cached at adjacent vehicles that are not directly on a communication path. In addition, it eliminates redundant caches, allowing a more diverse range of content storage. Extensive simulation results demonstrate that the proposed approach effectively reduces content delivery latency by 33% compared to the method using clusters without cooperative caching and by 19% compared to the ECV+ method.

1. Introduction

Vehicle-to-Vehicle (V2V) communication stands out as a pivotal technology in the field of autonomous driving [1]. The exchanged content can be categorized into two types: content related to safety and content related to comfort. The former typically has a brief lifespan and is seldom requested again. In comparison, the latter has a longer lifespan, and various vehicles may request the same information multiple times. When this type of content is delivered through separate pathways, the traffic volume increases rapidly with the growing number of requesting vehicles.
The Content-Centric Network (CCN) architecture [2] is specifically designed for the repeated delivery of identical content. In CCN, nodes use interest packets to request content and data packets to deliver it. Each node stores received or forwarded content as caches, enabling them to respond to requests instead of content providers. This approach effectively reduces communication delays and network traffic, making CCN widely adopted in wired networks. Recently, there has been a growing interest in applying CCN to vehicular networks [3,4,5], where a vehicle is treated as a CCN node.
There are limitations and issues with the basic CCN, however.
(1)
This approach cannot use content cached at nearby nodes that are not on communication routes.
(2)
Nodes autonomously decide whether to cache content, leading to the possibility of multiple vehicles storing the same information. This consumes cache buffers uselessly and hampers the caching of other content.
(3)
The mobility of vehicles makes it complex to establish concrete communication paths.
Collaborative caching provides an effective solution to address (1) and (2). As for (1), Ref. [6] solves it by letting vehicles periodically broadcast beacons informing the names of cached content with great overhead due to the extensive communication cost. As for (2), Refs. [3,7,8] suggest that each vehicle make a probabilistic decision on whether to cache content. While this reduces the redundancy of storing identical content in nearby vehicles, it does not completely eliminate it.
To further improve the performance of vehicular networks by CCN, this paper proposes a novel collaborative caching method. This method groups vehicles into clusters and manages caches in each cluster unitarily. Specifically, each cluster comprises a cluster head (CH) and several cluster members (CM). A CH is responsible for dealing with content requests and managing content in its cluster. When a CH receives a content request, it searches for the requested content in its cluster, enabling the CH to respond to requests based on caches from any vehicles within its cluster. When a CH receives content, it determines whether to store it in its cluster and if it does, it also decides which vehicle in its cluster stores it, which prevents duplicated content from being stored in a cluster.
The contributions of this paper are two-fold as follows:
  • To the best of our knowledge, this is the first work that proposes to consolidate content management within a cluster to optimize the utilization of cached content and cache buffers in V2V communications.
  • We implement unitary cache management by letting each vehicle overhear data packets, which suppresses the control overhead.
Part of this paper has been presented in a conference [9]. In this work, we added the estimation of content popularity and on this basis refined cache management. In addition, we evaluated the proposed method in more scenarios: both the freeway scenario used in [9] and an urban model scenario that has been newly added in this paper. We also added a comparison with ECV+ [3]. Extensive simulation evaluations demonstrated that the proposed method effectively increased the cache hit rate within a cluster and improved network efficiency.
The rest of this paper is organized as follows. Section 2 reviews the basic CCN and related works. Section 3 presents the proposed method, and Section 4 confirms its effectiveness by simulation evaluation under different settings. Finally, Section 5 concludes this paper.

2. Related Work

Here, we introduce the basic CCN and review related work on clustering algorithms, collaborative caching methods, and routing protocols for vehicular networks. Table 1 shows a brief explanation of each related method.

2.1. Basic CCN [2]

When a node requests content in CCN, it sends an interest packet containing the content name towards the content provider. If a node receiving the interest packet has the requested content, it replies with a data packet containing the content. Otherwise, it records the content name and the ID of the requesting node in its Pending Information Table (PIT) and forwards the interest packet towards the content provider according to its Forwarding Information Base (FIB). When a node receives a data packet, it stores the content included in the packet in the Content Store (CS) and forwards the packet towards the requesting node according to its PIT. Generally, CCN was designed for wired networks and has no node mobility considerations.

2.2. Cluster Construction

Clustering is a typical method of node management in vehicular networks and thus is a target of research [10,11]. These methods, however, are designed for general vehicular communication and are not optimized for CCN.

2.3. Collaborative Caching

Ref. [4] studies a combination of CCN and clustering whereby a CH stores content received from a Road-Side Unit (RSU), making it possible to reply to requests from CMs even if they are outside of the transmission ranges of RSUs. However, this method does not use cache buffers at CMs at all.
In the cooperative scheduling method [12], each RSU on a road divides vehicles in its transmission range into clusters in a centralized way. In each cluster, each vehicle takes turns to broadcast content according to the specified schedule. Although this helps to improve channel efficiency, it is not applicable to the scenario wherein vehicles request different content.
Refs. [3,7,8] studied how to reduce the redundancy in content caching by storing received content with a probability. These methods help to reduce the chance that neighbor vehicles save duplicated content, but they cannot prevent it entirely.
In [13], vehicles decide whether to cache content based on the vehicle’s power, content popularity, the gain acquired by caching the content, and the hop count of the content.
In [14], both vehicles and RSUs form clouds: a vehicular cloud (VC) and an RSU cloud (RC), respectively. When a vehicle requests content, it first asks its VC. If all vehicles in the cloud do not have the content, it queries its associated RSU. The RSU replies if one of its RCs has the content or fetches the content from the remote server otherwise.
Some research has applied existing algorithms to V2V communications, e.g., a cuckoo search for route calculation [15], the 0–1 knapsack problem for cache decisions [16], and the Minimum Weight Vertex Cover Problem for cache placement [20].
With the emergence of AI, some studies suggest combining AI and vehicular networks: Ref. [17] calculates the cache policy with deep reinforcement learning, Ref. [18] uses machine learning to decide which content to store and where to store it, and Ref. [19] uses reinforcement learning to decide which vehicle would be the best data carrier node.
To share content information, Ref. [6] suggests letting each vehicle broadcast the names of its cached content to neighbor vehicles by using periodical beacon messages. Neighbor vehicles, on receiving this message, store the information and use it to check which vehicle stores content as a cache. However, the notification of content information via beacons causes huge overhead and increases the possibility of packet collision.
Ref. [21] targets a stationary environment wherein each node requests a time-shifted TV program. Each content router (CR) has a label n of a positive integer, where n < k . CR n stores only chunks c, where c mod k = n , which prevents multiple nodes from caching the same content. It is difficult, however, to apply this method in the V2V environment where nodes move.

2.4. Determining Request/Relay Vehicle

Some research has discussed how to select a vehicle to request or relay content. In [22], a vehicle that lies in the common communication area of a packet sender and other potential relay vehicles is preferentially selected. In [5], the content requester requests content from vehicles moving in the same lane first and then to vehicles moving in adjacent lanes upon failure.

2.5. Comparison with Related Work

The purpose of this paper is not to construct clusters but to make cache management more efficient by using the clusters. Each CH manages content in its cluster unitarily, which solves the problem of cache duplication that was a challenge in previous works [3,7]. In this way, vehicles can cache more content in a cluster, which enables CHs to reply to diverse requests locally instead of further fetching them from other vehicles or content providers.

3. The Proposed Method

Figure 1 shows the overview of the proposed method. It mainly consists of four components as follows:
  • Cluster construction: Vehicles form clusters autonomously based on their moving speeds and directions. Each cluster has a cluster head (CH) and a variable number of cluster members (CMs).
  • Content fetching: On requesting content, each CM sends an interest packet to its associated CH. Each CH is responsible for forwarding interest and data packets and replies to an interest packet if the requested content is found in its cluster.
  • Popularity management: Vehicles update the popularity of the content by monitoring requests from its cluster and the interest packets it forwards.
  • Cache management: Each CH manages cache buffers for itself and all its CMs. On receiving a data packet, it decides whether and where to store the content. If there is a shortage of cache buffers, the content with the lowest popularity is replaced by the new content.
Figure 1. The overview of the proposed method.
Figure 1. The overview of the proposed method.
Electronics 13 00883 g001
With these steps, the proposed method removes duplicated caches and enables each cluster to cache more content locally. On this basis, it reduces the number of packet transmissions and the response time accordingly.

3.1. Cluster Construction and Maintenance

3.1.1. Detecting Neighbors by Exchanging Beacons

Each vehicle periodically broadcasts a beacon to alert nearby vehicles of its presence. Before broadcasting a beacon, the sender removes from its neighbor list all neighbor vehicles for which their information has not been updated within t neighbor . Each beacon contains the senders’ vehicle ID, position, speed, the number of neighbor vehicles, a flag indicating whether the sender is a CH, the ID of the associated CH if the sender is a CM, and the connection expired predicted (CEP) CH ID if the sender vehicle is a CH. When a vehicle receives a beacon, it records the sender as its neighbor vehicle.

3.1.2. CM Joining/Leaving

Every vehicle is equipped with a CM table designed to handle CM information in case it becomes a CH later. The CM table comprises two essential components for each CM entry: (i) a list of content with names and timestamps that denote the reception time and (ii) a cluster-keep-alive timer. This timer is refreshed upon receiving beacons from the CM and triggers the removal of the CM entry if it expires.
Each vehicle initially assumes the role of a CH. A CH without any CMs in its cluster is identified as an Orphan Vehicle (OV). When an OV receives a beacon from another CH, it transmits a beacon claiming its joining into the CH’s cluster as a CM if specific conditions are satisfied as follows:
  • Both the beacon sender and receiver move in the same direction.
  • The distance between the two vehicles is less than or equal to d c l u s t e r .
  • The relative velocity between the two vehicles is less than or equal to v c l u s t e r .
When a CH receives a beacon from a CM for the first time, it adds the CM’s information to its CM table. When the CH receives a beacon from the CM again, it resets the CM’s cluster-keep-alive timer. When a CM receives a beacon from its CH, it resets its cluster-keep-alive timer.
When one of the following conditions is satisfied, a CM leaves its cluster, and its CH considers that the CM has left its cluster.
  • The cluster-keep-alive timer expires at either a CM or a CH.
  • The distance between a CH and a CM is more than d c l u s t e r .

3.1.3. Splitting/Merging Clusters

If the distance between two adjacent CHs changes drastically, it is necessary to split a cluster into two (when the distance is too large) to avoid relay failure or to merge two clusters (when the distance is too short) to improve relay efficiency.
To this end, a CEP CH ID is added to the beacons sent from CHs. A CEP CH is a neighbor CH of the sender CH that satisfies the following conditions.
  • The CEP CH, as a relay of the sender CH, has the shortest distance to the CH in either the front or the back of the CH.
  • The connection between the CEP CH and the sender CH is predicted to break within t break .
A beacon may contain at most two CEP CH IDs to notify the potential relay failure of the sender CH briefly. Figure 2 shows an example, where CH1, predicting that the connection between itself and CH2 will break soon, sends a beacon containing CH2’s ID as a CEP CH ID.
When a CM receives a beacon from any CH containing a CEP CH ID and it can communicate with both the sender CH and the CEP CH, it tries to leave its cluster immediately and create a new one. Because multiple CMs simultaneously receive a beacon with a CEP CH ID, each CM calculates its CH suitability according to (1). Here, N i denotes the number of neighbors of vehicle i, and d ¯ i is the average distance from vehicle i to its neighbors.
s i = N i d ¯ i .
Then, the vehicle with the largest s becomes a new CH, and other vehicles choose to become its CMs or remain as the CMs of the previous clusters. Figure 3 shows that CH3 constructs a new cluster based on information in a beacon following the actions in Figure 2.
Two clusters start to merge when all of the following conditions are satisfied.
  • The CHs of both clusters are moving in the same direction.
  • A CH can communicate with all of the CHs that the other can.
  • The distance between the two CHs is less than or equal to d cluster .
  • The CH has not sent a merge request to another CH.
When CH1 decides to merge its cluster with CH2’s, it sends a merge request to CH2 and schedules a merge-cancel timer. CH2 accepts the request if the following conditions are satisfied and ignores the request otherwise.
  • The cluster merging requirements are satisfied.
  • CH2 is not requesting to merge clusters with another CH.
Once CH2 accepts the request, it selects a new CH from CH1 and CH2 by calculating both of the CH’s CH suitability according to (1) and choosing the CH with higher suitability. Then, it sends CH1 a merge ACK packet containing the information on whether CH2 becomes a CH or CM. All CMs are notified of the cluster merging by CH’s beacon or the merge ACK packet.
Cluster merging is canceled if the merge-cancel timer expires.

3.2. Content Fetching

Here, we explain how content is fetched in the proposed method using Figure 4.
When a CM requests content, it sends an interest packet containing the name of the requested content to its CH (step 1 in the figure). The CH tries to reply to the request in the following order.
  • If the CH itself has the requested content, it replies with a data packet.
  • If one of its CMs has the content, the CH adds the ID of the request vehicle to the interest packet and forwards it to the CM that has the cached content (step 4 in the figure).
  • If no vehicle in its cluster has the requested content, the CH first tries to forward the interest packet to other CHs according to its FIB (steps 2 and 3 in the figure).
When a CM receives an interest packet from its CH and has the requested content, it sends a data packet to the requester (step 5 in the figure). The CM communicates with the requester directly if it can, and otherwise, it relays the content via its CH. Due to potential packet loss in the wireless environment, a CM may not have the requested content. In such cases, the CM sends a content-not-found message to its CH. Then, the CH, accordingly, forwards the interest packet towards the content provider via other CHs.
A CH, on receiving a data packet, forwards it to the content requester according to the PIT (steps 6 and 7 in the figure).
In some cases, there may be no CH available as a forwarder. A CH forwards an interest packet to the CM that is closest to the content provider, using this CM as a relay. When a CM receives an interest packet from a CH outside of its cluster, it sends the interest packet to its CH which further forwards this towards the content provider.
Algorithm 1 is the pseudo-code for interest and packet handling.
Algorithm 1: Packet management
Electronics 13 00883 i001

3.3. Content Popularity Inference

Each vehicle has a request counter table (Figure 5) to count how many times a content is requested to infer the content’s popularity. It involves both requests from inside the cluster and the requests from other clusters that pass this cluster. Interest packets and data packets contain a timestamp. This timestamp, together with the content name, serves as a unique Request ID. When a vehicle receives one of these packets, it checks whether the Request ID in this packet is fresher than the one in the corresponding entry of the table, and if affirmative, it increases the counter by 1 and updates the Request ID. In this way, the counter of the content reflects its popularity.
Algorithm 2 is the pseudo-code for updating content’s popularity.
Algorithm 2: Updating content’s popularity
Electronics 13 00883 i002

3.4. Cache Management

A CH manages how to cache content in its cluster. When a CH receives or forwards a data packet, it uses the following operation.
  • If the packet destination is its CM and the CM has free cache space, the CH adds the timestamp and the name of the content to the corresponding CM’s entry in its CM table.
  • Otherwise, the CH decides whether to save the content or not by checking if the content is already stored in the cluster. If it saves the content, it also decides which vehicle in the cluster to save it to. If the target vehicle is a CM, it sends a request-to-cache message to the CM and adds the timestamp and the name of the content to the entry of the corresponding CM in its CM table (step 8 in Figure 4). If the buffer of the CM has no free space, the CM decides which content to replace based on content popularity. Figure 6 shows this operation.
A CM overhears data packets and stores content in its temporary buffer (step 6’ in Figure 4). When a CM receives a request-to-cache message from its CH, it moves the specified content from its temporary buffer to its CS (Content Store). If a specified time has passed and the CM has not received a request-to-cache message, it discards overheard content silently.
When a CH selects a CM to store content, it selects the vehicle with the maximum free space. If no CMs have free cache space, the CH selects a CM with content with the least popularity. Then, the selected vehicle replaces the content with the new content using the same cache policy as the CH.
When a CM joins a cluster, it sends a sync message that contains information about the n sync most popular content data (with content names and their timestamps) in its buffer to its CH. The CH updates its CM table accordingly. When a CH considers that a CM has left its cluster (cluster-keep-alive timer expires), it removes the CM’s information from its CM table.
Algorithm 3 is the pseudo-code for storing content in a cluster.
Algorithm 3: Storing content in a cluster
Electronics 13 00883 i003

4. Simulation Evaluation

We evaluated the proposed method with clustering, collaborative caching, and content popularity inference (C-CoCach-Pop) using Scenargie [23]: a commercial simulator supporting V2V communications.

4.1. Comparison Methods and Evaluation Metrics

We compared the proposed method with the following three methods. ’Pop’ in a method name indicates that popularity inference is used in the method.
  • V2V communication using basic CCN [2] without clustering (BCCN-Pop): A vehicle on a communication path caches the content on receiving a data packet, while the cache buffer for a vehicle off the path is not exploited.
  • V2V communication with clustering enabled but collaborative caching disabled (C-BCCN-Pop): Here, clusters are constructed for forwarding packets in the same way as in the proposed method.
  • V2V communication with clustering and collaborative caching but without content popularity inference (C-CoCach [9]): In this method, cache management is based on the Least Recently Used (LRU) policy.
  • ECV+ method [3] in which each vehicle decides the cache probability based on channel usage.
A vehicle sends a beacon periodically in both the proposed and comparison methods. A beacon only contains the sender’s position and velocity in BCCN-Pop, while it includes the information for constructing and managing clusters in C-BCCN-Pop, C-CoCach, and C-CoCach-Pop. The simulation also evaluates the impact of overhead for managing CMs’ caches.
We measured the median latency of data packets, channel usage, the success ratio of receiving requested content, and the cache hit ratio as metrics to evaluate the performance improvement of the proposed method.

4.2. Simulation Condition

We assume that a vehicle requests road condition information and weather information. The provider of content is the vehicle closest to it. Vehicles form clusters in the first 10 s, during which they do not request content. After that, each vehicle sends an interest packet periodically to fetch content. The content requests follow the Zipf law [24], where the number of requests of the k-th most popular content is proportional to 1 / k .
Detailed simulation conditions, simulation scenarios, the values of parameters, and the sizes of interest/data packets and control messages are shown in Table 2, Table 3, Table 4, and Table 5, respectively. Figure 7 is the open street map used in Scenario 3, which simulates real roads and buildings around Chofu Station.
We chose the values of parameters by initial experiments. Figure 8 and Figure 9 show the median latency and the successful reception ratio to fetch content, respectively. From these results, we chose d cluster = 150 m because with this value the successful reception ratio is high while the median latency is low.

4.3. Simulation Results

First, we evaluated the impact of popularity in the proposed method. Figure 10, Figure 11, Figure 12 and Figure 13 compare C-CoCach without popularity inference and C-CoCach-Pop with popularity inference using Scenario 2 with 200 vehicles. By inferring content popularity and storing content with high popularity in a cluster, the in-cluster cache hit ratio is improved (Figure 13), and thus, the median latency for retrieving content is decreased (Figure 10). By shortening content delivery paths, the number of failures in content delivery is also decreased (Figure 11). Note that the overall cache hit ratio of C-CoCach-Pop is less than that of C-CoCach (Figure 12). The cache policy of C-CoCach is LRU, for which content popularity is irrelevant. Hence, vehicles store both popular and unpopular content in C-CoCach, which increases content variety, resulting in a higher cache-hit ratio. But this increases the probability that some popular content is replaced by non-popular content that must be fetched again later.
These results confirm the effect of using popularity inference in the proposed method. In the following, we consider that all methods use popularity inference and evaluate their performance using Scenario 1 at first.
Figure 14 and Figure 15 show the median latency and the successful content reception ratio with respect to cache buffer size, respectively. Generally, a large cache buffer size enables the caching of more content, leading to small latency and a high success ratio in all methods. Compared with BCCN-Pop, the latency may be degraded a little in C-CoCach-Pop because almost all content goes through the CH, which increases the hop count. Compared with C-BCCN-Pop, which also uses clusters, the proposed C-CoCach-Pop method reduces the latency and increases the successful reception ratio in all ranges. This is because, with collaborative caching, more content can be fetched from local cache buffers, which reduces the hop count and also increases the successful reception ratio, as illustrated in Figure 15.
Figure 16 and Figure 17 show the cache hit ratio and in-cluster cache hit ratio, respectively, with respect to the cache buffer size. Both increase with the cache buffer size. The in-cluster cache hit ratio indicates the percentage of content that is retrieved in one hop within a cluster, and it is effectively improved by C-CoCach-Pop. This is also confirmed in Figure 18, which illustrates the distribution of hop counts required to retrieve content.
Figure 19 shows the average channel usage when the buffer size is 100. BCCN-Pop has the lowest channel usage, and implementing collaborative caching increases the overhead. When considering the methods using clustering, C-CoCach-Pop may cause more overhead in cache management. But actually, the results show that C-CoCach-Pop has less channel usage than C-BCCN-Pop, which indicates that the reduction in channel usage by collaborative caching is greater than that caused by the management overhead.
Next, we investigated the impact of vehicle density. Figure 20 and Figure 21 show the results for Scenario 2, for which the number of vehicles increases from 100 to 200, and the channel becomes more congested than in Scenario 1. Compared with Figure 14, the cache hit ratio in C-CoCach-Pop is improved. This is because clusters tend to contain more vehicles, and thus, more content can be cached in a cluster by exploiting the increased cache buffer. However, the successful reception ratio is degraded in all methods due to channel congestion. Compared with C-BCCN-Pop, our proposed method C-CoCach-Pop with collaborative caching still reduces the latency and improves the success ratio.
Figure 22 and Figure 23 show the results for Scenario 3, for which the number of vehicles is 300. The proposed C-CoCach-Pop method still achieved better results than C-BCCN-Pop for each metric, although the differences get smaller than for the straight road in Scenarios 1 and 2 due to the obstruction of roadside buildings and complex vehicle mobility in the urban area.
Finally, we compared our method with an existing method: ECV+ [3]. Figure 24 and Figure 25 compare our method and ECV+ for Scenario 2. Our method achieved better results because (1) ECV+’s probabilistic caching makes it difficult to fill vehicles’ cache buffers, and (2) ECV+ does not consider content popularity.

5. Conclusions

This paper has proposed a new collaborative caching method for fetching content efficiently in vehicular networks by extending our previous method [9] with popularity inference. By organizing vehicles in clusters, the proposed method has the following merits: (i) a CH learns all content cached in its cluster to reply to content requests, which improves the cache hit rate, (ii) a CH efficiently uses all cache buffer space in its cluster by avoiding duplicate caches and storing more diverse content, which increases the probability of locally responding to a content request. Further, we let vehicles infer content popularity and store popular content in their clusters, which increases the number of requests that could be completed within a cluster. In this way, the proposed method reduces the fetching latency of data packets by 33% and improves the packet reception rate by 17% compared to C-BCCN-Pop in an environment where the number of vehicles is 200. The proposed method is expected to enable rapid acquisition of road conditions.
In the future, we will further refine cache management to improve cache usage to get better results in both straight roads and urban models and will evaluate our method using actual V2V communications.

Author Contributions

Conceptualization, S.T.; methodology, H.T.; software, H.T.; validation, H.T.; writing—original draft preparation, H.T.; writing—review and editing, S.T.; visualization, H.T.; supervision, S.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Segawa, Y.; Tang, S.; Ueno, T.; Ogishi, T.; Obana, S. Improving Performance of C-V2X Sidelink by Interference Prediction and Multi-Interval Extension. IEEE Access 2022, 10, 42518–42528. [Google Scholar] [CrossRef]
  2. Jacobson, V.; Smetters, D.K.; Thornton, J.D.; Plass, M.F.; Briggs, N.H.; Braynard, R.L. Networking Named Content. In Proceedings of the CoNEXT ’09: Proceedings of the 5th International Conference on Emerging Networking Experiments and Technologies, Rome, Italy, 1–4 December 2009; pp. 1–12. [Google Scholar] [CrossRef]
  3. Nakazawa, T.; Tang, S.; Obana, S. CCN-Based Inter-Vehicle Communication for Efficient Collection of Road and Traffic Information. Electronics 2020, 9, 112. [Google Scholar] [CrossRef]
  4. Huang, W.; Song, T.; Yang, Y.; Zhang, Y. Cluster-Based Cooperative Caching With Mobility Prediction in Vehicular Named Data Networking. IEEE Access 2019, 7, 23442–23458. [Google Scholar] [CrossRef]
  5. Quan, W.; Xu, C.; Guan, J.; Zhang, H.; Grieco, L.A. Social cooperation for information-centric multimedia streaming in highway VANETs. In Proceedings of the IEEE International Symposium on a World of Wireless, Mobile and Multimedia Networks 2014, Sydney, Australia, 19 June 2014; pp. 1–6. [Google Scholar] [CrossRef]
  6. Duan, M.; Zhang, C.; Li, Y.; Xu, W.; Ji, X.; Liu, B. Neighbor Cache Explore Routing Protocol for VANET based on Trajectory Prediction. In Proceedings of the 2018 IEEE 3rd Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 12–14 October 2018; pp. 771–776. [Google Scholar] [CrossRef]
  7. Zhao, W.; Qin, Y.; Gao, D.; Foh, C.H.; Chao, H.C. An Efficient Cache Strategy in Information Centric Networking Vehicle-to-Vehicle Scenario. IEEE Access 2017, 5, 12657–12667. [Google Scholar] [CrossRef]
  8. Deng, G.; Wang, L.; Li, F.; Li, R. Distributed Probabilistic Caching strategy in VANETs through Named Data Networking. In Proceedings of the 2016 IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), San Francisco, CA, USA, 10–14 April 2016; pp. 314–319. [Google Scholar] [CrossRef]
  9. Tokunaga, H.; Tang, S. Efficient V2V communications by clustering-based collaborative caching. In Proceedings of the 2022 IEEE International Conference on Vehicular Electronics and Safety (ICVES), Bogota, Colombia, 14–16 November 2022; pp. 1–6. [Google Scholar] [CrossRef]
  10. Qi, W.; Song, Q.; Wang, X.; Guo, L.; Ning, Z. SDN-Enabled Social-Aware Clustering in 5G-VANET Systems. IEEE Access 2018, 6, 28213–28224. [Google Scholar] [CrossRef]
  11. Zhang, D.; Ge, H.; Zhang, T.; Cui, Y.Y.; Liu, X.; Mao, G. New Multi-Hop Clustering Algorithm for Vehicular Ad Hoc Networks. IEEE Trans. Intell. Transp. Syst. 2019, 20, 1517–1530. [Google Scholar] [CrossRef]
  12. Wang, J.; Liu, K.; Xiao, K.; Chen, C.; Wu, W.; Lee, V.C.S.; Son, S.H. Dynamic Clustering and Cooperative Scheduling for Vehicle-to-Vehicle Communication in Bidirectional Road Scenarios. IEEE Trans. Intell. Transp. Syst. 2018, 19, 1913–1924. [Google Scholar] [CrossRef]
  13. Gupta, D.; Rani, S.; Singh, A.; Rodrigues, J.J.P.C. ICN Based Efficient Content Caching Scheme for Vehicular Networks. IEEE Trans. Intell. Transp. Syst. 2022, 24, 15548–15556. [Google Scholar] [CrossRef]
  14. Xue, Z.; Liu, Y.; Han, G.; Ayaz, F.; Sheng, Z.; Wang, Y. Two-Layer Distributed Content Caching for Infotainment Applications in VANETs. IEEE Internet Things J. 2022, 9, 1696–1711. [Google Scholar] [CrossRef]
  15. Zaman, S.K.U.; Mustafa, S.; Abbasi, H.; Maqsood, T.; Rehman, F.; Khan, M.A.; Ahmed, M.; Algarni, A.D.; Elmannai, H. Cooperative Content Caching Framework Using Cuckoo Search Optimization in Vehicular Edge Networks. Appl. Sci. 2023, 13, 780. [Google Scholar] [CrossRef]
  16. Yang, J.; Tan, Y.; Xie, J.; Teng, B.; Dong, S. Vehicle clustering based edge caching scheme in internet of vehicles. IET Commun. 2023, 17, 1829–1836. [Google Scholar] [CrossRef]
  17. Xing, Y.; Sun, Y.; Qiao, L.; Wang, Z.; Si, P.; Zhang, Y. Deep Reinforcement Learning for Cooperative Edge Caching in Vehicular Networks. In Proceedings of the 2021 13th International Conference on Communication Software and Networks (ICCSN), Chongqing, China, 4–7 June 2021; pp. 144–149. [Google Scholar] [CrossRef]
  18. Zhang, Z.; Lung, C.H.; St-Hilaire, M.; Lambadaris, I. Smart Proactive Caching: Empower the Video Delivery for Autonomous Vehicles in ICN-Based Networks. IEEE Trans. Veh. Technol. 2020, 69, 7955–7965. [Google Scholar] [CrossRef]
  19. Wu, C.; Yoshinaga, T.; Ji, Y.; Murase, T.; Zhang, Y. A Reinforcement Learning-Based Data Storage Scheme for Vehicular Ad Hoc Networks. IEEE Trans. Veh. Technol. 2017, 66, 6336–6348. [Google Scholar] [CrossRef]
  20. Liu, L.C.; Xie, D.; Wang, S.; Zhang, Z. CCN-based cooperative caching in VANET. In Proceedings of the 2015 International Conference on Connected Vehicles and Expo (ICCVE), Shenzhen, China, 19–23 October 2015; pp. 198–203. [Google Scholar] [CrossRef]
  21. Li, Z.; Simon, G. Time-Shifted TV in Content Centric Networks: The Case for Cooperative In-Network Caching. In Proceedings of the 2011 IEEE International Conference on Communications (ICC), Kyoto, Japan, 5–9 June 2011; pp. 1–6. [Google Scholar] [CrossRef]
  22. Dora, D.P.; Kumar, S.; Kaiwartya, O. Efficient dynamic caching for geocast routing in VANETs. In Proceedings of the 2015 2nd International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 19–20 February 2015; pp. 979–983. [Google Scholar] [CrossRef]
  23. Space Time Engineering. Available online: https://www.spacetime-eng.com/jp/products?page=Products_Top_jp (accessed on 20 November 2023).
  24. Li, W. Zipf’s Law everywhere. Glottometrics 2002, 5, 14–21. [Google Scholar]
Figure 2. CH sends a beacon containing a connection-expired-predicted CH ID.
Figure 2. CH sends a beacon containing a connection-expired-predicted CH ID.
Electronics 13 00883 g002
Figure 3. After receiving the beacon, a new cluster will be created.
Figure 3. After receiving the beacon, a new cluster will be created.
Electronics 13 00883 g003
Figure 4. Communication process of the proposed method.
Figure 4. Communication process of the proposed method.
Electronics 13 00883 g004
Figure 5. Request counter table.
Figure 5. Request counter table.
Electronics 13 00883 g005
Figure 6. Data packet forwarding operation.
Figure 6. Data packet forwarding operation.
Electronics 13 00883 g006
Figure 7. The map used in Scenario 3 (around Chofu Station in Japan) (©OpenStreetMap contributors).
Figure 7. The map used in Scenario 3 (around Chofu Station in Japan) (©OpenStreetMap contributors).
Electronics 13 00883 g007
Figure 8. Median latency for retrieving the requested content with respect to cluster radius (Scenario 2).
Figure 8. Median latency for retrieving the requested content with respect to cluster radius (Scenario 2).
Electronics 13 00883 g008
Figure 9. Success ratio of retrieving requested content with respect to cluster radius (Scenario 2).
Figure 9. Success ratio of retrieving requested content with respect to cluster radius (Scenario 2).
Electronics 13 00883 g009
Figure 10. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 2).
Figure 10. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 2).
Electronics 13 00883 g010
Figure 11. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 2).
Figure 11. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 2).
Electronics 13 00883 g011
Figure 12. Cache hit ratio with respect to cache buffer size (Scenario 2).
Figure 12. Cache hit ratio with respect to cache buffer size (Scenario 2).
Electronics 13 00883 g012
Figure 13. In-cluster cache hit ratio with respect to cache buffer size (Scenario 2).
Figure 13. In-cluster cache hit ratio with respect to cache buffer size (Scenario 2).
Electronics 13 00883 g013
Figure 14. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 1).
Figure 14. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 1).
Electronics 13 00883 g014
Figure 15. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 1).
Figure 15. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 1).
Electronics 13 00883 g015
Figure 16. Cache hit ratio with respect to cache buffer size (Scenario 1).
Figure 16. Cache hit ratio with respect to cache buffer size (Scenario 1).
Electronics 13 00883 g016
Figure 17. In-cluster cache hit ratio with respect to cache buffer size (Scenario 1).
Figure 17. In-cluster cache hit ratio with respect to cache buffer size (Scenario 1).
Electronics 13 00883 g017
Figure 18. Distribution of hop count required to retrieve the requested content (Scenario 1, buffer size: 100).
Figure 18. Distribution of hop count required to retrieve the requested content (Scenario 1, buffer size: 100).
Electronics 13 00883 g018
Figure 19. Average channel usage (Scenario 1, buffer size: 100).
Figure 19. Average channel usage (Scenario 1, buffer size: 100).
Electronics 13 00883 g019
Figure 20. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 2).
Figure 20. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 2).
Electronics 13 00883 g020
Figure 21. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 2).
Figure 21. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 2).
Electronics 13 00883 g021
Figure 22. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 3).
Figure 22. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 3).
Electronics 13 00883 g022
Figure 23. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 3).
Figure 23. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 3).
Electronics 13 00883 g023
Figure 24. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 2).
Figure 24. Median latency for retrieving the requested content with respect to cache buffer size (Scenario 2).
Electronics 13 00883 g024
Figure 25. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 2).
Figure 25. Success ratio of retrieving requested content with respect to cache buffer size (Scenario 2).
Electronics 13 00883 g025
Table 1. Comparison of related methods.
Table 1. Comparison of related methods.
ReferenceKindDescription
 [10]Cluster constructionConstruct clusters based on inter-vehicle distances, vehicle heights, and route stability
 [11]Cluster constructionClusters may be constructed with multi-hop distances
 [3]Reducing the cache redundancyCache probability is determined by channel usage
 [7]Reducing the cache redundancyCache probability is determined by content similarity and vehicle mobility
 [8]Reducing the cache redundancyCache probability is determined by the characteristics of vehicles’ mobility
 [4]Collaborative cachingCHs receive content from RSU; CMs receive content from both of them
 [12]Collaborative cachingVehicles broadcast alternately
 [5]Collaborative cachingA vehicle preferentially requests content from others moving in the same lane
 [13]Collaborative cachingA vehicle decides whether to cache content based on the vehicle’s power, the content’s popularity, the gain acquired by the content, and the distance (hop count) to the content provider
 [14]Collaborative cachingBoth vehicles and RSUs construct clusters
 [15]Collaborative cachingVehicles within an RSU are divided into clusters, and they send requests to the RSU
 [16]Collaborative cachingVehicles construct clusters and receive requested content from either their CHs or RSUs; part of the mathematical model is solvable as a knapsack problem
 [6]Collaborative cachingVehicles broadcast the names of cached content
[17]Collaborative cachingCache policy is determined with deep reinforcement learning
[18]Collaborative cachingThe problem of which content to store and where to store it is determined by machine learning
[19]Collaborative cachingData carrier node is selected by reinforcement learning
 [20]Cache place determinationCache placement is treated as an MWVCP problem
 [21]Cache place determinationNode n  ( n < k ) only stores content c, where c mod k = n
 [22]Relay vehicle determinationVehicles that lie in the common communication area are preferentially selected
Table 2. Simulation conditions.
Table 2. Simulation conditions.
KeyValue
SimulatorScenargie [23]
MAC protocolIEEE 802.11p
Velocity of vehicles20 m/s
Vehicle transmission range350 m
Beacon transmission interval100 ms
Buffer size (number of content)0, 10, …, 100
The interval for sending interest packets1 s
The lifetime of content in temporary buffer500 ms
Simulation time320 s
Number of trials8
Table 3. Simulation scenarios.
Table 3. Simulation scenarios.
No.Road ShapeNumber of Vehicles
Scenario 14 km straight road100
Scenario 24 km straight road200
Scenario 31.6 km × 1.6 km urban area300
Table 4. Values of parameters.
Table 4. Values of parameters.
Variable NameValue
t neighbor 300 ms
d cluster 150 m
v cluster 1.4 m/s
n sync 100
t break 1 s
Table 5. Sizes of interest/data packets and control messages.
Table 5. Sizes of interest/data packets and control messages.
Packet NameSize (Bytes)
Interest packet128
Data packet512
Content-Not-Found128
Request-To-Cache128
Sync (for n content) 24 + 12 n
Beacon (clustering disabled)48
Beacon (clustering enabled)80
Merge request16
Merge ack20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tokunaga, H.; Tang, S. Efficient V2V Communications by Clustering-Based Collaborative Caching. Electronics 2024, 13, 883. https://doi.org/10.3390/electronics13050883

AMA Style

Tokunaga H, Tang S. Efficient V2V Communications by Clustering-Based Collaborative Caching. Electronics. 2024; 13(5):883. https://doi.org/10.3390/electronics13050883

Chicago/Turabian Style

Tokunaga, Hiroki, and Suhua Tang. 2024. "Efficient V2V Communications by Clustering-Based Collaborative Caching" Electronics 13, no. 5: 883. https://doi.org/10.3390/electronics13050883

APA Style

Tokunaga, H., & Tang, S. (2024). Efficient V2V Communications by Clustering-Based Collaborative Caching. Electronics, 13(5), 883. https://doi.org/10.3390/electronics13050883

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop