1. Introduction
By virtue of its versatility and utility convenience, mobile wireless communication has now become a necessity. Over the years, different technologies have been developed to increase the data rate in wireless networks. Technologies like WiMAX (Worldwide Interoperability for Microwave Access) and 3G/4G devices can support 100 Mbps per data channel [
1], and LTE (Long Term Evolution) can go as high as 200 Mbps. This has created room for different types of applications in different specialties. Users have come to desire not only the basic voice functionality of wireless devices, but also different forms of mobile multimedia communication. Video communication via a VANET (Vehicular Adhoc NETwork) is one such application.
Safety and environmental issues in managing existing road networks have become one of the greatest concerns of both the automotive industry and governmental bodies. A VANET consisting of clusters of vehicles communicating with each other and also with roadside infrastructure can provide a solution to these issues. Telematics for the telecommunications of safety information is one major tool used in this area. Progressively, sharing real-time driving video traffic among vehicles and roadside infrastructures would be the next natural step to enhance the safety feature. One motivation of this paper comes from the statistics of road traffic fatalities which shows that there were 44,192 accidental deaths in Canada from the years 2000 to 2004 [
2]. California, with a population of 37 million people reports, one million vehicle accidents each year [
3]. The increase in road traffic volume over the years has led to a concern among researchers to create an effective, efficient and safe automotive environment with the help of video communication. One example is the “smart car” in the state of Nevada which, as a first step towards this goal, uses video to assist hands-free driving.
Unlike other wireless environments that are mostly stationary or under slow motion, transmitting video signals in a VANET poses more challenges to resolve. Electromagnetic interferences such as those from engine electronics, from Additive White Gaussian Noise, or those due to weather conditions can affect the QoS (Quality of Service) as seen by a wireless user. Since the topology is constantly changing, a vehicle could move out of sight to cause an outage in data transmission. Obstacles along LOS (Line of Sight) communication are another major reason to cause an outage. Finally, packet loss due to data traffic congestion will definitely make things worse, and communication disruptions may become inevitable.
Understanding the characteristics of real-time video would allow the design of efficient real-time transmission/transportation protocols to address the above challenges and to allow us to carry out further study. There has been continuous and extensive work on VBR video from various specific video sources, (e.g., from video conferencing cameras, web cameras, broadcast TV or HDTV [
4,
5], and non real-time processing/storage of data taken from the cameras installed in a real public safety law enforcement car system [
6]. However, images from a vehicular camera have higher activity around all edges than the middle portion. This is very different from a normal camera constructed for stationary scenarios, and can be addressed/utilized in the design of video communication. So far, we are not aware of any comprehensive study related to real-time driving video traffic originating in the actual vehicle camera, such as GV1500 from GenieView Inc. [
7]. Furthermore, all these prior arts focus on the modeling of separate and independent video traffic before the wireless networks, and there is no work done on the video model of the traffic coming out of a wireless link, as will be needed in our later study of highway video through a VANET.
Higher layer protocols are usually required to support various video applications. TCP (Transport Control Protocol) [
8] and UDP (User Datagram Protocol) [
9] for real-time transportation of video have been shown to be ineffective. As is further reviewed in
Section 2, TCP is best used for lossless data transfer as it ensures packet delivery by its support of packet retransmission. This poses a problem for real-time traffic, as there is no time for retransmission [
10]. UDP is better for video traffic, but unfortunately still cannot guarantee real-time and loss performance. RTP (Real-Time Transport Protocol) [
11] offers an alternative but has other shortcomings that need to be remedied, as is discussed later. Most of the aforementioned transport algorithms and the traffic control protocols below are built on top of TCP and have only been tested in wired networks. Quite often, packets in a wireless network are stored in a buffer waiting for proper bandwidth allocation before they can be delivered. In real-time communication, a poorer quality can be tolerated, but delay should be reduced to a minimum if not eliminated.
Since a large volume of video traffic can easily consume the limited wireless bandwidth, there is a need to throttle the video traffic for better bandwidth utilization and avoid congestion. Traffic/Congestion control can provide smooth throughput and high link utilization for real-time data. Of the two main types of traffic congestion, control in computer networking (reviewed in
Section 2), implicit congestion control does not provide smooth throughput for real-time and low link utilization. Also, the control is based on packet loss-events which are not a big concern for real-time traffic as mentioned earlier. Explicit congestion control of traffic from the router node becomes impractical in a high speed when the CPU cannot catch up with the updating of events arising from changing driving speeds. Consequently, a different protocol (desirably with a low complexity) should be considered.
Finally, it is safer and more cost efficient to simulate possible solutions than actual field experimenting with highway driving. An effective VANET simulation platform/model has pertinent importance in research and industry to emulate a real world situation. One major challenge is the development of an effective platform that can bring all issues described earlier under one simulation model. A major issue is the integration of an effective mobility model that considers vehicle-to-vehicle interaction and vehicle-to-infrastructure interaction along with the full functionalities of a communication device capable of effective receiving, processing and transmitting capabilities. Although there has been work on creating a similar platform, e.g., [
12,
13,
14], little work has been done on certain issues related to video streaming from vehicle camera such as the characterization of the data traffic (e.g., its unique packet size distribution) and the behavior under different driving conditions (e.g., highway
vs. local).
In view of the above and others reviewed in the related work in the next section, we seek to create a robust platform for effective VANET simulation, one that can provide an environment to test, develop and deploy effective communication protocols that would enhance real-time communication of video information from vehicle camera. As forerunners, in this paper, we would like to:
- (i)
create a real-time camera video model under VANET conditions;
- (ii)
integrate a realistic VANET mobility model to test the testbed at various speeds;
- (iii)
integrate a practical algorithm to survive network congestion;
- (iv)
demonstrate the capability of such a platform.
To achieve the above objectives, we first adopt the protocol architecture of our simulation platform as shown in
Figure 1. We need to adopt appropriate protocols for the lower layers consisting of the Data Link Layer and the Physical Layer, and the upper layers for the remainder. After a comparison of several viable technologies for the lower layers such as the pros and cons as reviewed in
Section 2, we have chosen WiMAX (Worldwide Interoperability for Microwave Access) for its cost-effective approach that can provide a high data rate that can satisfy the needs of our vehicle camera users (low latency and high coverage) at high speed at an affordable cost. It is scalable because its uses OFDMA (Orthogonal Frequency Division Multiple Access) to allow flexible frequency re-use. It implements full MIMO (Multiple-Input and Multiple-Output) which is good for mobile applications by enhancing timely information delivery to save lives and improve quality of life. In view of the complexity problem in implementing an explicit congestion control mechanism in the upper layers to control streaming traffic from driving video in real-time, our present approach is not to “cure” congestion but to “survive” congestion. We propose to reduce the compression rate of the video packets once congestion is detected in the network. We use the RTCP (Real Time Control Protocol) to feed such information back to the sender.
Figure 1.
Overall Architecture.
Figure 1.
Overall Architecture.
The next step is to find a simulation tool and a realistic vehicular mobility model for integration. Of the few available network simulation languages such as Qualnet and NS-2, we have chosen OPNET because it has different wireless models including WiMAX. It has a rich set of video streaming protocols, and it provides a good framework, which allows us to integrate our mathematical driving video model effectively so that we can explore different scenarios during network traffic congestion. After searching, we have also decided that VanetMobiSim [
15,
16] is the only VANET mobility model that would allow the output code generator to be modified in order to generate and output traces useable by other simulators. Consequently, we can integrate its traces into our existing OPNET model using the open-loop approach.
To support the experiments from our simulation platform, we need to use realistic driving video traffic streams. Since we do not have an additional ACE [
17] license to support the feeding of life data into our OPNET simulator, we shall generate our traffic streams based on a model established from real video traffics transported over wireless networks. Not only do we want to use this model to generate traffic traces quickly, we would also use it to alter the traffic characteristic under different scenarios later on without having to recapture traffic traces from life experiments, which can be both costly and time consuming.
The contributions of this paper lie in the following:
- (i)
The design and implementation of a theoretical but more realistic driving video model than the traditional video model. To the best of our knowledge, there is no comprehensive study related to real-time driving video traffic originated from a vehicle camera, nor any work done on the modeling of the video traffic from the vehicular camera system through the wireless link;
- (ii)
The integration of a VANET mobility model with such a driving video model to provide a practical VANET simulation platform for the evaluation of video traffic under different scenarios;
- (iii)
The design and implementation of a practical traffic controller for driving video over RTP applications.
The rest of the paper is organized as follows.
Section 2 discusses existing works related to our research, especially their deficiencies.
Section 3 discusses the network operations and models used in our testbed. To illustrate the procedure of modeling the driving video traffic in VANET,
Section 4 presents an example of the trace collection and stochastic analysis of the trace. We then discuss the OPNET simulation of different scenarios and evaluate their performance in
Section 5.
3. Operation and Models of the VANET Testbed
Figure 2 is a simple example of a VANET consisting of cars communicating their video information with each other and with the Internet (IP cloud) via the RSUs, which can be implemented as a WiMAX-enabled Base Station. The server is to forward or to archive the streaming video from the cars. Each car is acting as a WiMAX Mobile Station (MS) capable of receiving, processing and forwarding data to the destination and consequently to the backhaul network via the WiMAX Base Station.
In our VANET testbed, each vehicle has the ability to communicate with any neighboring vehicles. Depending on the nature of the messages, they will either stay within the VANET or exit the backhaul network via the RSU. For instance, tailgating, collision and braking (from preceding cars) are warning messages that can remain within the VANET network, while detailed regional weather forecast, notification of traffic jams and collision are usually exchanged between the backhaul network and the VANET. In many video applications we envision that video messages from the vehicular camera are usually required to be forwarded from the point of interest (such as traffic congestion, road block, and accidents) to a control center in the backhaul network to aid various parties (such as traffic personnel, emergency agents, insurance companies) to respond to the situations more effectively. We shall refer to these types of traffic as “safety driving video” or simply “safety video” later on.
To implement a testbed for the study safety driving video traffic among vehicles, we shall provide the operations at different layers as well as different models used in the testbed in the following two sections.
Figure 2.
Vehicle camera network operation.
Figure 2.
Vehicle camera network operation.
3.1. Protocol Layers and Operations
With reference to
Figure 2, an RSU (BS 1 or BS 2) has the capability of handling up to N cars simultaneously. Each car/MS (Mobile Station) is associated with the closest RSU in the coverage area. Safety driving video packets are routed and given priority by recognizing their service class and scheduling type. For example, the silver service class and the rtPS (real-time Polling Service) scheduling [
17] are assigned to the safety video. For each service class, there is a maximum sustainable traffic and reserved traffic rates (e.g., 384 kbps for the silver service class). Each video packet arriving from the higher layer at a MS is expected to be within a size limit. Otherwise, it will be segmented before being encapsulated into a PDU (Protocol Data Unit) with the appropriate header information, and then transmitted. For example, video packets longer than 1500 B must be segmented to fit the SDU (Service Data Unit) for the silver (safety video) service.
The driving video traffic is generated from the application layer when an MS wants to transmit. We use UDP (User Datagram Protocol) as the transport layer protocol to cater for the end-to-end delivery of packets. RTP (Real Time Protocol) is used at the application layer to combat the possible unreliability in UDP. RTP is a real-time streaming protocol designed for streaming audio and video. In conjunction, we use RTCP (Real-Time Transport Control Protocol) for controlling, signaling and quality service monitoring.
The unique nature of driving video traffic is due to its burstiness and persistence. The size of each frame generated is dependent on its previous frames and the driving speed, and the exact time a frame is generated or received also depends on previous frames and road conditions. These two correlated uncertainties contribute to the high variable bit rate characteristics of driving video communication. Coupled with the real-time and accuracy requirement of our safety applications, these factors pose a modeling challenge. Of the various video codec schemes that have been developed e.g., MPEG2/4, H.263/4, we have chosen the MPEG4 video compression for our vehicular cameras and studied its traffic model for our testbed. Our array-based MPEG encodes the input video into three types of compressed frames called the I, the Peven, and the Podd frames. I-frames are compressed using intra frame information only. Peven or Podd-frames are coded similarly but with motion compensation for even or odd lines respectively according to the previous I frame. Depending on the vehicle driving speed or congestion condition, Peven and Podd frames may be dropped alternatively, or altogether. In general, I–frames are 10 times bigger than P frames. After coding, the frames are arranged in a deterministic order, which is called GOP (Group Of Pictures). This pattern is not specified hence different coders may use different patterns for subsequent GOPs. A common feature, however, is that all video frames exhibit strong time dependent correlations among themselves. This gives rise to different queuing statistics and time-dependent rush hour congestion phenomenon. This is the subject of streaming traffic analysis and modeling later on, as well as the need for a delicate adaptive congestion controller design.
Each car has a WiMAX capability in transmitting (and receiving) at the physical and the MAC layers after receiving the safety video packets from (and delivering to) the network layer. At the MAC layer, the destination address is examined and the nearest RSU is chosen to handle the request. The packet is sent to that RSU, and further forwarded accordingly. In our investigation, the IP cloud in
Figure 2 can be set to its default values and acts a router. The server is configured to accept packets generated by our simulator.
As suggested above, we shall use mobile WiMAX technology at the physical and MAC layers for our VANET. This is a revision of the fixed WiMAX IEEE 802.16e-2005 standard [
48] and it provides functionalities such as Base Station handoffs, MIMO (Multiple Input Multiple output) transmit/receive diversity, and scalable FFT (Fast Fourier Transform) sizes.
Table 1 lists some of the more relevant features. Although WiMAX can support high data rates providing up to 128 Mbps downlink and 56 Mbps uplink using its MIMO antenna techniques, we shall use SISO antenna technique, which supports up to 1 Mbps uplink and downlink. We use the WiMAX MAC architecture to define service flows that can be mapped into gradual IP sessions to enable end-to-end IP based QoS.
Table 1.
IEEE 802.16e characteristics.
Table 1.
IEEE 802.16e characteristics.
Sub-layers 2 | Standard | IEEE 802.16e–2005 |
Multiplexing | OFDMA |
FFT size | Scalable (512, 1024, 2048, etc.) |
Duplexing mode | TDD |
Sub-layers 1 | Modulation scheme | QPSK, 16-QAM, and 64-QAM |
Subcarrier spacing | 10.94 kHz |
Signal bandwidths | 5, 7, 8.75, and 10 MHz |
Spectrum | 2.3, 2.5, and 3.5 GHz |
Of the two-duplexing modes allowed by mobile WiMAX, we shall use the TDD (Time Division Duplexing) mode (the other is FDD (Frequency Division Duplexing) mode) as done in most deployments in order to accommodate the asymmetric VANET traffic. That is, our uplink traffic requires more bandwidth than the downlink, due to the uploading of safety (not downloading entertainment video that the other researchers are focused on) driving video traffic to the RSU. The WiMAX technology uses OFDMA frames for multiple users to receive data from the BS at the same time thereby increasing bandwidth utilization. The frames are structured into UL (Uplink) and DL (Downlink) subframe as shown in
Figure 3 [
49].
Figure 3.
WiMAX frame structure.
Figure 3.
WiMAX frame structure.
All the MSes receive the frame control section of a DL subframe. The DL mapping field in the frame control section specifies the location and duration of DL data burst in the frame from a particular MS. Conversely; the UL mapping field specifies the location and number of time slots allotted in the frame for a particular MS to transmit its data on UL. All MSes listen to the control section in the DL frame, and receive or transmit data during the slots assigned to them by the BS [
50]. As seen from
Figure 3, this structure enables dynamic allocation of DL and UL resources to efficiently support asymmetric DL/UL traffic. A subchannel is allocated to each user there by reducing the channel interference in the frequency domain. OFDMA (Orthogonal Frequency Division Multiple Access) is the scheme used allowing multiple accesses to every user on our network.
3.2. Network Models
Figure 4 summarizes the various components of our Driving Video VANET OPNET model. It consists mainly of the video model and the VANET model. OPNET modeler then provides the platform for the communication model and allowed for the integration of the various components of the Driving Video VANET OPNET model and other considerations like the congestion controller. They are discussed in the following.
Figure 4.
Driving video VANET OPNET Model structure.
Figure 4.
Driving video VANET OPNET Model structure.
3.2.1. VANET Model
The VANET model consists of the VANET mobility model for the movement of a vehicle, and a communication model for information exchange with neighboring MS and BS.
3.2.1.1. VANET Mobility Model
Of the four categories of VANET mobility models described previously, we adopted both the macroscopic and microscopic models to allow the modeling of vehicle-to-vehicle and vehicle-to-infrastructure interaction. VanetMobiSim is chosen to integrate these models in OPNET due to its flexibility to manipulate output files (by coding its output generator file to produce a format desired by our simulator). It also generates traces in a special Universal Trace Format [
15] in the form of (
time node_id pos_X pos_Y velocity acceleration) where
time records the simulated time;
node_id identifies the specific node; pos
_X and
pos_Y provide the X and Y coordinates of the specific node at the given time respectively; and the
velocity and
acceleration show respectively the velocity at which the node moves and its acceleration at that time. With these capabilities, VanetMobiSim can model traffic lights, stop signs, human mobility dynamics and safe inter-distance management. It allows the integration of road topologies from databases such as GDF [
51] or TIGER [
52] to user-defined topologies as well as random topologies. VanetMobiSim provides a flexible platform in which the user can configure the shortest path during a trip based on the criteria of Dijkstra, road speed and road density. The trip can either be generated by random source-destination or according to activities [
15].
3.2.1.2. RSU and Vehicle Communication Model
The RSU and the vehicles are the major communication nodes in VANET. Our RSU is a WiMAX BS. Each RSU is non-application-sensitive so that a wide range of information can be sent and received. Each vehicle has WiMAX receiving and transmitting capabilities to enable vehicle to vehicle and vehicle to infrastructure (RSU in our case) interaction.
Table 2 shows the basic essential characteristics of our model along with some typical settings.
Table 2.
RSU parameters.
Parameter | Value | Unit |
---|
Physical Layer | IEEE 802.16e | NA |
BS TX Power | 5 | W |
Number of TX | SISO | NA |
BS Antenna Gain | 15 | dBi |
Minimum Power Density | −80 | dBc |
Maximum Power Density | −30 | dBc |
Link Bandwidth | 20 | MHz |
Base Frequency | 5.8 | GHz |
Physical Layer Profile | OFDM | NA |
3.2.2. Driving Video Model
The driving video model is used to analyze and characterize the video trace of a live driving scene captured by a vehicle camera. The results are then used to generate models for use in our simulator for real-time driving video communication in a VANET environment. The implementation of this wireless driving video traffic model allows us to handle a wide range of burstiness in terms of packet inter-arrival and packet size arising from the unique nature of wireless video traffic during driving as mentioned above. A guideline to check the efficiency of a traffic model is, however, needed as to set a standard to be followed. Parsimony, analytic, relative accuracy, flexibility, implement ability and absolute accuracy are factors listed in [
53] showing the engineering tradeoffs among different traffic models.
Using these factors on a scale of 1 to 3, 3 being the highest,
Table 3 shows the settings of a few popular models including the FBM (Fractional Brownian Model) and the OPNET-emulated TCP traffic. By comparison, we adopt the Mini-Pareto model to be used later due to its overall score. In particular, we have taken a systematic approach in developing our mini-Pareto model. Driving video traffic traces were collected using the same camera used for a road test. In our testbed, we traced the frames from the receiver instead of the transmitter in order to correspond to video traveling from vehicle to vehicle, and through the RSU to the city’s road monitoring center in a safety network. The traces were analyzed and stochastically represented and plugged into our simulation platform. A detailed discussion about these steps is presented in
Section 4.
Table 3.
Traffic model comparison.
Table 3.
Traffic model comparison.
Items | FBM | TCP Model | Mini-Pareto |
---|
Parsimony | 3 | 1 | 2 |
Analytical | 1 | 1 | 2 |
Flexibility | 1 | 1 | 1 |
Implemental | 2 | 1 | 3 |
Accuracy | 2 | 3 | 2 |
Mobility | 2 | 1 | 3 |
3.2.3. Traffic Controller Model
As discussed in
Section 1 and
Section 2.4, we only need to survive congestion to meet the QoS requirement of our real-time safety application scenario. This has given us the idea of reducing the complexity of the explicit control algorithm in order to cut down the CPU consumption when the vehicular camera is moving at higher speeds. Our careful study shows that reducing the sending bit rate is the key factor of success, and this can be achieved in two ways: dropping some frames (packets) or reducing quantization level per frame. There are different consequences and tradeoffs.
For driving video traffic, the drop pattern is very important to the quality of service as seen by user. Dropping just the even frames or just the odd frames is usually better than dropping both consecutively. The frequency of a drop can be a more critical factor than the percentage of drop. For instance, dropping 10 consecutive packets out of 100 packets is a 10% drop, which is the same as randomly dropping 1 out of 10 packets from a stream of 100 packets. However, the latter is more tolerable than the former in our system. Since there are many more P frames than I frames, we only drop P frames, first even then odd in our testbed. Thus we can use an automatic predictive source control to assist the explicit congestion control. That is, when P frames start piling up, we drop the P-frames by following a particular drop patter. For example, dropping the even frames first and then the odd frames without waiting for control commands to come from the base station. We can also set an upper limit of approximately 25% beyond which the video becomes less comfortable to watch (and may not satisfy customer expectations).
According to our industrial partners, changing the quantization level will be the next practical and cost-effective approach to reduce bit rate in terms of complexity and ease of deployment. We lower the quantization level when we use fewer coefficients from the coarse DCT (Discrete Cosine Transformation) algorithm, and thus reducing the packet size as well. Depending on the end-to-end delay, the receiver notifies the sender of the quantization level to use in encoding the subsequent video frames. This method will bring the total bit rate down to 50% of the original coding/sending rate. Beyond 50%, we need to drop the color components, thus leaving the video in its black-and-white mode.
In summary, the sending bit rate = frameRate*FrameSize which can be lowered by lowering one or both of the frame rate or frame size. There are tradeoffs between the two. Dropping packets (frames) will lower the frame rate while reducing the quantization level that will lower the resolution. The challenge in lowering the resolution is to keep the user experience/satisfaction high as long as possible, and various experiments have been conducted to find out the optimum point. Although our real GenieView camera has implemented in both methods in firmware to improve user experience, our OPNET model has chosen to drop the frames/packet only. This is reflected in
Figure 5 and
Figure 6 below where the called party collaborates with the calling party to determine the end-to-end delay based on how our controller uses one byte to notify the source to reduce its sending rate by dropping selected frames as discussed above. We also make use of RTP/RTCP at the application layer to provide an effective feedback mechanism.
Figure 5 depicts the video called party (receiver) procedure. Once the variables are initialized successfully, and a connection is established, data transmission can begin upon request. As soon as a frame is received, the frame count is incremented until it reaches the threshold value
p. Once
p is reached (e.g.,
p = 5 in our experiments), the frame count is reset to zero and the end-to-end delayassociated with the packet is checked. If the end-to-end delay is less than the threshold value
t set by the end user on the receiving end (e.g., 100 ms), a default/designed rate of
x2 is fed back to sender via the ACK packet. Else a lower rate
x1 is returned instead. The
x1 value can be a default value such as half of
x2, or computed according to some formula)
Figure 5.
Video called party flow diagram with RTCP implementation.
Figure 5.
Video called party flow diagram with RTCP implementation.
Figure 6 shows the flow chart of the controller operation controller at the calling party manager (sender). Once the initialization process is successful, the video characteristics/properties are collected and a connection is established with the receiver. For every packet received, the algorithm checks if it is a control ACK packet. If not, the required statistics on packet type are collected and the packet is destroyed. If the packet is an ACK packet, the rate is retrieved and used to determine whether the sender should use the normal sending rate
x2 or the reduced rate of
x1. The parameters for the call establishment are also updated according to the new value obtained.
Figure 6.
Video calling party flow diagram with traffic controller implementation.
Figure 6.
Video calling party flow diagram with traffic controller implementation.
4. Experimental Setup
By collecting and studying video traces we want to come up with a model that can generate video traffic with characteristics as close to real life traffic as possible. This section discusses on the setup of our simulator testbed illustrates the trace collection process, its analysis and modeling. This will set the stage for traffic generation and performance evaluation in the next section.
4.1. Trace Collection
As shown in the setup of
Figure 7, we use the monitor to play a series of driving video clips at 10 min each from driving action movies. Each video was captured by a PTS (Pan Tilt Zoom) camera taken from a vehicle. The video is encrypted and highly compressed at a ratio of 250:1 at the transmitter block. Using one processor for each section of the screen, the 96-array processors in the transmitter block can achieve fast independent image processing. The transmitter sends the highly compressed encrypted video images to the control center via a wireless link using the non-licensed 900 MHz FHSS radio-band. Please note that 900 MHz is more affordable to us presently than WiMax and DSRC. Since they are all wireless links, they will all be subjected to the same noise sources, such as EM interference. We only need to obtain similar type of interference and their statistics first to verify our platform.
Figure 7.
Setup for taking video traces.
Figure 7.
Setup for taking video traces.
The receiver block decodes and decompresses the received video frames and plays the image at the control center at about 20 fps. The control-center laptop and the receiver are connected using a RS232 serial port. The control center is then connected through a router to a computer hosting the packet trace-capturing tool—Wireshark™ [
54]. Once the system is turned on, the computer with the Wireshark™ software is set to access the “capture” folder in the control center which would then stream the video via a router for trace capture.
A video trace (a record sequence of time and size of each arriving packet) captured by Wireshark is shown in
Figure 8. Due to the loss and duplication that will change its original characteristics, this type of traffic trace is closer to those from our application in the real environments. We want to characterize such a life driving video trace using proven stochastic processes and integrate it into a model that has as few parameters as possible but which can produce traces to match the real measurements.
Figure 8.
Live driving video trace.
Figure 8.
Live driving video trace.
In order to model video traces of different correlation/fractal characteristics arising from different activities, we collected traces playing three different video clips. The video clips were chosen based on their activity rates. Actions like highway chasing movies (with a lot of movement on the edge of the screen/camera and hence unique variation in frame sizes), stories in a parking lot (with less and slower movement) and dramas on the street (on the average driving speed) were chosen. We combine these clips to form a two-hour video trace (consisting of the time and size of each arriving frame), which is sufficient to characterize the fractal nature of the driving video traffic as well as the wireless channel effect.
4.2. Trace Analysis and Modeling
The next challenge is to analyze the trace and to create a driving video traffic model for traffic generation later. To verify the correlation of video frame and to obtain the modeling parameter, we obtain the frame-to-frame correlation of the video traffic.
Figure 9 is an example using our video trace, which shows correlation is present up to 20 seconds apart.
Figure 9.
Correlation length of driving video.
Figure 9.
Correlation length of driving video.
We have used the Pareto distribution to characterize the general correlation of “action” video frame duration in wireless video streaming as well as the strong correlation between a cluster of byte loss and/or duplication caused by the air interface. The Pareto distribution is a long-tailed distribution with the following pdf (probability density function) g(x) for the consecutive I-frame duration, where the edge activity level is high. Note that x here is not the frame size but the group length rather.
where α > 0 is the location parameter and
c > 0 is the shape parameter. The location parameter value can be obtained from the mean correlation length E(x) of the trace where E(x) = cα/(α − 1). On the other hand, the shape parameter has to be obtained via the Hurst parameter. That is,
Note that the Hurst parameter is a useful parameter to reflect the invariance according to the entropy conservation property [
26]. Since our video is highly compressed with little loss of information, the entropy/uncertainty remains the same. Therefore, we can determine c from Equation (2).
An indirect but easy way to obtain the Hurst parameter value is via the
slope of the IDC (Index of Dispersion for Count) curve as a function of the level of aggregation. Here IDC is defined to be normalized variance of aggregated variables [
26], and the level of aggregation refers the logarithm of the number of variables base-10 (therefore level-0 has 1 variable, level-1 has 10 variables, level-2 has 100 variables and so on). Mathematically, the relationship is given by [
26]
The slope of the curve is usually measured from the first point to the last point on the graph. For example,
Figure 10 gives the IDC of the video trace we use. From this figure, the slope is measured to be 0.33. Hence the Hurst parameter H is 0.67, and the shape parameter is
c = 1.66.
Figure 10.
Entire trace IDC curve.
Figure 10.
Entire trace IDC curve.
Once we obtain the pdf g(x) to characterize the activity duration (via the correlation of I-frames), we shall next determine the packet (frame) size distribution f(x), which is not Pareto in general. Furthermore, unlike the traditional video (e.g., video conference) with the moving part in the middle, our driving video has the moving part on the edge. Whereas the traditional video has a normal distribution in the middle for its packet size after 2D-DCT (Discrete Cosine Transform), the distribution for driving video is at the edge, which becomes very hard to model analytically to match the measured curve. After many trials, we have determined that an empirical model using nine mini-sources laid out in a grid can capture the distribution properly. The surrounding sources will capture the distribution of packets further away from the middle (edges and corners) which are usually moving and which produce large packets. The mini-sources closer to the center will capture the scenes in the middle which are relatively still and that produce smaller packets.
Figure 11 demonstrates the empirical analysis we performed on frame size (horizontal axis) of our traffic trace. For simplicity, here we are using the basic camera configuration, which produces one packet per frame. We use a tool called EasyFit™ [
55] to obtain the pdf of the
n = 9 video traces. The orange curve in the figure shows that the lognormal distribution is the best distribution to fit the given histogram data.
Figure 11.
The pdf f(x) of video trace frames.
Figure 11.
The pdf f(x) of video trace frames.
Finally, we want to determine the mean off-time of each source in order to model the inter-arrival time between each driving scene (such as trees or buildings on the road side). We have determined that the distribution is relatively memoryless, and therefore its pdf is exponential given by
where 1/λ is the mean inter-arrival time.
4.3. Driving Video Traffic Generation
Having obtained all the characterization parameters for traffic generation in previous sections, we are ready to emulate our wireless video traffic.
Figure 12 shows the schematics of our wireless driving video traffic generation model from the superposition of
n mini-sources [
56]. Each mini-source represents a video traffic stream regulated by a switch to give a long-range dependency. We shall model the on-time switch by a Pareto distribution given in Equation (1) and the off-time by a Poisson process [an exponential distribution as given in Equation (4)].
Figure 12.
Model description for driving video traffic generation.
Figure 12.
Model description for driving video traffic generation.
Table 4 shows an implementation of using nine mini-sources, each representing one of the 3 × 3 sectors of a camera screen. Here, frame duration is the transmission time of a frame (
i.e., frame size divided by bit rate). Assume the camera is facing the front when driving on a road. The middle is usually relatively still when compared to the sides that see more activities. The left turns, right turns, and uphills or downhills are all different scenes that can lead to special packet size distributions. Big packet size is produced by a fast side view of the camera while the small size is produced by the almost still middle view. In the characterization of each mini-Pareto source in the table, we use the same shape parameter obtained from the Hurst parameter, but the location parameter is different for different packet size bin.
Table 4.
Mini-sources with characterization values.
Table 4.
Mini-sources with characterization values.
Bins | Mean Off-Time (ms) | Mean On-Time (ms) | Frame Duration (ms) | Mean Packet Size (Bytes) |
---|
mini1 | 7531 | 195 | 8 | 125 |
mini2 | 20603 | 453 | 18 | 290 |
mini3 | 6911 | 702 | 28 | 450 |
mini4 | 5515 | 960 | 38 | 615 |
mini5 | 7800 | 1216 | 32 | 779 |
mini6 | 10400 | 1472 | 39 | 943 |
mini7 | 18200 | 1723 | 46 | 1104 |
mini8 | 27300 | 1984 | 53 | 1272 |
mini9 | 10400 | 2232 | 59 | 1430 |
We have also tried other numbers such as n = 3 × 4 = 12 mini-sources to match the horizontal and vertical ratio of the camera. As verified by our experiments, the more mini-sources created, the higher the accuracy. However, the computational complexity also increases due to the more parameters needed to match the trace. This is translated into longer simulation time. So we have chosen n = 9 for our model for the best trade-off we found between parsimony and accuracy.
Figure 13 shows the video traffic generated mini-pareto video traffic, which can be compared to the live video traffic in
Figure 8. Although not exactly the same, one can verify the simulated traffic captures the level of fluctuation of the driving video traffic.
Figure 13.
Mini-Pareto video traffic.
Figure 13.
Mini-Pareto video traffic.
4.4. OPNET Simulation Setup
Table 5 gives the values of the general parameters we used in our simulations. The terrain has an area of 1300 × 1250 m. The relative position (x, y) on the terrain is used to integrate the VANET mobility model trajectories and obtain the initial positions of the vehicles. Vehicular environment for the path loss parameter is modeled according to the description in the “Radio Tx Technologies for IMT2000” white paper of the ITU [
57]. The shadow fading standard deviation was set to 10 dB. The mobility pattern (trajectory) during the simulation is predefined using the
VanetMobiSim discussed in
Section 3.2.1.1.
Table 5.
Simulation parameters.
Table 5.
Simulation parameters.
Parameter | Value | Note |
---|
Physical Layer | IEEE802.16e | WiMax |
Data Rate | 10 | Mbps |
BS TX Power | 5 | W |
MS Tx Power | 1 | W |
Antenna Type | Omni-directional | Horizontal |
BS Antenna Gain | 15 | dBi |
MS Antenna Gain | 9 | dBi |
Link Bandwidth | 20 | MHz |
Modulation Scheme | 16-QAM | Model |
Path Loss Parameter | Vehicular Environment | ITU |
Number of Vehicles | 10 | Cars |
Mobility Model | VanetMobiSim | Open |
Number of RSUs | 2 | Base |
Simulated Time | 3600 | secs |
Seeds | 127 | random |
Terrain Dimensions | 1300 × 1250 | m2 |
We have investigated different scenarios such as highway and residential, and have studied the effect of varying traffic congestion (density), wireless interference and traffic speed limits. Due to space limitation, we shall present the highway scenario for this paper.
In the highway scenario shown in
Figure 14, we consider cars moving along the highway with a minimum speed of 60 km/h and a maximum speed of 100 km/h. There is one traffic light, just before the cars enter the highway. The car trajectories are represented by the white lines. We use the free space loss model for our pathloss. Unless otherwise stated, the following are assumptions taken throughout the paper for our simulation study:
- (1)
WiMAX BS is a “stationary” node. This is required due to the limitation of our OPNET model and we need it to act as an intermediate node for packet forwarding to the destination;
- (2)
OFDMA is used by the RSU and there is always a slot available for each SS sending video traffic once the SS is within the communication range;
- (3)
No disruption in a communication channel once the node is allocated a dedicated channel;
- (4)
No obstacle along the LOS between the RSU and the mobile cars to allow the simulation to monitor the vehicles within the range of the RSU;
- (5)
Finite buffer size for each transmitter in order to be more realistic so that the trade-off between buffer size and end-to-end delay can be determined.
Figure 14.
Highway scenario.
Figure 14.
Highway scenario.