1. Introduction
Driven by increasing communication needs worldwide, next generation optical networks are expected to provide huge bandwidth as well as support for diverse service demands. These requirements can be fulfilled by the high capacity offered by the next generation optical transport network (OTN) [
1]. To fully utilize the provided capacity, different kinds of switching paradigms in the optical domain are employed to eliminate the processing bottleneck in the electronic domain. A lot of attention has been paid to hybrid approaches [
2,
3], which combine optical circuit switching (OCS) with optical packet switching (OPS) and/or optical burst switching (OBS) to support different QoS demands. In this paper, we focus on the OPS part of such hybrid architectures, which is used to switch the lower bandwidth demand service class. Since OPS has smaller switching granularity and much more immature production technology compared with OCS, its architecture would dominate the total physical size and economic cost of the hybrid switching node.
In the past, various OPS architectures have been proposed, and their performance evaluations have been done in terms of the packet loss ratio (PLR) and average packet delay. Refs. [
4,
5] described and evaluated the performance of the broadcast and select OPS architecture proposed in the European Advanced Communication Technologies and Services (ACTS) KEOPS (keys to optical packet switching). Ref. [
6] proposed a scalable design for OPS architecture with two stages of tunable wavelength converters (TWCs), and its performance was analyzed in detail. However, most of the proposed OPS architectures are mainly focused on the packet contention resolution (optical buffering, properly designed fiber delay lines, deflection routing) [
7,
8,
9,
10], fast packet switching (device, packet synchronization) [
11], and QoS [
12]. Not much attention is paid regarding the physical size and the economic cost, which are also major issues for the OPS architecture.
In this paper, we proposed a novel OPS architecture with input concentrators, which are composed of fixed length fiber delay lines (FDLs) and work as multi-input single-output optical buffers. According to the input wavelength, the specific concentrator aggregates the incoming traffic from all input fibers into its respective output link, which is followed by the TWC and the input channel of the switching fabric. Correspondingly, the physical size, the number of needed TWCs, and the economic cost of the total OPS node are reduced dramatically. In addition, the incoming traffic pattern will be smoothed significantly by the concentrator, leading to very small and negligible packet loss from the contention at the switch output. However, the deployment of input concentrators introduces additional packet loss and delay from the traffic aggregation. In this paper, we present a Markov analytical model to study the input PLR and the average packet delay introduced by input concentrators. The corresponding closed form expressions are given. The model also demonstrates that the buffer size provided by each concentrator will be used up when the total incoming traffic load is larger than 0.69315. This analytical model is validated by the simulations under different network scenarios.
The proposed OPS architecture realizes having a smaller physical size, needing fewer TWCs, a higher link utilization, and a lower economic cost. Therefore, it will play a very important role in a number of network scenarios. As mentioned above, it can improve some hybrid optical switch architectures like OpMiGua, “Optical Migration Capable Networks with Service Guarantees” [
13], which uses OPS to switch the lower bandwidth demand’s service classes. Since only a small amount of incoming traffic are mapped into this kind of service class, our proposed OPS architecture can be used for handling them and to realize the more compact, lower economic cost, and smaller physical size of the total hybrid switch architecture. The proposed OPS architecture can also be used in the access network, which is responsible for collecting end-users’ traffic and has a higher demand of the granularity and flexibility. The deployment of the proposed OPS node will lead to higher link utilization, larger cost saving, higher transmission speeds, smaller physical size, and more flexibility for the access network. In addition, the proposed architecture of the concentrator can also be used for other network scenarios, like ORION (overspill routing in optical networks) [
14] and packet enhanced OTN [
15]. Ref. [
16] proposed the constrained ORION and pointed out the considerable advantages compared with the unconstrained ORION. However, the detailed architecture of how to implement the concentration is not described. The proposed concentrator is a good candidate for this.
The rest of this paper is organized as follows. We present the novel OPS node, the detailed architecture, and the operation of the employed concentrator in
Section 2.
Section 3 introduces the analytical model and presents the closed form expressions of both the PLR and the average packet delay. The results from both simulation and the analytical model are shown in
Section 4. Finally, our conclusions are given in
Section 5.
3. Analytical Model
In this section, we study the performance of the proposed OPS structure. Due to the incoming traffic pattern having been smoothed heavily by input concentrators, the PLR introduced by the contention at the switch output is very small and can be neglected. In the following analysis, we study the PLR and the average packet delay given by input concentrators. A Markov analytical model is built based on the following assumptions:
Uniform traffic pattern for each input fiber, i.e., each input fiber gives the equal traffic load);
Uniform wavelength distribution of the incoming requests, i.e., each incoming wavelength gives the same traffic load). We can restrict our study to only one incoming wavelength from different input fibers;
The Poisson arrival processes for the packets from each input wavelength, i.e., a sufficient large number of flows are aggregated in core network to allow the Poisson arrival process for the packets carried by each input wavelength;
The packet length is fixed, and its size is same as the FDL length. In this paper, we consider the fixed length packets, which is the case for most proposed asynchronous OPS networks [
17].
As described in
Section 2, the same wavelengths from all input fibers are connected to one corresponding input concentrator which works as a multiple-input single-output optical buffer. Since the packet arrives following the Poisson process, the inter-arrival time between the consecutive packets follows a negative exponential distribution (n.e.d). We assume the packet arrival intensity per wavelength is
λ, the service time for each packet is
. Note that due to the packet size being the same as the FDL length, the service time for each packet is the same as the delay time provided by one FDL element. In the following analysis, we restrict our study to only one input concentrator.
Figure 3 shows a timing paradigm to present the detailed packet arrival and departure process from the studied concentrator. PCK
above each packet denotes the
th arriving packet to the studied concentrator. FDL
denotes the
th FDL element the respective arriving packet should go through. The number of FDLs denotes the delay time of this packet introduced by the concentrator. For PCK
, its delay is equal to
.
denotes the arriving time and
denotes its leaving time out of the concentrator.
in the left of the timing paradigm indicates the current queue length (from 0 to
) just after the packet arrives in the concentrator. Its value depends on the current inter-arrival time (
) with this new arriving packet.
indicates the transition probability from the state
to the state
. The bottom of the timing diagram presents the packets’ departure process from the concentrator. Hence, the link utilization can be obtained by this departure process.
We can explain
Figure 3 in the following way. The concentrator works like a memoryless optical buffer. The queue length’s evolution depends on the current queue length and the current inter-arrival time (
) with the new arriving packet. If
, the new packet arrives before the previous one finishes its transmission into the concentrator. The current queue length
will increase by 1. In
Figure 3, from
to
, once the
is smaller than
,
increases by 1. If
, when the new packet arrives, the previous packet has transmitted into the concentrator, however it has not finished its transmission time through one FDL element. Accordingly, the current queue length will not change. Using the same method, if
, the queue length will decrease by 1. If
, the queue length will decrease
. In
Figure 3, when
arrives,
, the
has already gone through the 2th, 1th, and the 0th position in the queue, so
will be inserted in the 0th position in the current queue and the
Q decreases by 2 and becomes 0.
According to the discussion above, the Markov model is built as shown in
Figure 4.
The maximum state is , which denotes the state that the current queue length is , and if the new packet arrives on this state, this packet will be dropped directly.
In order to get the needed PLR and average packet delay from
Figure 4, we have to calculate the probability of each state and the transition probabilities among them.
Next, we derive the transition probabilities in
Figure 4. If
, the value of
will increase by 1. The packet arrival process follows the Poisson process, hence:
where
.
If
, the value of
Q will not change, so:
For
,
is equal to
. Note that when
, the calculation for the probability is different. When
the current queue length is smallest as 0, the state will not change unless
is smaller than
. Its probability is:
For
the current queue length is largest as
, the state will not change unless
is larger than
, which leads to the current queue length decreasing, and the state returns to other smaller states. Otherwise, if
, the state keeps same, and the new arrival packet is dropped directly. Hence its probability is:
If
, the current queue length will be decreased by 1. Hence:
For
,
is equal to
. Note that
is different, if
is larger than
, the current state 1 will go back to the state 0. Hence:
If
, the current queue length will decrease
. So:
For
,
is equal to
. Note that
is different, if
is larger than
, the current state
will go back to the state 0. Hence:
According to the derived transition probabilities and the node equations, the probability of each state can be obtained. In the following formulas, we use
to denote the probability of state
. The node equations are listed as follows:
Submitting transition probabilities into the node equations above, the probability of each state can be derived as:
The solution can be obtained as:
where
. Hence,
. As discussed in [
5], only when
, the Markov chain is positive recurrent and has stationary distribution. If
:
Therefore, we can conclude that no matter how large the buffer size provided by each concentrator is, it will be used up when the total incoming traffic load is larger than 0.69315.
In order to calculate the studied PLR, first we have to find the difference between the electronic buffer and the optical buffer. For the electronic buffer, when the queue length is
, all the positions in the queue are occupied, then all the new incoming packets will be dropped directly. Hence, for the electronic buffer, the packet loss probability is equal to
. However, for the optical buffer, the new arriving packet still can be served if
(the current inter-arrival time) is larger than
and smaller than
Therefore, the packet loss only happens when the current inter-arrival time is smaller than
while the current queue length is
. Note that when
, the studied concentrator has no FDL element to provide time delay, and all the incoming packets will be sent out or dropped directly. The studied concentrator can be considered as an
system, and the corresponding PLR can be obtained by the Erlang loss formula. Combined with Equation (1), the closed form of the studied PLR is:
According to
Figure 4, we also get the average packet delay for all the transmitted packets. For instance, state
Q denotes the current queue length, and it is same as the inserted position of the respective packet. Hence, for state
Q, the delay time for the inserted packet is
. Note that when
, all the incoming packets to the concentrator will be sent out or dropped directly. The average packet delay is 0. We use
to denote the average delay time, and its closed form expression is:
4. Simulation and Analytical Results
A simulation model for an asynchronous OPS node equipped with input concentrators is built in the discrete event modeling on Simula (DEMOS) software. Ten independent simulations were performed for each parameter setting. For all the simulation results, we calculated the average value. All the analytical results are obtained by Equations (11), (13), and (14).
Table 1 summarizes all the employed variables in our simulations. The “values” in
Table 1 present all possible used values under the different scenarios.
For simulations, we consider one Poisson source which generates packets into the proposed OPS node. For the case where the incoming packets have a fixed length, the packet size is fixed as 800 bits. For the case where the incoming packets have a variable length, the packet size follows n.e.d. with the mean value of 800 bits. The link capacity is set as 10 Gbps. Accordingly, the mean service time for both cases is equal to (i.e., ). The packet inter-arrival time for both cases follows n.e.d. Its mean value is set according to the total traffic load () and . In this section we will give the results for the two cases separately.
Case I: The incoming packets have fixed length
In a network with fixed length packets, a natural choice for the buffer is to let the basic delay line map the packet length. In this part we set the length of the basic delay line element in the input concentrator as the same as the packet length to evaluate its performance.
For the proposed OPS node, the channels with the same wavelength from all different input fibers are connected to the specified input concentrator. For each input concentrator, when the packet arrives, it will check the current queue length inside the concentrator. If the queue length is equal to zero, the packet will be sent out directly, otherwise the packet will choose the suitable inserted position and will be sent out by passing through the respective delay time. In the case that the last position of the queue is occupied, this new arriving packet is dropped out directly.
Figure 5 presents the input PLR of the system as a function of the normalized incoming traffic load arrived per concentrator with different configurations, i.e., the buffer size or the FDL number per concentrator is changed from 0 to 75 (
). Here the traffic load denotes the total incoming traffic load for each concentrator. It is equal to
(the number of input fibers times the normalized system load per wavelength).
Results from both the analytical model and simulations are shown. S. denotes the results from the simulation model and A. indicates the results from the built analytical model. The main observation is that the analytical values approximate the simulation results very well under different network scenarios.
A number of observations can be achieved from
Figure 5. When the number of FDL elements is equal to 0, all the arriving packets will be sent out or dropped directly; the studied concentrator can be considered as an
system; the PLR is obtained by the Erlang loss formula directly.
Figure 5 shows the PLR under a different incoming traffic load per concentrator. As expected, under the fixed traffic load, which is not larger than 0.69, i.e.,
, the PLR of whole system decreases very fast as the buffer size increases. For instance, when the normalized traffic load is 0.5, the PLR is about 1.0 × 10
−3 with the number of FDL elements is 12, and it decreases to 1.0 × 10
−8 when K increases to 39. Hence, the system performance can be greatly improved by increasing the number of FDL elements under the low traffic loads. However, as with the discussion given in
Section 3, the buffer size provided by each concentrator will be used up when the total incoming traffic load is larger than 0.69315. This means that when the traffic load is larger than 0.69315, the concentrator can be considered as an overflow system, and its performance cannot be improved even with the larger buffer size. As shown in
Figure 5, when the total system load is 0.7, the obtained PLR values are very close if the number of FDL elements is larger than 39, which means that the improvement of the system performance is very limited by increasing the buffer size per concentrator. When the traffic load is 0.8, the system PLR values are equal when the number of FDL elements is larger than 21, which means the buffer size is used up and cannot be improved by simply increasing the FDL elements under the high traffic load. To sum up, for the proposed OPS architecture with input concentrators, we can conclude that when the traffic load is small, and especially smaller than 0.69315, the PLR can be reduced significantly by adding the number of FDL elements in each concentrator. However, for a large traffic load, the addition of the FDL elements per concentrator cannot improve the performance while leading to a longer average packet delay.
Figure 6 shows the average packet delay of the system as a function of the normalized incoming traffic load under the different buffer sizes provided per concentrator. As in
Figure 5, the traffic load denotes the total incoming traffic load per concentrator. It is also equal to
, as shown in
Table 1. The analytical results are obtained by Equation (3). Compared with the simulation results, we can conclude that the analytical values are in good agreement with the simulation results, which validates the correctness and the feasibility of the proposed analytical model.
When the number of FDL elements is equal to 0, i.e.,
, the arriving packet will be sent out or dropped directly, and no delay exists, which is not given in
Figure 6. We can see that the average packet delay increases as the traffic load grows up under the fixed buffer size. For instance, when the number of FDL elements is fixed as 12, the average packet delay is smaller than 1.0 × 10
−7 when the traffic load is 0.3 and becomes very close to 1.0 × 10
−6 under the traffic load of 0.8. Furthermore, when the traffic load is smaller than 0.69315, we can see that, as the FDL number increases, the average packet delay increases very fast at first and becomes quite stable by increasing the buffer size. As shown in
Figure 6, when the traffic load is 0.4, the average delay can be improved by increasing the buffer size and becomes stable when the FDL number is larger than 12. When the traffic load is 0.5, the average delay cannot be improved when
. When the traffic load is 0.6, the average delay becomes stable under
. Combined with the results shown in
Figure 5, which validates that the PLR can be reduced dramatically as the FDL number increases, we can conclude that for the low incoming traffic load, increasing the FDL number can improve the performance of the studied OPS structure while no additional delay will be introduced. However, when the traffic load is larger than 0.69315, the buffer size will be used up no matter how large the buffer size provided by each concentrator, hence the average packet delay will increase and cannot be stationery for the large traffic load. As shown in
Figure 6, when the system load is 0.7 and 0.8, the average packet delay does not have the threshold value from which the delay improvement is marginal. This can be explained by the difference between the electronic buffer and the optical buffer. For the electronic buffer, the buffered packets can be transmitted back-to-back, and no voids exist between two consecutive packets. The electronic buffer will be used up when the system load is larger than 1. However, for the optical buffer, as discussed in
Section 3, due to the voids between any two consecutive transmitted packets, the buffer size will be used up when the system load is larger than 0.69315. Hence, we can conclude that for the large traffic load larger than 0.69315, we cannot improve the performance of the studied OPS structure by adding the number of FDL elements per concentrator.
Case II: The incoming packets have variable length
Here we consider the incoming packets following n.e.d. with the mean packet size of 800 bits. The length of the basic delay line element is set as the same as the mean packet size. Therefore, the delay time offered by each delay line element is same as the mean packet service time (i.e., ).
Figure 7 presents the
values and the average packet delay per input concentrator as the function of the offered buffer size (
, i.e., the number of FDL elements). All simulation results are given under different system loads. As can be seen from
Figure 7a, when traffic load
, as buffer size
increases, the
values decrease very fast. Accordingly, the network performance can be improved significantly by increasing the buffer size under low traffic loads.
Figure 7b shows that as
increases the average packet delay time increases very fast at first and then becomes stable after one turning point. Therefore, we can conclude that increasing
value (i.e., increasing the buffer size) will improve the performance of the optical buffer dramatically under the low system loads (
): the
value decreases significantly without any additional delay time. However, when traffic load
, as
increases, the
values decrease very fast at first and become stable after one turning point. Moreover, the average packet delay value increases very fast as
increases. No threshold after which the delay time will become stable exists. Thus, we can conclude that the performance of the optical buffer cannot be improved by increasing the buffer size under high traffic loads (
): the
value cannot be reduced while the packet delay value increases as the buffer size grows.
Figure 7 shows that the total incoming traffic load arriving into each input concentrator should be smaller than 0.7, for the considered case that the length of the basic FDL element is same as the mean packet length. For any higher incoming traffic load value, the buffer will become an overflow and the corresponding performance will deteriorate by increasing the buffer size, i.e.,
cannot be improved while the delay time increases dramatically.
5. Discussions
In this section we discuss the economic cost of the proposed OPS node compared with the classic OPS node without input concentrators.
In order to investigate the economic cost of the proposed OPS node architecture, the number of used optical gates (OGs) and TWCs are required, since both of which are expected to be the most expensive components in the proposed architecture. For the space switch of the input concentrator, a single state implementation is considered. The number of OGs is evaluated as the product between the number of inputs and outputs, i.e.,
for each concentrator in
Figure 2. Hence, the total number of OGs for the proposed OPS node is
, while no OG is used in the OPS node without input concentrators. Furthermore, the introduction of the input concentrator reduces the number of TWCs, which is foreseen as the most expensive in the switch. Correspondingly, the proposed OPS node only needs
TWCs for handling the incoming traffic while the OPS node without input concentrators has to employ
TWCs.
Once the number of OGs and TWCs has been obtained, the cost comparison at the present time (indicated with
, time counter in year) can be carried out. The cost of an OG at the present year (
) is taken as the reference cost. The cost of a TWC is assumed to be
, with
, since a TWC is undoubtedly much more complex and expensive. In the following, we use
,
to denote the proposed OPS node and the OPS node without input concentrators at the present time, which can be expressed as:
The cost ratio
, which is defined as the economic cost ratio between the proposed OPS node with input concentrators and the OPS node without concentrators, can be obtained. The cost ratio at present time
can be defined as:
In order to evaluate how the economic cost evolves as a function of time, the extended learning curve model in [
18] is adopted. The model defines the cost evolution of a device
in this paper) and is given as:
where the parameters are:
, the production cost of the device in the reference year
;
, the relative accumulated volumes sold at the reference year;
, the growth period, which means the time for the accumulated volume to grow from 10% to 90% of the total production;
, the learning curve coefficient, which indicates the cost reduction when the production volume doubles. The values of those parameters for OG and TWC are assigned as in
Table 2 according to the technology classification in [
18]. We include the OGs into the New Fast Class (new components with technology improving fast) while TWC is included in the Emerging Medium and Emerging Slow classes, which are not already present in the market and expected to evolve in different ways. According to [
18],
is set as 0.8 for optical components and 0.7 for advanced optical components. Then
can be calculated according to its expression.
Figure 8 gives the cost ratio evaluation as a function of time (years) in case of
and
. Nine different OPS node configurations are considered. The main observation is that the cost ratio
R(
t) is smaller than 1 in future 20 years, i.e., the proposed OPS node is much cheaper than the OPS node without input concentrators even 20 years into the future. Furthermore, as with the theoretical analysis, its evaluation only depends on the fiber number and is not affected by the wavelength number per fiber for fixed
and
. We also observe that
decreases as
increases. This means that as the node degree grows (i.e., the number of input/output fibers per node grows), the proposed OPS architecture will have larger cost savings. In addition,
will diminish in the first 6 years and then increase for all nine configurations, as shown in
Figure 8. This shows that, according to the cost evolution of TWCs and OGs, the advantage in the economic cost of the proposed OPS architecture will become more and more obvious in the next 6 years. However, as the development of the production technology for TWC, whose growth period is set as 20, it will become much cheaper as time goes on. Hence, the cost ratio will start to increase after 6 years. To conclude, the proposed OPS architecture has significant improvements compared with the OPS node without input concentrators due to the reduction of TWCs, the large cost saving at the present time for all considered configurations, and better cost performance in at least the next 20 years.