SlideShare une entreprise Scribd logo
1  sur  14
Télécharger pour lire hors ligne
Performance Evaluation and Comparison of
     Westwood+, New Reno, and Vegas TCP Congestion
                        Control

                                          Luigi A. Grieco and Saverio Mascolo
                          Dipartimento di Elettrotecnica ed Elettronica, Politecnico di Bari, Italy
                                                   {a.grieco,mascolo}@poliba.it



Abstract                                                                   misinterpreted as a symptom of congestion by current TCP
TCP congestion control has been designed to ensure Internet                schemes and lead to an undue reduction of the transmission rate.
stability along with fair and efficient allocation of the network          Thus, TCP requires supplementary link layer protocols such as
bandwidth. During the last decade, many congestion control                 reliable link-layer or split-connections approach to efficiently
algorithms have been proposed to improve the classic Tahoe/Reno            operate over wireless links [2,10,21,33,34].
TCP congestion control. This paper aims at evaluating and                  Vegas TCP was the first attempt to depart from the loss-driven
comparing three control algorithms, which are Westwood+, New               paradigm of the TCP by introducing a mechanism of congestion
Reno and Vegas TCP, using both Ns-2 simulations and live                   detection before packet losses [9]. In particular, Vegas TCP
Internet measurements. Simulation scenarios are carefully                  computes the difference between the actual input rate (cwnd/RTT)
designed in order to investigate goodput, fairness and friendliness        and the expected rate (cwnd/RTTmin), where RTT is the Round
provided by each of the algorithms. Results show that Westwood+            Trip Time and RTTmin is the minimum measured round trip time,
TCP is friendly towards New Reno TCP and improves fairness in              to infer network congestion. In particular, if the difference is
bandwidth allocation whereas Vegas TCP is fair but it is not able          smaller than a threshold α then the cwnd is additively increased,
to grab its bandwidth share when coexisting with Reno or in the            whereas if the difference is greater than another threshold β then
presence of reverse traffic because of its RTT-based congestion            the cwnd is Additively Decreased; finally, if the difference is
detection mechanism. Finally results show that Westwood+                   smaller than β and greater than α, then the cwnd is kept constant
remarkably improves utilization of wireless links that are affected        [9]. In [11] it has been shown that Vegas TCP ensures network
by losses not due to congestion.                                           stability but it is not able to grab its own bandwidth share when
                                                                           interacting with algorithms that systematically hits network queue
                                                                           capacity as Reno.
1. INTRODUCTION                                                            Westwood TCP is a new congestion control algorithm that is
Internet stability is still largely based on the congestion control        based on end-to-end bandwidth estimate [12]. In particular,
algorithm proposed by Van Jacobson at the end of the eighties [1],         Westwood TCP estimates the available bandwidth by counting
which is known as Tahoe TCP, on its first modification, which is           and filtering the flow of returning ACKs and adaptively sets the
known as Reno TCP, and other variants described in [3]-[5]. The            cwnd and the ssthresh after congestion by taking into account the
Van Jacobson congestion control algorithm has been designed by             estimated bandwidth. The original bandwidth estimation
following the end-to-end principle and has been quite successful           algorithm fails to work properly in the presence of ACK
from keeping the Internet away from congestion collapse [18]-              compression [41]. Thus a slightly modified version of the
[20]. Two variables, congestion window (cwnd) and slow start               bandwidth estimation algorithm has been proposed in [14] to cope
threshold (ssthresh), are used to throttle the TCP input rate in           with ACK compression effect. We call Westwood+ the original
order to match the network available bandwidth. All these                  Westwood algorithm with the enhanced bandwidth estimate.
congestion       control    algorithms    exploit    the   Additive-       Furthermore, in [14] it has been shown via a mathematical
Increase/Multiplicative-Decrease (AIMD) paradigm, which                    analysis that Westwood+ is friendly towards Reno TCP and fairer
additively increases the cwnd to grab the available bandwidth and          than Reno in bandwidth allocation.
suddenly decreases the cwnd when network capacity is hit and               This paper aims at comparing Westwood+, New Reno and Vegas
congestion is experienced via segment losses, i.e. timeout or              TCP. New Reno is an improved version of Reno that avoids
duplicate acknowledgments. AIMD algorithms ensure network                  multiple reductions of the cwnd when several segments from the
stability but they don’t guarantee fair sharing of network resources       same window of data get lost [3]. New Reno TCP has been
[1], [6],[ 7], [21].                                                       considered because it is the leading Internet congestion control
After the introduction of the Van Jacobson algorithm research on           protocol [27]. Vegas TCP has been considered because it also
TCP congestion control become very active and several end-to-              proposes, as Westwood+, a new mechanism for throttling the
end congestion control algorithms have been proposed since then            congestion window that is based on measuring the network
to improve network stability, fair bandwidth allocation and                congestion status via RTT measurements. Moreover, Vegas TCP
resource utilization of high-speed networks and wireless networks          provides the basic ideas behind the new Fast TCP congestion
[2,3,4,9,12,14,20,32,33,39]. In fact, today TCP is not well suited         control algorithm, which has been recently proposed by
for wireless links since losses due to radio channel problems are



                                                                       1
researchers at Caltech [39]. In authors’ words, “Fast TCP is a sort       Westwood+ additively increases the cwnd as Reno, when ACKs
of high-speed version of Vegas” [40]. At the time of this paper           are received. On the other hand, when a congestion episode
Fast TCP is still in a trial phase and authors do not have released       happens, Westwood employs an adaptive setting of cwnd and
any kernel code or ns-2 implementation. Being based on RTT                ssthresh so that it can be said that Westwood+ follows an
measurements to infer congestion, it could inherit all drawbacks          Additive-Increase/Adaptive-Decrease paradigm [14].
of Vegas that will be illustrated in this paper, mainly the               It is worth noting that the adaptive decrease mechanism employed
incapacity to grab bandwidth when coexisting with Reno traffic or         by Westwood+ TCP improves the stability of the standard TCP
in the presence of reverse traffic.                                       multiplicative decrease algorithm. In fact, the adaptive window
For evaluation and comparison purposes, computer simulations              shrinking provides a congestion window that is decreased enough
using the ns-2 simulator [16], and measurements using a Linux             in the presence of heavy congestion and not too much in the
implementation, over the real Internet, have been collected. In           presence of light congestion or losses that are not due to
particular, ns-2 simulations have been carried out over single and        congestion, such as in the case of unreliable radio links.
multi bottleneck scenarios with link capacities ranging from              Moreover, the adaptive setting of the control windows increases
1Mbps to 100Mbps, for various buffer sizes and in the presence of         the fair allocation of available bandwidth to different TCP flows.
homogeneous and heterogeneous traffic sources. Moreover,                  This result can be intuitively explained by considering that the
Medium Earth Orbit (MEO) and Geosynchronous Earth Orbit                   window setting of Westwood+ TCP tracks the estimated
(GEO) satellite scenarios in the presence of lossy links have been        bandwidth so that, if this estimate is a good measurement of the
simulated. Simulation results have shown that: (1) Westwood+              fair share, then the fairness is improved. Alternatively, it could be
TCP is friendly towards New Reno TCP; (2) Weswtood+ TCP                   noted that the setting cwnd = B×RTTmin sustains a transmission
improves the fairness in bandwidth sharing with respect to New            rate (cwnd/RTT) = (B×RTTmin)/RTT that is smaller than the
Reno TCP; (3) Vegas TCP is not able to grab its own bandwidth             bandwidth B estimated at the time of congestion: as a
share when coexisting with New Reno TCP [11] or in the                    consequence, the Westwood+ TCP flow clears out its path
presence of reverse traffic; (4) Westwood+ improves the                   backlog after the setting thus leaving room in the buffers for
utilization of wireless (i.e. satellite) links with respect to both       coexisting flows, which improves statistical multiplexing and
Vegas and New Reno in the presence of uniform or bursty losses.           fairness.
The paper is organized as follows: Section 2 outlines the
Westwood+ algorithm; Section 3 compares New Reno, Vegas and
Westwood+ using the ns-2 simulator; Section 4 compares New                2.2 The end-to-end bandwidth estimate
Reno and Westwood+ using live Internet experiments; finally, the          The AIMD algorithm can be viewed as an end-to-end method to
last section draws the conclusions.                                       obtain a “rough” but robust measurement of the best effort
                                                                          bandwidth that is available along a TCP path.
                                                                          The first attempt to exploit ACK packets to improve bandwidth
2. WESTWOOD+ TCP                                                          estimation is the packet pair (PP) algorithm, which tries to infer
This section describes the Westwood+ congestion control
                                                                          the bottleneck available bandwidth at the starting of a connection
algorithm. In particular, Section 2.1 describes the control
                                                                          by measuring the interarrival time between the ACKs of two
algorithm, Section 2.2 the bandwidth estimation algorithm used
                                                                          packets that are sent back to back [23]. Hoe proposes a refined PP
by the control algorithm and section 2.3 summarizes some results
                                                                          method for estimating the available bandwidth in order to properly
on mathematical evaluation of fairness and friendliness of Reno
                                                                          initialize the ssthresh: the bandwidth is calculated by using the
and Westwood+ TCP.
                                                                          least-square estimation on the reception time of three ACKs
                                                                          corresponding to three closely-spaced packets [24]. Allman and
2.1 The algorithm                                                         Paxson evaluate the PP techniques and show that in practice they
The Westwood+ algorithm is based on end-to-end estimation of              perform less well than expected [25]. Lai and Baker propose an
the bandwidth available along the TCP connection path [12],[14].          evolution of the PP algorithm for measuring the link bandwidth in
The estimate is obtained by filtering the stream of returning ACK         FIFO-queuing networks [26]. The method consumes less network
packets and it is used to adaptively set the control windows when         bandwidth while maintaining approximately the same accuracy of
network congestion is experienced. In particular, when three              other methods, which is poor for paths longer than few hops. The
DUPACKs are received, both the congestion window (cwnd) and               inaccuracy of the algorithms based on the packet pair approach is
the slow start threshold (ssthresh) are set equal to the estimated        due to the fact that the interarrival times between consecutive
bandwidth (BWE) times the minimum measured round trip time                segments at the receiver can be very different from the interarrival
(RTTmin); when a coarse timeout expires the ssthresh is set as            times between the corresponding ACKs at the sender. It will be
before while the cwnd is set equal to one.                                shown in the following section that this effect is much more
The pseudo code of the Westwood+ algorithm is reported below:             significant in the presence of congestion along the ACK path. Jain
a)    On ACK reception:                                                   and Dovrolis propose to use streams of probing packets to
      cwnd is increased accordingly to the Reno algorithm;                measure the end-to-end available bandwidth, which is defined as
      the end-to-end bandwidth estimate BWE is computed;                  the maximum rate that the path can provide to a flow, without
b)    When 3 DUPACKs are received:                                        reducing the rate of the rest of the traffic. The estimate is
      ssthresh =max(2, (BWE* RTTmin) / seg_size);                         computed over an averaging interval [36]. Finally, they focus on
      cwnd = ssthresh;                                                    the relationship between the available bandwidth in a path they
                                                                          measure and the throughput of a persistent TCP connection. They
c)    When coarse timeout expires:
                                                                          show that the averaged throughput of a TCP connection is about
      ssthresh = max(2,(BWE* RTTmin) / seg_size);
                                                                          20-30% more than the available bandwidth measured by their tool
      cwnd = 1;
                                                                          due to the fact that the TCP probing mechanism gets more
From the pseudo-code reported above, it turns out that



                                                                      2
bandwidth than what was previously available in the path,                                                1.0E+09
grabbing part of the throughput of other connections. We notice
that the latter result is not surprising being a consequence of the                                      1.0E+08




                                                                              Bandwidth estimate (bps)
fundamental property of the TCP congestion control algorithm for                                         1.0E+07
which a new joining TCP flow must get its bandwidth share from                                           1.0E+06
existing flows.
                                                                                                         1.0E+05
A similar technique, based on using streams of probing packets,
has been proposed by Melander et. al [37]. It uses sequences of                                          1.0E+04
packet pairs at increasing rates and estimates the available                                             1.0E+03
bandwidth by comparing input and output rates of different packet                                        1.0E+02
pairs.                                                                                                                 Fair share
Westwood+ TCP proposes an end-to-end estimate of the “best-                                              1.0E+01
effort” available bandwidth by properly counting and filtering the                                                 0        20      40        60   80           100
flow of returning ACKs [12]. A sample of available bandwidth                                                                              s
 bk = d k / ∆ k is computed every RTT, where dk is the amount of                                                                    (a)
data acknowledged during the last RTT = ∆ k . The amount dk is                                           1.0E+09
determined by a proper counting procedure that considers delayed                                         1.0E+08
ACKs and duplicate ACKs: a duplicate ACK counts for one


                                                                              Bandwidth estimate (bps)
delivered segment, a delayed ACK for two segments, a                                                     1.0E+07
cumulative ACK counts for one segment or for the number of                                               1.0E+06
segments exceeding those already accounted for by previous                                               1.0E+05
duplicate acknowledgements (see [12] for more details on this).
                                                                                                         1.0E+04
Bandwidth samples bk are low-pass filtered since congestion is
due to low frequency components [31], and because of delayed                                             1.0E+03
ACK option [13,22]. In [42] the following time-invariant low-                                            1.0E+02                                        Fair share
pass filter has been proposed as an alternative to the original time-                                    1.0E+01
varying filter of Westwood TCP [12]:                                                                               0        20      40        60   80           100
 ˆ        ˆ
 bk = α ⋅ bk −1 + (1 − α ) ⋅ bk                                    (1)                                                                    s
where α is a constant set equal to 0.9. The filter (1) reveals to be
particularly suited for kernel code implementation, where floating                                       (b)
point operations should be avoided [44].                                     Figure 1. Bandwidth estimates of 20 TCP flows in the presence
It should be noted that bk are samples of used bandwidth that                of ACK compression: (a) Westwood+; (b) Westwood.
coincide with the “best-effort” available bandwidth when the
connection hits network capacity and experiences losses.                     To conclude this section we report some considerations on the
Measuring the actual rate a connection is achieving during the               end-to-end bandwidth estimate of Westwood+ and on the backlog
data transfer as done by Westwood+ TCP is a different and much               estimate of Vegas. The bandwidth estimate employed by
easier task than estimating the bandwidth available at the                   Westwood+ TCP measures the low frequency components of the
beginning of a TCP connection going over a shared FIFO queuing               used bandwidth samples bk, which coincides with the “best-effort”
network.                                                                     available bandwidth of a TCP sender at the end of the slow-
To give an insight into the bandwidth estimation algorithm, Fig. 1           start/congestion-avoidance probing phase. We remark that the
shows the bandwidth computed at congestion instants by 20                    estimate (1) is different from measuring the low frequency
Westwood+ or 20 Westwood flows sharing a 10Mbps bottleneck                   components of the sending rate cwnd/RTT, where cwnd/RTT is the
in the presence of reverse traffic contributed by 10 TCP long lived          measure of the instantaneous throughput employed by Vegas
New Reno connections. Fig. 1(a) shows that all the 20                        TCP. In fact, the Vegas actual rate cwnd/RTT is a measure of the
Westwood+ connections estimate a best-effort available                       available bandwidth that is based on the number of sent packets
bandwidth that reasonably approaches the fair share of 0.5Mbps.              (cwnd) and not on the number of acknowledged packets dk. As a
On the other hand, Fig. 1(b) highlights that Westwood                        consequence, Vegas samples do not take into account that a
overestimates up to 100 times the fair share due to ACK                      fraction of sent packets could be lost thus leading to available
compression.                                                                 bandwidth overestimate. To illustrate this point, we simulate a
                                                                             single Westwood+ connection that sends data through a 1Mbps
                                                                             bottleneck link. Fig. 2 shows the bandwidth estimates obtained by
                                                                             filtering the bk samples or the cwnd/RTT samples using the filter
                                                                             (1). Results confirm that the bandwidth estimate obtained by
                                                                             filtering the ACKs is more accurate and less oscillating than the
                                                                             one obtained by filtering the input rate, which, in fact, provides an
                                                                             overestimate of the available bandwidth.




                                                                         3
1.2E+06                                                                       based on RTT measurements [9, 22]. Packets are 1500 Bytes long
                                                BWE                             and buffer sizes are set equal to the link bandwidth times the
                                                Input Rate
                                                Bottleneck Capacity             maximum round trip propagation time unless otherwise specified.
  1.1E+06
                                                                                The initial congestion window has been set equal to 3 segments
                                                                                [43]. The receiver window size and the initial ssthresh are set
  1.0E+06                                                                       equal to a value greater than the pipe network capacity so that
                                                                                flows are always network constrained and grab the available
bps




  9.0E+05                                                                       bandwidth using slow-start when the connection starts.

  8.0E+05                                                                       3.1 A simple single-connection scenario
                                                                                In order to analyze the fundamental dynamics of the considered
                                                                                TCP congestion control algorithms, we start by considering the
  7.0E+05
                                                                                single connection scenario depicted in Fig. 3. The TCP1
            0               20   40   s    60          80         100
                                                                                connection is persistent and sends data over a 2Mbps bottleneck
Figure 2. Bandwidth estimate and input rate.                                    link. The RTT is 250ms. Ten ON-OFF New Reno TCP senders
                                                                                inject traffic along the ACK path of the single TCP connection,
2.3 Mathematical evaluation of fairness and                                     i.e. they generate reverse traffic for the TCP connection on the left
                                                                                side of Fig. 3. Reverse traffic aims at provoking congestion along
    friendliness                                                                the ACK path and at exciting ACK compression, which is
This section investigates the intra-protocol fairness in bandwidth              important to be considered since it exacerbates the bursty nature
allocation provided by Reno and Westwood+ and their inter-                      of the TCP [41].
protocol friendliness using the mathematical models of the Reno
and Westwood+ throughputs reported in [14,27,28]. In particular,
it has been shown that when the average segment loss probability                                                                       TCP1
                                                                                      TCP1
p is low, the Reno throughput is proportional to the inverse of the                                                                    Sink
                                                                                     source
average round trip time RTT and to 1 / p as follows [28]:
                                                                                                         Forward Traffic
            1       2( 1 − p )
T Reno =                                                              (2)
           RTT           p
Under the same assumptions, it has been shown that Westwood+
TCP provides the following steady state throughput [14]:                                                    Reverse Traffic            10 TCP
                                                                                     10 TCP
                1         1− p                                                                                                         sources
T West =                                                              (3)             Sinks
           RTT ⋅ Tq        p
                                                                                                   Figure 3. Simulation scenario.
where Tq is the average queuing time equal to the difference
between RTT and the minimum round trip time RTTmin.                             The TCP1 connection starts at t=0. The 10 TCP connections on
By comparing (2) and (3) it results that both throughputs of                    the backward path follow an OFF-ON-OFF-ON pattern in order to
                                                                                investigate the effect of reverse traffic. In particular, the reverse
Westwood+ and Reno depend on 1 / p , that is Westwood+ and
                                                                                traffic is ON during the intervals [250s, 500s] and [750s, 1000s]
Reno are friendly to each other. Moreover, since flows with                     and is silent during the intervals [0s, 250s] and [500s, 750s].
different RTTs experience the same mean queuing time Tq, Eq. (3)                Tables I reports the goodputs that have been measured during
shows that the throughput of Westwood+ depends on round trip                    each interval.
time as 1 / RTT whereas the throughput of Reno as 1 / RTT ,
that is, Westwood+ increases fair sharing of network capacity                            Table I. Goodputs of a single TCP connection.
between flows with different RTTs. Friendliness and fairness will                               [0s,250s]      [250s,500s]    [500,750s]      [750s,1000s]
                                                                                                 (Mbps)          (Mbps)        (Mbps)           (Mbps)
be investigated through simulations in the next sections to confirm
                                                                                New Reno          1.86            1.62           1.99             1.64
these theoretical results.                                                      Vegas             1.97            0.48           1.97             0.51
                                                                                Westwood+         1.86            1.68           1.99             1.69
3. SIMULATION-BASED COMPARISON
In this section we evaluate and compare Westwood+, New Reno
and Vegas TCP using the ns-2 simulator [16,38]. Simple scenarios                The Goodput has been computed as follows:
are considered in order to illustrate the fundamental features of the
considered protocol dynamics whereas more complex topologies                                  ( sent _ data − retransmitted _ data )
are considered to test the protocols in more realistic settings. In             Goodput =
                                                                                                          transfer _ time
particular, single, multi-bottleneck and mixed wired/wireless
scenarios are considered. Each considered scenario is particularly
useful to highlight a particular feature of the dynamic behavior of             When the reverse traffic is OFF, goodputs of all considered TCPs
the protocols or to evaluate a specific metric. In all considered               approaches the bottleneck link capacity. However, when the
scenarios, unless otherwise specified, the timestamp option is                  reverse traffic is ON Vegas provides the worst goodput whereas
enabled; destinations implement the delayed ACK option except                   Westwood+ obtains a slightly better goodput with respect to New
when Vegas is used, since its congestion avoidance mechanism is                 Reno TCP.




                                                                            4
To get a further insight, Figs. 4-5 plot the cwnd and the ssthresh                100
dynamics. New Reno and Westwood+ TCP achieve a larger cwnd                                                                     cwnd
                                                                                      90                                       ssthresh
during the time intervals when the reverse traffic is off. The
                                                                                      80
reason is that, when the reverse traffic is silent, New Reno and
Westwood+ TCP increase the cwnds up to the bandwidth delay                            70
product, which is roughly 40 segments, plus the bottleneck queue                      60




                                                                           Segments
size, which is 40 segments too, before experiencing congestion;                       50
this is shown in Figs. 4 and 5 where the cwnds of New Reno and                        40
Westwood+ systematically reach the value of 80 segments before                        30
being reduced by following the Multiplicative Decrease or                             20
Adaptive Decrease mechanism, respectively. On the other hand,                         10
when the reverse traffic is turned on, both New Reno and                               0
Westwood+ can experience a congestion episode as soon as the                               0   100 200 300 400 500 600 700 800 900 1000
cwnd is larger than the buffer size because of the burstiness of the                                            s
TCP due to ACK compression. However, it should be noted that
                                                                           Figure 6. cwnd and ssthresh of a single Vegas TCP flow with
the ssthresh dynamics of Westwood+ is less sensitive to
                                                                           reverse traffic contributed by 10 New Reno TCP flows
congestion along the ACK path with respect to New Reno because
of the bandwidth estimate used for its setting.
                                                                           As a consequence, the Vegas connection on the forward path
Regarding Vegas, Fig. 6 shows that the cwnd is constant and
                                                                           additively shrinks the cwnd thus obtaining a poor utilization of the
approaches the bandwidth delay product when the reverse traffic
                                                                           bottleneck link. We have investigated more this point to see if the
is off, thus providing an efficient link utilization. On the other         kind of used reverse traffic is of importance. Fig. 7 shows that
hand, when the reverse traffic is on, the cwnd keeps very low
                                                                           Vegas is not able to grab the bottleneck link also when the reverse
values thus preventing link utilization. The reason is that the            traffic is contributed by 10 Vegas connections, that is, Vegas does
reverse traffic provokes queuing delays along the backward path,
                                                                           not provide full link utilization whatever source type of reverse
which increases the RTT.                                                   traffic is considered. To complete the comparison we report, for
       100                                                                 this time only, the behavior of Reno TCP. Fig. 8 shows that Reno
                                                     cwnd                  is heavily affected by the presence of reverse traffic. Therefore,
           90
                                                     ssthresh
           80                                                              since now on, we will always consider the New Reno TCP.
           70                                                                     100
                                                                                                                                 cwnd
           60
Segments




                                                                                      90                                         ssthresh
           50                                                                         80
           40                                                                         70
           30                                                                         60
                                                                           Segments




           20                                                                         50
           10                                                                         40
            0                                                                         30
                0   100 200 300 400 500 600 700 800 900 1000                          20
                                     s
                                                                                      10
Figure 4. cwnd and ssthresh of a single New Reno TCP flow                              0
with reverse traffic contributed by 10 New Reno TCP flows.                                 0   100 200 300 400 500 600 700 800 900 1000
       100                                                                                                      s
                                                    cwnd
           90                                                              Figure 7. cwnd and ssthresh of a single Vegas TCP flow with
                                                    ssthresh
           80                                                              reverse traffic contributed by 10 Vegas TCP flows.
           70
           60                                                              Based on the simulations above reported and on others that we do
Segments




                                                                           not report here, we can conclude that the reverse traffic heavily
           50
                                                                           affects protocol behaviors. Therefore, the reverse traffic will be
           40
                                                                           always active in all scenarios we will consider in the sequel.
           30
                                                                           Moreover, since we have seen that the effect of reverse traffic
           20                                                              does not depend significantly on the TCP control algorithm that
           10                                                              generates it, we will consider always New Reno type reverse
            0                                                              traffic because it is the more efficient and today used TCP.
                0   100 200 300 400 500 600 700 800 900 1000
                                     s
Figure 5. cwnd and ssthresh of a single Westwood+ TCP flow
with reverse traffic contributed by 10 New Reno TCP flows.




                                                                       5
120                                                                                        1.0E+07
                    cwnd
                    ssthresh                                                                      9.0E+06
       100                                                                                        8.0E+06
                                                                                                  7.0E+06




                                                                            Total Goodput (bps)
           80
                                                                                                  6.0E+06
Segments




           60                                                                                     5.0E+06
                                                                                                  4.0E+06
           40                                                                                     3.0E+06
                                                                                                                                                                              New Reno
                                                                                                  2.0E+06                                                                     Vegas
           20                                                                                     1.0E+06                                                                     Westwood+
                                                                                                  0.0E+00
            0                                                                                                0        20        40        60        80    100    120   140   160   180   200
                0   100 200 300 400 500 600 700 800 900 1000
                                     s                                                                                                     M=No. of TCP connections

Figure 8. cwnd and ssthresh of a single Reno TCP flow with                 Figure 12. Total goodput of M TCP connections
reverse traffic contributed by 10 New Reno TCP flows.

We conclude this section by noting that Fig. 4 and 5 show that             Fig. 13 plots the Jain fairness index [17]:
Westwood+ and Reno exhibit a cyclic behavior that continuously
probes for network capacity. The main consequence of this                  J FI =
                                                                                                     (∑iM1bi )2
                                                                                                        =
                                                                                                                  2
behavior is that, when Westwood+ and New Reno coexist, they                                             M b
                                                                                                    M ∑ i =1 i
are friendly to each other; on the other hand, when Vegas coexists         where bi is the goodput of the ith connection and M are the
with New Reno or Westwood+ it gives up bandwidth to New                    connections sharing the bottleneck. The Jain fairness index
Reno or Westwood+ since Vegas lacks of this probing behavior               belongs to the interval [0,1] and increases with fairness reaching
[11].                                                                      the maximum value at one.
                                                                           Fig. 13 shows that Westwood+ improves fairness in bandwidth
3.2 Single bottleneck scenario                                             sharing with respect to New Reno TCP when M<60. The reason is
The scenario depicted in Fig. 11, where M TCP sources with                 that RTTs are uniformly spread over the interval [20+230/M,
different RTTs share a 10Mbps bottleneck link, is particularly             250]ms so that, for smaller M, RTTs are more distant, which
suited for evaluating goodput and fairness in bandwidth                    increases the unfair behavior of TCP as stated by Eqs. (2) and (3).
allocation. The M TCP flows are persistent and controlled by the           Regarding Vegas, it exhibits the best jain fairness index, but with
same algorithm in order to evaluate the intra-protocol fairness.           the lowest goodput (see Fig. 12).
RTTs are uniformly spread in the interval [20+230/M, 250]ms,
                                                                                                  1.00
with M ranging from 10 to 200, to investigate the fairness with
                                                                                                  0.98
respect to the round trip time. Simulations last 1000s during which
                                                                                                  0.96
all the TCP sources send data. In order to obtain ACK
                                                                                                  0.94
                                                                             Fairness Index




compression, 10 TCP New Reno senders inject traffic along the
                                                                                                  0.92
ACKs path of the M connections.
                                                                                                  0.90
                                                                                                  0.88
                                                          TCP1
                 TCP1                                                                             0.86
                                                                                                                                                                              New Reno
                                                          Sink
                source                                                                            0.84                                                                        Vegas
                                                                                                  0.82                                                                        Westwood+
                                 Forward Traffic
                                                          TCPM                                    0.80
                TCPM                                      Sink                                           0       20        40        60        80        100    120    140   160   180   200
                source                                                                                                                M

                                  Reverse Traffic        10 TCP            Figure. 13. Jain fairness index over a 10Mbps bottleneck.
                10 TCP
                                                         sources
                 Sinks                                                     To provide a “visual” look into the fairness issue, the sequence
                      Figure 11. Single bottleneck scenario.               numbers of 20 New Reno or 20 Westwood+ or 20 Vegas
                                                                           connections sharing a 10Mbps bottleneck are shown in Figs. 14-
Fig. 12 shows the total goodput, which is defined as the sum of            16, respectively. Figs. 14 and 15 show that the New Reno final
the goodputs of all the M TCP connections on the forward path. In          sequence numbers are spread in the interval [26693, 64238]
particular, the figure shows that when M is larger than 40 the total       whereas the Westwood+ ones are in the shorter interval [28822,
goodput approaches the bottleneck link capacity. On the other              53423]. Fig. 16 shows that Vegas is fair but provides very low
hand, when M is smaller than 40, Vegas provides a very low total           goodput due to the presence of reverse traffic (see Fig. 12).
goodput. Again this phenomenon is due to the TCP traffic on the            To summarize simulation results of this section, we can say that
backward path, which has a significant impact on Vegas TCP (see            both Westwood+ and New Reno achieve full link utilization with
also Figs. 6, 7).                                                          Westwood+ providing improved intraprotocol fairness with
                                                                           respect to New Reno. On the other hand, Vegas is fair but it is
                                                                           unable to utilize the network bandwidth.



                                                                       6
70000
                                                                                            persistent and starts at time t = 10s. Notice that the described
                                                                                            scenario is a ”worst case” scenario for the source C1 since: (1) C1
 Sequence Numbers (Segments)




                               60000                                                        starts data transmission when the network bandwidth has been
                               50000
                                                                                            grabbed by the cross traffic sources; (2) C1 has the longest RTT
                                                                                            and experiences drops at each router it goes through.
                               40000
                                                                                                                                        Sink3            C3         Sink5 C5
                               30000

                               20000
                                                                                             C1                                                                                               Sink1
                               10000
                                                                                                                                                 R            R        R         R
                                   0
                                       0   200      400       600     800        1000
                                                          s

                                Figure 14. Sequence numbers of 20 New Reno TCP                                                           C2              Sink2 C4              Sink4
                                                   connections
                               70000
                                                                                                                                                 1th hop                2th hop
 Sequence Numbers (Segments)




                               60000
                                                                                                                                      Figure 17. Multi bottleneck topology.
                               50000

                               40000                                                        We will consider the following 4 scenarios:
                                                                                            Scenario 1. The C2,C3,C4 … C2N+1 sources of cross traffic are
                               30000
                                                                                            controlled by New Reno TCP whereas the C1 connection is
                               20000                                                        controlled by New Reno, Vegas or Westwood+, respectively. This
                                                                                            scenario aims at comparing New Reno, Vegas or Westwood+,
                               10000
                                                                                            when going through an Internet dominated by New Reno traffic.
                                   0                                                        In other terms, this scenario allows us to investigate the capacity
                                       0   200      400       600     800        1000       of New Reno, Vegas and Westwood+ to grab network bandwidth
                                                          s                                 when competing with New Reno cross traffic, which is the
                                                                                            friendliness of New Reno TCP towards Vegas or Weswtood+
                               Figure 15. Sequence numbers of 20 Westwood+ TCP.             TCP.
                               70000                                                        Fig. 18 shows the goodput of the C1 connection as a function of
                                                                                            the number of hops; the fair share is 5Mbps. The goodput of the
 Sequence Numbers (Segments)




                               60000
                                                                                            C1 connection monotonically decreases with the number of hops
                               50000                                                        because of increased loss ratio and RTT. Westwood+ roughly
                                                                                            achieves the same goodput as New Reno, whereas Vegas is again
                               40000
                                                                                            not able to grab its bandwidth share in a “New Reno
                               30000                                                        environment”.
                                                                                            Fig. 19 shows that the total goodput, which is now computed as
                               20000
                                                                                            the goodput of the C1 connection + average of the C2,C4..C2N
                               10000                                                        connection goodputs, does not vary significantly with the number
                                                                                            of hops. This is due to the fact that the total goodput mainly
                                   0
                                                                                            depends on the behavior of the cross traffic connections
                                       0   200      400       600     800        1000
                                                                                            C2,C4..C2N whereas the C1 connection has a negligible impact on
                                                          s
                                                                                            the total goodput.
 Figure 16. Sequence numbers of 20 Vegas TCP connections.                                                                          1.0E+07
                                                                                              Goodput of the C1 connection (bps)




3.3 Multi bottleneck scenario
The multi-bottleneck scenario is particularly suited to investigate                                                                1.0E+06
the inter-protocol friendliness of New Reno, Westwood+ and
Vegas TCP. The topology depicted in Fig. 17 is characterized by:
                                                                                                                                   1.0E+05
(a) N hops; (b) one persistent connection C1 going through all the
N hops; (c) 2N persistent sources C2;C3;C4 … C2N+1 transmitting                                                                                      New Reno
cross traffic data over every single hop. The capacity of the                                                                      1.0E+04           Vegas
entry/exit links is 100Mbps with 20ms propagation delay. The                                                                                         Westwood+
capacity of the links connecting the routers is 10Mbps with 10ms                                                                                     Fair share
propagation delay. Router queue sizes have been set equal to 125                                                                   1.0E+03
packets, which corresponds to the bandwidth delay product of a                                                                               1       2    3       4      5     6     7    8    9   10
typical RTT of 150ms. Simulation lasts 1000s during which the                                                                                                     No. of traversed hops
cross traffic sources are always active. The connection C1 is                               Figure 18. C1 Goodput vs. number of traversed hops in the
                                                                                            presence of New Reno cross traffic.



                                                                                        7
1.0E+07                                                                                                   1.0E+07

                                       9.0E+06                                                                                                   9.0E+06
  Total Goodput (bps)




                                                                                                            Total Goodput (bps)
                                       8.0E+06                                                                                                   8.0E+06

                                       7.0E+06                                                                                                   7.0E+06

                                       6.0E+06                                                                                                   6.0E+06

                                       5.0E+06                                                                                                   5.0E+06
                                                     New Reno                                                                                                  New Reno
                                       4.0E+06       Vegas                                                                                       4.0E+06       Vegas
                                                     Westwood+                                                                                                 Westwood+
                                       3.0E+06                                                                                                   3.0E+06
                                                 1   2    3      4      5     6     7    8   9   10                                                        1   2    3      4      5     6     7    8   9   10
                                                                 No. of traversed hops                                                                                     No. of traversed hops
Figure 19. Total Goodput vs. number of traversed hops in the                                              Figure 21. Total Goodput vs. number of traversed hops in the
presence of New Reno cross traffic.                                                                       presence of Westwood+ cross traffic.

                                                                                                          Scenario 3. The C2;C3;C4 … C2N+1 sources of cross traffic are
Scenario 2. The C2;C3;C4 … C2N+1 sources of cross traffic are                                             controlled by Vegas TCP whereas the C1 connection is
controlled by Westwood+ TCP whereas the C1 connection is                                                  alternatively controlled by Reno, Vegas or Westwood+. This
alternatively controlled by New Reno, Vegas or Westwood+. This                                            scenario investigates the friendliness of Vegas towards Reno and
scenario allows us to investigate the friendliness of Westwood+                                           Westwood+ TCP. Fig. 22 shows that Reno and Westwood+
towards New Reno and Vegas TCP. Fig. 20 shows the goodput of                                              basically achieve the same goodput, which is larger than the fair
the C1 connection as a function of the number of traversed hops.                                          share for any number of traversed hops: the reason is that the
Again, it shows that Vegas is not able to grab its bandwidth share.                                       Vegas cross traffic, differently from New Reno and Westwood+,
Moreover, a comparison of the New Reno curves in Fig. 18 and                                              avoids to systematically fill the queues thus leaving more room
Fig. 20 shows that the C1 New Reno achieves a slightly greater                                            for the C1 traffic. The total goodput is very low when the C1
goodput when going through Westwood+ cross-traffic than when                                              connection is controlled by Vegas, whereas it approaches the
going through New Reno cross-traffic, that is, Westwood+ is more                                          10Mbps link capacity when the C1 connection is controlled by
than friendly towards New Reno. Also in this case, the total                                              New Reno or Westwood+ (see Fig. 23). This result again shows
goodput, which is reported in Fig. 21, does not vary significantly                                        that the Vegas cross traffic connections C2,C4..C2N are far from
with N.                                                                                                   providing an efficient utilization of the network capacity.
                                       1.0E+07                                                                                                   1.0E+07
  Goodput of the C1 connection (bps)




                                                                                                            Goodput of the C1 connection (bps)




                                       1.0E+06                                                                                                   1.0E+06


                                       1.0E+05                                                                                                   1.0E+05

                                                      New Reno                                                                                                 New Reno
                                       1.0E+04        Vegas                                                                                      1.0E+04       Vegas
                                                      Westwood+                                                                                                Westwood+
                                                      Fair share                                                                                               Fair share
                                       1.0E+03                                                                                                   1.0E+03
                                                 1   2    3      4      5     6     7    8   9   10                                                        1   2    3      4      5     6     7    8   9   10
                                                                 No. of traversed hops                                                                                     No. of traversed hops
Figure 20. C1 Goodput vs. number of traversed hops in the                                                 Figure 22. C1 Goodput vs. number of traversed hops in the
presence of Westwood+ cross traffic.                                                                      presence of Vegas cross traffic.




                                                                                                      8
presence of homogeneous cross-traffic.
                                       1.0E+07

                                       9.0E+06                                                            To summarize, results of this section have shown the inter-
                                                                                                          protocol friendliness of New Reno and Westwood+ towards each
  Total Goodput (bps)




                                       8.0E+06
                                                                                                          other that is mainly due to the fact that they employs the same
                                       7.0E+06                                                            probing mechanism. On the other hand, the rtt-based congestion
                                                                                                          detection mechanism of Vegas TCP turns out an unfriendly
                                       6.0E+06                                                            behavior of New Reno or Westwood+ towards Vegas, which is
                                       5.0E+06                                                            not able to fully utilize the network bandwidth. For these reasons,
                                                     New Reno                                             we will not consider Vegas in the sequel of the paper.
                                       4.0E+06       Vegas
                                                     Westwood+
                                       3.0E+06                                                            3.4 Wireless scenarios
                                                 1   2    3      4      5     6     7    8   9   10       This section aims at investigating the behavior of TCP over
                                                                 No. of traversed hops                    wireless links that are affected by losses not due to congestion.
Figure 23. Total Goodput vs. number of traversed hops in the                                              This case is particularly interesting since it is well known that
presence of Vegas cross traffic.                                                                          protocols that react to losses by multiplicatively decreasing the
                                                                                                          control windows do not provide efficient utilization of lossy
Scenario 4. All traffic sources are controlled by the same control                                        channels. In this scenario we consider Westwood+, New Reno and
algorithm. This is a homogeneous scenario aiming at evaluating                                            also TCP SACK to investigate the efficiency of SACK to recover
New Reno, Westwood+ and Vegas TCP in absolute terms.                                                      from sporadic random losses. Since TCP SACK by default does
Fig. 24 shows that New Reno and Westwood+ provide roughly                                                 not use delayed ACK, we consider Westwood+ and New Reno
the same goodput, which monotonically decreases when the                                                  with delayed ACK (default case) and without delayed ACK in
number of traversed hops increases, whereas Vegas achieves the                                            order to get a fair comparison.
highest goodputs when the number of traversed hops is larger than
3 because the Vegas cross traffic is not efficient as Reno or                                             3.4.1 Terrestrial scenario
Westwood+. In fact, Fig. 25 shows that the total goodput obtained                                         The first scenario we consider is the hybrid wired/wireless
by using Vegas TCP scenario is much smaller than the total                                                topology shown in Fig. 26. The TCP1 connection goes through a
goodputs obtained by New Reno or Westwood+, which means                                                   wired path terminating with a last hop wireless link. The wireless
that the Vegas cross traffic do not use the share of link capacity                                        last hop models a mobile user accessing the Internet using a radio
left unused by the C1 connection.                                                                         link as in the case of a cellular telephone system. The one way
                                                                                                          delay of the TCP1 connection is 125ms with 20ms delay on the
                                       1.0E+07
                                                                                                          wireless link, which is a 2Mbps link [2]. RTTs of the 5 cross
  Goodput of the C1 connection (bps)




                                                                                                          traffic connections and of the 10 New Reno backward traffic
                                       1.0E+06                                                            connections are uniformly spread in the intervals [66ms,250ms]
                                                                                                          and [46ms,250ms], respectively. We consider a wireless link
                                                                                                          affected by bursty segment losses in both directions. We use the
                                       1.0E+05                                                            Gilbert two state Markov chain to model the loss process [10]. In
                                                                                                          particular, we assume a segment loss probability equal to 0, when
                                                     New Reno                                             the channel is in the Good state, and equal to 0.1 when the
                                       1.0E+04       Vegas
                                                                                                          channel is in the Bad state. The permanence time in the Good
                                                     Westwood+
                                                                                                          state is assumed deterministic and equal to 1s whereas the
                                                     Fair share
                                       1.0E+03
                                                                                                          permanence time in the Bad state is assumed also deterministic
                                                 1   2    3      4      5     6     7    8   9   10
                                                                                                          but this time we consider values ranging from 0.1ms to 100 ms.
                                                                 No. of traversed hops                    When the permanence time in a state elapses, the state can transit
                                                                                                          to a Good or Bad state with a probability p=0.5. For each
Figure 24. C1 Goodput vs. number of traversed hops in the
                                                                                                          considered case, we run 10 simulations by varying the seed of the
presence of homogeneous cross-traffic.
                                                                                                          random loss process. For each value of the BAD state duration we
                                       1.0E+07                                                            report the maximum, minimum and average goodputs.
                                       9.0E+06
                                                                                                                                                   5 TCP
                                                                                                                5 TCP
  Total Goodput (bps)




                                       8.0E+06                                                                                                     Sinks
                                                                                                               sources
                                       7.0E+06                                                                                                         Wireless link
                                                                                                                            Cross Traffic
                                       6.0E+06
                                                                                                           TCP1
                                       5.0E+06
                                                     New Reno                                              source
                                       4.0E+06       Vegas                                                                                                     TCP1
                                                     Westwood+
                                                                                                                            Reverse Traffic
                                                                                                                                                 10 TCP         sink
                                       3.0E+06                                                                 10 TCP                            sources
                                                 1   2    3      4      5     6     7    8   9   10             Sinks
                                                                 No. of traversed hops
Figure 25. Total Goodput vs. number of traversed hops in the                                              Figure 26. Mixed wired/wireless scenario.



                                                                                                      9
In order to analyze only the impact of bursty losses on the TCP                             100
behavior, we have first turned off both the cross and reverse
                                                                                                90
traffic sources. This simple scenario is particularly useful to
                                                                                                80
investigate the effectiveness of the adaptive decrease paradigm
when losses not due to congestion are experienced by the TCP.                                   70
Fig. 27 (a) shows the goodput of TCP1 Westwood+ and New                                         60




                                                                                     Segments
Reno, when the delayed ACK is enabled (default case), as a                                      50
function of the duration of the BAD state. It turns out that                                    40
Westwood+ improves the goodput for a large set of channel                                       30
conditions. In particular, when the delayed ACK option is                                       20
                                                                                                                                          cwnd
enabled, Westwood+ increases the utilization of the wireless link                               10                                        ssthresh
from 70% to 230% with respect to New Reno. Fig. 27 (b) shows                                     0
the goodput of Westwood+, New Reno and TCP SACK when the                                             0   100 200 300 400 500 600 700 800 900 1000
delayed ACK is disabled (default case for TCP SACK). In this                                                              s
case SACK TCP provides a goodput similar to that of New Reno,
                                                                                     Figure 28. Cwnd and ssthresh of Westwood+ when the
whereas Westwood+ improves the link utilization with respect to
                                                                                     duration of the BAD state is 0.01s.
SACK and New Reno from 34% up to 177%. The reason is that
the adaptive setting of cwnd and ssthresh performed by                                      100
Westwood+ takes into account the bandwidth used at time of                                      90                                      cwnd
congestion, so that the TCP sender does not lose ground when in                                                                         ssthresh
                                                                                                80
the presence of losses not due to congestion. To get a further                                  70
insight into this feature, Figs. 28 and 29 report the cwnd dynamics                             60
                                                                                     Segments

of Westwood+ and New Reno, respectively, obtained when the                                      50
duration of the BAD state is 0.01s and the delayed ACK is                                       40
enabled. In this case, Westwood+ TCP provides a ssthresh
                                                                                                30
approaching the bandwidth-delay product of 40 packets, whereas
                                                                                                20
New Reno TCP provides a ssthresh that is smaller than one fourth
the bandwidth-delay product.                                                                    10
                                                                                                 0
                2.0E+06
                                                                                                     0   100 200 300 400 500 600 700 800 900 1000
                1.8E+06                                                                                                   s
                1.6E+06
                                                                                     Figure 29. Cwnd and ssthresh of New Reno when the duration
                1.4E+06                                                              of the BAD state is 0.01s.
Goodput (bps)




                1.2E+06
                1.0E+06                                                              One further point valuable of investigation is when Westwood+
                8.0E+05                                                              shares the wired portion of the network with several TCP flows on
                6.0E+05                                                              the forward and backward paths. For that purpose, we turn on the
                4.0E+05        Westwood+, DACK enabled
                                                                                     cross and reverse traffic in Fig. 26 and we measure the goodput of
                                                                                     the TCP1 connections for various values of the BAD state
                2.0E+05
                               New Reno, DACK enabled                                duration. Fig. 30 shows that the delayed ACK option plays a
                0.0E+00                                                              major role in this scenario. In fact, protocols that do not employ
                      0.0001               0.001                0.01      0.1        the delayed ACK option provides goodputs that are roughly two
                                          Duration of the BAD state (s)
                                                                                     times larger than those obtained when the delayed ACK option is
                                                  (a)
                                                                                     enabled. The reason is that the delayed ACK option slows down
                2.0E+06                                                              the TCP probing phase. In these scenarios Westwood+ TCP
                1.8E+06                                                              (DACK disabled) still improves the goodput with respect to New
                1.6E+06                                                              Reno (DACK disabled) and SACK TCP, but the improvement is
                1.4E+06                                                              now only up to roughly 20%. The reason is that in this case the
Goodput (bps)




                1.2E+06                                                              TCP1 connection loses bandwidth in favor of the cross traffic that,
                1.0E+06
                                                                                     being wired, is not penalized by losses not due to congestion.
                8.0E+05
                6.0E+05
                4.0E+05        Westwood+, DACK disabled
                               New Reno, DACK disabled
                2.0E+05
                               SACK
                0.0E+00
                      0.0001               0.001                0.01      0.1
                                          Duration of the BAD state (s)
                             (b)
Figure 27. Goodput of the TCP1 connection without reverse
traffic: (a) DACK enabled; (b) DACK disabled.




                                                                                10
900000                        Westwood+, DACK disabled                                       10000000
                                               New Reno, DACK disabled                                         9000000
                 800000                        SACK
                                                                                                               8000000
                 700000                        Westwood+, DACK enabled




                                                                                        Total Goodput (bps)
                                               New Reno, DACK enabled                                          7000000
 Goodput (bps)




                 600000                                                                                        6000000
                 500000                                                                                        5000000
                                                                                                               4000000       Westwood+, DACK disabled
                 400000
                                                                                                               3000000       Westwood+, DACK enabled
                 300000                                                                                        2000000       SACK
                                                                                                                             New Reno, DACK disabled
                 200000                                                                                        1000000
                                                                                                                             New Reno, DACK enabled
                 100000                                                                                             0
                      0.0001    0.001                0.01                   0.1                                     0.0001         0.001           0.01          0.1   1
                               Duration of the BAD state (s)                                                                          Duration of the BAD state (s)
Figure 30. Goodput of the TCP1 connection in presence of                               Figure 32. Goodput over the GEO satellite scenario.
cross and reverse traffic.

3.4.2 Satellite scenario                                                               4. LIVE MEASUREMENTS
This section investigates the performances of New Reno and                             When a new protocol is proposed, it is necessary to collect a large
Westwood+ over a large leaky pipe such as in the case of a                             set of simulation and experimental results in order to assess its
satellite scenario. For that purpose we consider the scenario in                       validity and the advantages of its deployment in the real Internet
Fig. 31 where a 10mbps bottleneck link has a one-way delay equal                       [29]. In this section we test Linux implementations of New Reno
to 275ms, which corresponds to a GEO satellite connection [32].                        and Westwood+ [44] over the real Internet. More than 4000 files,
                                                                                       with different sizes, have been uploaded via ftp from a host at the
                                                                                       Laboratory of Communication and Control at Politecnico di Bari
                                                                                       (South of Italy) to three remote servers, which are located at
                                                                                       Parma (North of Italy), Uppsala University (Sweden) and
                 20 TCP                                             20 TCP             University of California, Los Angeles (Ucla). For each upload we
                 senders                                             sinks             have measured the goodput, the number of retransmitted
                                                                                       segments, and important variables such as cwnd, sshtresh, RTT
                                                                                       and bandwidth estimate. Each measurements session collects data
                                                                                       of many file uploads, which are alternatively executed by using
                                                                        10 TCP
            10 TCP                                                                     Westwood+ or New Reno. Table 2 summarizes the main
             sinks                                                      senders
                                                                                       characteristics of the sessions from Bari to the FTP server at
                                                                                       UCLA, whereas Fig. 33 shows the average goodputs achieved
                                                                                       during data transfer.
Figure 31. GEO satellite scenario.
                                                                                       Table 2: FTP uploads from Politecnico of Bari, Italy to UCLA,
We consider 20 TCP forward connections in the presence of
                                                                                       Los Angeles.
reverse traffic contributed by 10 long-lived New Reno
connections. RTTs of the forward connections are equal to 590ms.                       date                No. of Uploads        Size of uploaded
Simulations last 1000s. We assume the same error model used in                                                                   files (MBytes)
the previous sub-section except for the BAD state duration that                        Feb,21 2003         117                   32
has been increased up to 1s. The bottleneck link experiences                           Feb,26 2003         197                   3.2
segment losses in both directions. Fig. 32 shows the goodput                           Feb,28 2003         702                   3.2
provided by the two considered TCP control algorithms. When                            Mar,14 2003         54                    32
delayed ACK is enabled, Westwood+ TCP provides a goodput                               Mar,19 2003         79                    32
improvement with respect to New Reno TCP that ranges from                              Mar,21 2003         47                    32
20% to 160%, whereas, when the delayed ACK is disabled, the
improvements of Westwood+ with respect to SACK TCP and
New Reno are up to 80%. Again, the reason is that Westwood+
adaptively reduces the cwnd and sstresh by taking into account an                      The average goodput of a measurements session is obtained by
estimate of the available bandwidth: this mitigates the impact of                      averaging the goodputs of the session uploads. It turns out that
random losses not due to congestion that provokes multiplicative                       Westwood+ TCP provides goodput improvements ranging from
reductions of New Reno control windows.                                                23% to 53% with respect to New Reno. It is also worth noting that
                                                                                       improvements provided by Westwood+ TCP are not due to a more
                                                                                       aggressive behaviour, since Fig. 34 shows that Westwood+ and
                                                                                       New Reno TCP have similar retransmission ratios.

                                                                                       Table 3 summarizes the characteristics of the measurement
                                                                                       sessions from Bari to the server at the Uppsala University. Figs.
                                                                                       35 and 36 show the average goodputs and the average



                                                                                  11
retransmission ratio in this case. Westwood+ TCP provides                                                                                                                          12
goodput improvements ranging from 4% to 40% with respect to                                                                                                                                     New Reno
                                                                                                                                                                                                Westwood+
New Reno with similar retransmission ratios.                                                                                                                                       10




                                                                                                                                                              Retransmission ratio (%)
Table 3: FTP from Politecnico of Bari, Italy to Uppsala                                                                                                                                  8
University (Sweden)
date                No. of Uploads   Size of uploaded                                                                                                                                    6
                                     files (MBytes)
Dec,14 2002         253              32                                                                                                                                                  4
Dec,17 2002         200              3.2
Jan,10 2003         100              3.2                                                                                                                                                 2
Jan,12 2003         100              3.2
Jan,13 2003         150              32                                                                                                                                                  0
                                                                                                                                                                                             21 Feb 2003 26 Feb 2003 28 Feb 2003 14 Mar 2003 19 Mar 2003 21 Mar 2003
Jan,17 2003         1000             3.2                                                                                                                                                        32MB       3.2MB       3.2MB        32MB        32MB        32MB
Feb,3 2003          278              32
Feb,7 2003          500              32                                                                                                                                                  Figure 34. Retransmission ratio during ftp to UCLA.
                                                                                                                                                           250
Table 4 summarizes the characteristics of the measurement                                                                                                                                     New Reno



                                                                                                                    Average Goodput (KBytes/s)
sessions from Bari to the FTP server located at Parma. Figs. 37                                                                                                                               Westwood+
and 38 show the goodputs and the retransmission ratios that have                                                                                           200

been measured during these data transfer. In this case the
connection has a national extension and Westwood+ and New                                                                                                  150
Reno TCP provide similar goodputs.
                                                                                                                                                           100
Table 4: FTP uploads towards the server located at Parma
(Italy)
                                                                                                                                                                     50
date              No. of Uploads      Size of uploaded
                                      files (MBytes)
                                                                                                                                                                               0
Mar,26 2003       100                 32
                                                                                                                                                                                             14 Dec   17 Dec    10 Jan   12 Jan   13 Jan   17 Jan    3 Feb    7 Feb
Mar,27 2003       98                  3.2                                                                                                                                                      2002     2002      2003     2003     2003     2003     2003     2003
Mar,28 2003       1000                3.2                                                                                                                                                     32 MB    3.2MB     3.2MB    3.2MB    32MB     3.2MB    32MB     32MB

Apr,4 2003        390                 32
Apr,7 2003        200                 3.2                                                                            Figure 35. Average goodput during ftp to Uppsala.
Apr,9 2003        200                 3.2                                                                                                                  4.5
                                                                                                                                                                                                                                                          New Reno
Apr,11 2003       1000                3.2
                                                                                                                                                 Retransmission ratio (%)




                                                                                                                                                                            4                                                                             Westwood+
                                                                                                                                                           3.5

                                                                                                                                                                            3
                                50                                                               New Reno                                                  2.5
                                                                                                 Westwood+
                                45
                                                                                                                                                                            2
                                40
  Average Goodput (KBytes/s)




                                                                                                                                                           1.5
                                35
                                                                                                                                                                            1
                                30
                                                                                                                                                           0.5
                                25
                                                                                                                                                                            0
                                20                                                                                                                                                           17 Dec    10 Jan     12 Jan    13 Jan     17 Jan 3 Feb 1003 7 Feb 2003
                                                                                                                                                                                               2002      2003       2003      2003       2003   32MB       32MB
                                15
                                                                                                                                                                                              3.2MB     3.2MB      3.2 MB    32MB       3.2MB
                                10

                                5
                                                                                                                                                                                Figure 36. Retransmission ratio during ftp to Uppsala.
                                0
                                     21 Feb 2003 26 Feb 2003 28 Feb 2003 14 Mar 2003 19 Mar 2003 21 Mar 2003
                                        32MB       3.2MB       3.2 MB       32MB        32 MB       32 MB


                               Figure 33. Average goodput during ftp to Los Angeles.




                                                                                                               12
 "Performance Evaluation and Comparison of Westwood+, New Reno and Vegas TCP Congestion Control"
 "Performance Evaluation and Comparison of Westwood+, New Reno and Vegas TCP Congestion Control"

Contenu connexe

Tendances

TCP congestion control
TCP congestion controlTCP congestion control
TCP congestion controlShubham Jain
 
Clustering based Time Slot Assignment Protocol for Improving Performance in U...
Clustering based Time Slot Assignment Protocol for Improving Performance in U...Clustering based Time Slot Assignment Protocol for Improving Performance in U...
Clustering based Time Slot Assignment Protocol for Improving Performance in U...journal ijrtem
 
IEEE CAMAD 2014_LTE Uplink Delay Constraints for Smart Grid Applications
IEEE CAMAD 2014_LTE Uplink Delay Constraints for Smart Grid ApplicationsIEEE CAMAD 2014_LTE Uplink Delay Constraints for Smart Grid Applications
IEEE CAMAD 2014_LTE Uplink Delay Constraints for Smart Grid ApplicationsSpiros Louvros
 
Iaetsd an effective approach to eliminate tcp incast
Iaetsd an effective approach to eliminate tcp incastIaetsd an effective approach to eliminate tcp incast
Iaetsd an effective approach to eliminate tcp incastIaetsd Iaetsd
 
Analytical Research of TCP Variants in Terms of Maximum Throughput
Analytical Research of TCP Variants in Terms of Maximum ThroughputAnalytical Research of TCP Variants in Terms of Maximum Throughput
Analytical Research of TCP Variants in Terms of Maximum ThroughputIJLT EMAS
 
TLS in manet
TLS in manetTLS in manet
TLS in manetJay Patel
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
 
TCP Congestion Control By Owais Jara
TCP Congestion Control By Owais JaraTCP Congestion Control By Owais Jara
TCP Congestion Control By Owais JaraOwaîs Járå
 
TCP Performance Optimizations for Wireless Sensor Networks
TCP Performance Optimizations forWireless Sensor NetworksTCP Performance Optimizations forWireless Sensor Networks
TCP Performance Optimizations for Wireless Sensor NetworksTorsten Braun, Universität Bern
 
Congestion Control in Computer Networks - ATM and TCP
Congestion Control in Computer Networks - ATM and TCPCongestion Control in Computer Networks - ATM and TCP
Congestion Control in Computer Networks - ATM and TCPAttila Balazs
 
Fault tolerant wireless sensor mac protocol for efficient collision avoidance
Fault tolerant wireless sensor mac protocol for efficient collision avoidanceFault tolerant wireless sensor mac protocol for efficient collision avoidance
Fault tolerant wireless sensor mac protocol for efficient collision avoidancegraphhoc
 
A packet drop guesser module for congestion Control protocols for high speed ...
A packet drop guesser module for congestion Control protocols for high speed ...A packet drop guesser module for congestion Control protocols for high speed ...
A packet drop guesser module for congestion Control protocols for high speed ...ijcseit
 

Tendances (17)

Bg4101335337
Bg4101335337Bg4101335337
Bg4101335337
 
TCP congestion control
TCP congestion controlTCP congestion control
TCP congestion control
 
Clustering based Time Slot Assignment Protocol for Improving Performance in U...
Clustering based Time Slot Assignment Protocol for Improving Performance in U...Clustering based Time Slot Assignment Protocol for Improving Performance in U...
Clustering based Time Slot Assignment Protocol for Improving Performance in U...
 
Jm2516821688
Jm2516821688Jm2516821688
Jm2516821688
 
TCP Congestion Control
TCP Congestion ControlTCP Congestion Control
TCP Congestion Control
 
IEEE CAMAD 2014_LTE Uplink Delay Constraints for Smart Grid Applications
IEEE CAMAD 2014_LTE Uplink Delay Constraints for Smart Grid ApplicationsIEEE CAMAD 2014_LTE Uplink Delay Constraints for Smart Grid Applications
IEEE CAMAD 2014_LTE Uplink Delay Constraints for Smart Grid Applications
 
Iaetsd an effective approach to eliminate tcp incast
Iaetsd an effective approach to eliminate tcp incastIaetsd an effective approach to eliminate tcp incast
Iaetsd an effective approach to eliminate tcp incast
 
Analytical Research of TCP Variants in Terms of Maximum Throughput
Analytical Research of TCP Variants in Terms of Maximum ThroughputAnalytical Research of TCP Variants in Terms of Maximum Throughput
Analytical Research of TCP Variants in Terms of Maximum Throughput
 
Congestion control in TCP
Congestion control in TCPCongestion control in TCP
Congestion control in TCP
 
TLS in manet
TLS in manetTLS in manet
TLS in manet
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
 
Congestion Control
Congestion ControlCongestion Control
Congestion Control
 
TCP Congestion Control By Owais Jara
TCP Congestion Control By Owais JaraTCP Congestion Control By Owais Jara
TCP Congestion Control By Owais Jara
 
TCP Performance Optimizations for Wireless Sensor Networks
TCP Performance Optimizations forWireless Sensor NetworksTCP Performance Optimizations forWireless Sensor Networks
TCP Performance Optimizations for Wireless Sensor Networks
 
Congestion Control in Computer Networks - ATM and TCP
Congestion Control in Computer Networks - ATM and TCPCongestion Control in Computer Networks - ATM and TCP
Congestion Control in Computer Networks - ATM and TCP
 
Fault tolerant wireless sensor mac protocol for efficient collision avoidance
Fault tolerant wireless sensor mac protocol for efficient collision avoidanceFault tolerant wireless sensor mac protocol for efficient collision avoidance
Fault tolerant wireless sensor mac protocol for efficient collision avoidance
 
A packet drop guesser module for congestion Control protocols for high speed ...
A packet drop guesser module for congestion Control protocols for high speed ...A packet drop guesser module for congestion Control protocols for high speed ...
A packet drop guesser module for congestion Control protocols for high speed ...
 

Similaire à "Performance Evaluation and Comparison of Westwood+, New Reno and Vegas TCP Congestion Control"

Improving Performance of TCP in Wireless Environment using TCP-P
Improving Performance of TCP in Wireless Environment using TCP-PImproving Performance of TCP in Wireless Environment using TCP-P
Improving Performance of TCP in Wireless Environment using TCP-PIDES Editor
 
Tcp santa cruz
Tcp santa cruzTcp santa cruz
Tcp santa cruzlosalamos
 
ANALYSIS AND EXPERIMENTAL EVALUATION OF THE TRANSMISSION CONTROL PROTOCOL CON...
ANALYSIS AND EXPERIMENTAL EVALUATION OF THE TRANSMISSION CONTROL PROTOCOL CON...ANALYSIS AND EXPERIMENTAL EVALUATION OF THE TRANSMISSION CONTROL PROTOCOL CON...
ANALYSIS AND EXPERIMENTAL EVALUATION OF THE TRANSMISSION CONTROL PROTOCOL CON...IRJET Journal
 
Lecture 19 22. transport protocol for ad-hoc
Lecture 19 22. transport protocol for ad-hoc Lecture 19 22. transport protocol for ad-hoc
Lecture 19 22. transport protocol for ad-hoc Chandra Meena
 
Improvement of Congestion window and Link utilization of High Speed Protocols...
Improvement of Congestion window and Link utilization of High Speed Protocols...Improvement of Congestion window and Link utilization of High Speed Protocols...
Improvement of Congestion window and Link utilization of High Speed Protocols...IOSR Journals
 
Studying_the_TCP_Flow_and_Congestion_Con.pdf
Studying_the_TCP_Flow_and_Congestion_Con.pdfStudying_the_TCP_Flow_and_Congestion_Con.pdf
Studying_the_TCP_Flow_and_Congestion_Con.pdfIUA
 
A dynamic performance-based_flow_control
A dynamic performance-based_flow_controlA dynamic performance-based_flow_control
A dynamic performance-based_flow_controlingenioustech
 
VEGAS: Better Performance than other TCP Congestion Control Algorithms on MANETs
VEGAS: Better Performance than other TCP Congestion Control Algorithms on MANETsVEGAS: Better Performance than other TCP Congestion Control Algorithms on MANETs
VEGAS: Better Performance than other TCP Congestion Control Algorithms on MANETsCSCJournals
 
Comparative Analysis of Different TCP Variants in Mobile Ad-Hoc Network
Comparative Analysis of Different TCP Variants in Mobile Ad-Hoc Network Comparative Analysis of Different TCP Variants in Mobile Ad-Hoc Network
Comparative Analysis of Different TCP Variants in Mobile Ad-Hoc Network partha pratim deb
 
A Survey of Different Approaches for Differentiating Bit Error and Congestion...
A Survey of Different Approaches for Differentiating Bit Error and Congestion...A Survey of Different Approaches for Differentiating Bit Error and Congestion...
A Survey of Different Approaches for Differentiating Bit Error and Congestion...IJERD Editor
 
Proposition of an Adaptive Retransmission Timeout for TCP in 802.11 Wireless ...
Proposition of an Adaptive Retransmission Timeout for TCP in 802.11 Wireless ...Proposition of an Adaptive Retransmission Timeout for TCP in 802.11 Wireless ...
Proposition of an Adaptive Retransmission Timeout for TCP in 802.11 Wireless ...IJERA Editor
 
An Implementation and Analysis of RTS/CTS Mechanism for Data Transfer in Wire...
An Implementation and Analysis of RTS/CTS Mechanism for Data Transfer in Wire...An Implementation and Analysis of RTS/CTS Mechanism for Data Transfer in Wire...
An Implementation and Analysis of RTS/CTS Mechanism for Data Transfer in Wire...iosrjce
 
Performance Evaluation of High Speed Congestion Control Protocols
Performance Evaluation of High Speed Congestion Control  ProtocolsPerformance Evaluation of High Speed Congestion Control  Protocols
Performance Evaluation of High Speed Congestion Control ProtocolsIOSR Journals
 
4..[26 36]signal strength based congestion control in manet
4..[26 36]signal strength based congestion control in manet4..[26 36]signal strength based congestion control in manet
4..[26 36]signal strength based congestion control in manetAlexander Decker
 
11.signal strength based congestion control in manet
11.signal strength based congestion control in manet11.signal strength based congestion control in manet
11.signal strength based congestion control in manetAlexander Decker
 

Similaire à "Performance Evaluation and Comparison of Westwood+, New Reno and Vegas TCP Congestion Control" (20)

Improving Performance of TCP in Wireless Environment using TCP-P
Improving Performance of TCP in Wireless Environment using TCP-PImproving Performance of TCP in Wireless Environment using TCP-P
Improving Performance of TCP in Wireless Environment using TCP-P
 
U01725129138
U01725129138U01725129138
U01725129138
 
Tcp santa cruz
Tcp santa cruzTcp santa cruz
Tcp santa cruz
 
Cp Fu03a
Cp Fu03aCp Fu03a
Cp Fu03a
 
ANALYSIS AND EXPERIMENTAL EVALUATION OF THE TRANSMISSION CONTROL PROTOCOL CON...
ANALYSIS AND EXPERIMENTAL EVALUATION OF THE TRANSMISSION CONTROL PROTOCOL CON...ANALYSIS AND EXPERIMENTAL EVALUATION OF THE TRANSMISSION CONTROL PROTOCOL CON...
ANALYSIS AND EXPERIMENTAL EVALUATION OF THE TRANSMISSION CONTROL PROTOCOL CON...
 
Lecture 19 22. transport protocol for ad-hoc
Lecture 19 22. transport protocol for ad-hoc Lecture 19 22. transport protocol for ad-hoc
Lecture 19 22. transport protocol for ad-hoc
 
Improvement of Congestion window and Link utilization of High Speed Protocols...
Improvement of Congestion window and Link utilization of High Speed Protocols...Improvement of Congestion window and Link utilization of High Speed Protocols...
Improvement of Congestion window and Link utilization of High Speed Protocols...
 
Studying_the_TCP_Flow_and_Congestion_Con.pdf
Studying_the_TCP_Flow_and_Congestion_Con.pdfStudying_the_TCP_Flow_and_Congestion_Con.pdf
Studying_the_TCP_Flow_and_Congestion_Con.pdf
 
A dynamic performance-based_flow_control
A dynamic performance-based_flow_controlA dynamic performance-based_flow_control
A dynamic performance-based_flow_control
 
VEGAS: Better Performance than other TCP Congestion Control Algorithms on MANETs
VEGAS: Better Performance than other TCP Congestion Control Algorithms on MANETsVEGAS: Better Performance than other TCP Congestion Control Algorithms on MANETs
VEGAS: Better Performance than other TCP Congestion Control Algorithms on MANETs
 
F45013942
F45013942F45013942
F45013942
 
Comparative Analysis of Different TCP Variants in Mobile Ad-Hoc Network
Comparative Analysis of Different TCP Variants in Mobile Ad-Hoc Network Comparative Analysis of Different TCP Variants in Mobile Ad-Hoc Network
Comparative Analysis of Different TCP Variants in Mobile Ad-Hoc Network
 
A Survey of Different Approaches for Differentiating Bit Error and Congestion...
A Survey of Different Approaches for Differentiating Bit Error and Congestion...A Survey of Different Approaches for Differentiating Bit Error and Congestion...
A Survey of Different Approaches for Differentiating Bit Error and Congestion...
 
Proposition of an Adaptive Retransmission Timeout for TCP in 802.11 Wireless ...
Proposition of an Adaptive Retransmission Timeout for TCP in 802.11 Wireless ...Proposition of an Adaptive Retransmission Timeout for TCP in 802.11 Wireless ...
Proposition of an Adaptive Retransmission Timeout for TCP in 802.11 Wireless ...
 
B010611015
B010611015B010611015
B010611015
 
An Implementation and Analysis of RTS/CTS Mechanism for Data Transfer in Wire...
An Implementation and Analysis of RTS/CTS Mechanism for Data Transfer in Wire...An Implementation and Analysis of RTS/CTS Mechanism for Data Transfer in Wire...
An Implementation and Analysis of RTS/CTS Mechanism for Data Transfer in Wire...
 
Performance Evaluation of High Speed Congestion Control Protocols
Performance Evaluation of High Speed Congestion Control  ProtocolsPerformance Evaluation of High Speed Congestion Control  Protocols
Performance Evaluation of High Speed Congestion Control Protocols
 
4..[26 36]signal strength based congestion control in manet
4..[26 36]signal strength based congestion control in manet4..[26 36]signal strength based congestion control in manet
4..[26 36]signal strength based congestion control in manet
 
11.signal strength based congestion control in manet
11.signal strength based congestion control in manet11.signal strength based congestion control in manet
11.signal strength based congestion control in manet
 
Sky x technology
Sky x technologySky x technology
Sky x technology
 

Plus de losalamos

Exp user guide_4.6
Exp user guide_4.6Exp user guide_4.6
Exp user guide_4.6losalamos
 
Security flawsu pnp
Security flawsu pnpSecurity flawsu pnp
Security flawsu pnplosalamos
 
Zmap fast internet wide scanning and its security applications
Zmap fast internet wide scanning and its security applicationsZmap fast internet wide scanning and its security applications
Zmap fast internet wide scanning and its security applicationslosalamos
 
Effective Java Second Edition
Effective Java Second EditionEffective Java Second Edition
Effective Java Second Editionlosalamos
 
Swf File Format Spec V10
Swf File Format Spec V10Swf File Format Spec V10
Swf File Format Spec V10losalamos
 
Developing Adobe AIR 1.5 Applications with HTML and Ajax
Developing Adobe AIR 1.5 Applications with HTML and AjaxDeveloping Adobe AIR 1.5 Applications with HTML and Ajax
Developing Adobe AIR 1.5 Applications with HTML and Ajaxlosalamos
 
Bshield osdi2006
Bshield osdi2006Bshield osdi2006
Bshield osdi2006losalamos
 
"Start-up dynamics of TCP's Congestion Control and Avoidance Schemes"
"Start-up dynamics of TCP's Congestion Control and Avoidance Schemes""Start-up dynamics of TCP's Congestion Control and Avoidance Schemes"
"Start-up dynamics of TCP's Congestion Control and Avoidance Schemes"losalamos
 
Conficker summary-review-07may10-en
Conficker summary-review-07may10-enConficker summary-review-07may10-en
Conficker summary-review-07may10-enlosalamos
 
Jscriptdeviationsfromes3
Jscriptdeviationsfromes3Jscriptdeviationsfromes3
Jscriptdeviationsfromes3losalamos
 
Sourcefire Vulnerability Research Team Labs
Sourcefire Vulnerability Research Team LabsSourcefire Vulnerability Research Team Labs
Sourcefire Vulnerability Research Team Labslosalamos
 
Mixing Games And Applications
Mixing Games And ApplicationsMixing Games And Applications
Mixing Games And Applicationslosalamos
 
Astaro Orange Paper Oss Myths Dispelled
Astaro Orange Paper Oss Myths DispelledAstaro Orange Paper Oss Myths Dispelled
Astaro Orange Paper Oss Myths Dispelledlosalamos
 
Conociendo Db2 Express V9.5
Conociendo Db2 Express V9.5Conociendo Db2 Express V9.5
Conociendo Db2 Express V9.5losalamos
 
Mision De Cada Signo
Mision De Cada SignoMision De Cada Signo
Mision De Cada Signolosalamos
 
Lectura+Y+Mujeres%2c+Im%C3%81 Genes+De+Una+Aventura
Lectura+Y+Mujeres%2c+Im%C3%81 Genes+De+Una+AventuraLectura+Y+Mujeres%2c+Im%C3%81 Genes+De+Una+Aventura
Lectura+Y+Mujeres%2c+Im%C3%81 Genes+De+Una+Aventuralosalamos
 
Buenas Maneras
Buenas ManerasBuenas Maneras
Buenas Maneraslosalamos
 

Plus de losalamos (20)

Exp user guide_4.6
Exp user guide_4.6Exp user guide_4.6
Exp user guide_4.6
 
Remote api
Remote apiRemote api
Remote api
 
Security flawsu pnp
Security flawsu pnpSecurity flawsu pnp
Security flawsu pnp
 
Zmap fast internet wide scanning and its security applications
Zmap fast internet wide scanning and its security applicationsZmap fast internet wide scanning and its security applications
Zmap fast internet wide scanning and its security applications
 
Effective Java Second Edition
Effective Java Second EditionEffective Java Second Edition
Effective Java Second Edition
 
Swf File Format Spec V10
Swf File Format Spec V10Swf File Format Spec V10
Swf File Format Spec V10
 
Developing Adobe AIR 1.5 Applications with HTML and Ajax
Developing Adobe AIR 1.5 Applications with HTML and AjaxDeveloping Adobe AIR 1.5 Applications with HTML and Ajax
Developing Adobe AIR 1.5 Applications with HTML and Ajax
 
Bshield osdi2006
Bshield osdi2006Bshield osdi2006
Bshield osdi2006
 
"Start-up dynamics of TCP's Congestion Control and Avoidance Schemes"
"Start-up dynamics of TCP's Congestion Control and Avoidance Schemes""Start-up dynamics of TCP's Congestion Control and Avoidance Schemes"
"Start-up dynamics of TCP's Congestion Control and Avoidance Schemes"
 
Conficker summary-review-07may10-en
Conficker summary-review-07may10-enConficker summary-review-07may10-en
Conficker summary-review-07may10-en
 
Jscriptdeviationsfromes3
Jscriptdeviationsfromes3Jscriptdeviationsfromes3
Jscriptdeviationsfromes3
 
Sourcefire Vulnerability Research Team Labs
Sourcefire Vulnerability Research Team LabsSourcefire Vulnerability Research Team Labs
Sourcefire Vulnerability Research Team Labs
 
Mixing Games And Applications
Mixing Games And ApplicationsMixing Games And Applications
Mixing Games And Applications
 
Cryptointro
CryptointroCryptointro
Cryptointro
 
Astaro Orange Paper Oss Myths Dispelled
Astaro Orange Paper Oss Myths DispelledAstaro Orange Paper Oss Myths Dispelled
Astaro Orange Paper Oss Myths Dispelled
 
Apache Eng
Apache EngApache Eng
Apache Eng
 
Conociendo Db2 Express V9.5
Conociendo Db2 Express V9.5Conociendo Db2 Express V9.5
Conociendo Db2 Express V9.5
 
Mision De Cada Signo
Mision De Cada SignoMision De Cada Signo
Mision De Cada Signo
 
Lectura+Y+Mujeres%2c+Im%C3%81 Genes+De+Una+Aventura
Lectura+Y+Mujeres%2c+Im%C3%81 Genes+De+Una+AventuraLectura+Y+Mujeres%2c+Im%C3%81 Genes+De+Una+Aventura
Lectura+Y+Mujeres%2c+Im%C3%81 Genes+De+Una+Aventura
 
Buenas Maneras
Buenas ManerasBuenas Maneras
Buenas Maneras
 

Dernier

DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):comworks
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostZilliz
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DaySri Ambati
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfRankYa
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfPrecisely
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 

Dernier (20)

DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdf
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 

"Performance Evaluation and Comparison of Westwood+, New Reno and Vegas TCP Congestion Control"

  • 1. Performance Evaluation and Comparison of Westwood+, New Reno, and Vegas TCP Congestion Control Luigi A. Grieco and Saverio Mascolo Dipartimento di Elettrotecnica ed Elettronica, Politecnico di Bari, Italy {a.grieco,mascolo}@poliba.it Abstract misinterpreted as a symptom of congestion by current TCP TCP congestion control has been designed to ensure Internet schemes and lead to an undue reduction of the transmission rate. stability along with fair and efficient allocation of the network Thus, TCP requires supplementary link layer protocols such as bandwidth. During the last decade, many congestion control reliable link-layer or split-connections approach to efficiently algorithms have been proposed to improve the classic Tahoe/Reno operate over wireless links [2,10,21,33,34]. TCP congestion control. This paper aims at evaluating and Vegas TCP was the first attempt to depart from the loss-driven comparing three control algorithms, which are Westwood+, New paradigm of the TCP by introducing a mechanism of congestion Reno and Vegas TCP, using both Ns-2 simulations and live detection before packet losses [9]. In particular, Vegas TCP Internet measurements. Simulation scenarios are carefully computes the difference between the actual input rate (cwnd/RTT) designed in order to investigate goodput, fairness and friendliness and the expected rate (cwnd/RTTmin), where RTT is the Round provided by each of the algorithms. Results show that Westwood+ Trip Time and RTTmin is the minimum measured round trip time, TCP is friendly towards New Reno TCP and improves fairness in to infer network congestion. In particular, if the difference is bandwidth allocation whereas Vegas TCP is fair but it is not able smaller than a threshold α then the cwnd is additively increased, to grab its bandwidth share when coexisting with Reno or in the whereas if the difference is greater than another threshold β then presence of reverse traffic because of its RTT-based congestion the cwnd is Additively Decreased; finally, if the difference is detection mechanism. Finally results show that Westwood+ smaller than β and greater than α, then the cwnd is kept constant remarkably improves utilization of wireless links that are affected [9]. In [11] it has been shown that Vegas TCP ensures network by losses not due to congestion. stability but it is not able to grab its own bandwidth share when interacting with algorithms that systematically hits network queue capacity as Reno. 1. INTRODUCTION Westwood TCP is a new congestion control algorithm that is Internet stability is still largely based on the congestion control based on end-to-end bandwidth estimate [12]. In particular, algorithm proposed by Van Jacobson at the end of the eighties [1], Westwood TCP estimates the available bandwidth by counting which is known as Tahoe TCP, on its first modification, which is and filtering the flow of returning ACKs and adaptively sets the known as Reno TCP, and other variants described in [3]-[5]. The cwnd and the ssthresh after congestion by taking into account the Van Jacobson congestion control algorithm has been designed by estimated bandwidth. The original bandwidth estimation following the end-to-end principle and has been quite successful algorithm fails to work properly in the presence of ACK from keeping the Internet away from congestion collapse [18]- compression [41]. Thus a slightly modified version of the [20]. Two variables, congestion window (cwnd) and slow start bandwidth estimation algorithm has been proposed in [14] to cope threshold (ssthresh), are used to throttle the TCP input rate in with ACK compression effect. We call Westwood+ the original order to match the network available bandwidth. All these Westwood algorithm with the enhanced bandwidth estimate. congestion control algorithms exploit the Additive- Furthermore, in [14] it has been shown via a mathematical Increase/Multiplicative-Decrease (AIMD) paradigm, which analysis that Westwood+ is friendly towards Reno TCP and fairer additively increases the cwnd to grab the available bandwidth and than Reno in bandwidth allocation. suddenly decreases the cwnd when network capacity is hit and This paper aims at comparing Westwood+, New Reno and Vegas congestion is experienced via segment losses, i.e. timeout or TCP. New Reno is an improved version of Reno that avoids duplicate acknowledgments. AIMD algorithms ensure network multiple reductions of the cwnd when several segments from the stability but they don’t guarantee fair sharing of network resources same window of data get lost [3]. New Reno TCP has been [1], [6],[ 7], [21]. considered because it is the leading Internet congestion control After the introduction of the Van Jacobson algorithm research on protocol [27]. Vegas TCP has been considered because it also TCP congestion control become very active and several end-to- proposes, as Westwood+, a new mechanism for throttling the end congestion control algorithms have been proposed since then congestion window that is based on measuring the network to improve network stability, fair bandwidth allocation and congestion status via RTT measurements. Moreover, Vegas TCP resource utilization of high-speed networks and wireless networks provides the basic ideas behind the new Fast TCP congestion [2,3,4,9,12,14,20,32,33,39]. In fact, today TCP is not well suited control algorithm, which has been recently proposed by for wireless links since losses due to radio channel problems are 1
  • 2. researchers at Caltech [39]. In authors’ words, “Fast TCP is a sort Westwood+ additively increases the cwnd as Reno, when ACKs of high-speed version of Vegas” [40]. At the time of this paper are received. On the other hand, when a congestion episode Fast TCP is still in a trial phase and authors do not have released happens, Westwood employs an adaptive setting of cwnd and any kernel code or ns-2 implementation. Being based on RTT ssthresh so that it can be said that Westwood+ follows an measurements to infer congestion, it could inherit all drawbacks Additive-Increase/Adaptive-Decrease paradigm [14]. of Vegas that will be illustrated in this paper, mainly the It is worth noting that the adaptive decrease mechanism employed incapacity to grab bandwidth when coexisting with Reno traffic or by Westwood+ TCP improves the stability of the standard TCP in the presence of reverse traffic. multiplicative decrease algorithm. In fact, the adaptive window For evaluation and comparison purposes, computer simulations shrinking provides a congestion window that is decreased enough using the ns-2 simulator [16], and measurements using a Linux in the presence of heavy congestion and not too much in the implementation, over the real Internet, have been collected. In presence of light congestion or losses that are not due to particular, ns-2 simulations have been carried out over single and congestion, such as in the case of unreliable radio links. multi bottleneck scenarios with link capacities ranging from Moreover, the adaptive setting of the control windows increases 1Mbps to 100Mbps, for various buffer sizes and in the presence of the fair allocation of available bandwidth to different TCP flows. homogeneous and heterogeneous traffic sources. Moreover, This result can be intuitively explained by considering that the Medium Earth Orbit (MEO) and Geosynchronous Earth Orbit window setting of Westwood+ TCP tracks the estimated (GEO) satellite scenarios in the presence of lossy links have been bandwidth so that, if this estimate is a good measurement of the simulated. Simulation results have shown that: (1) Westwood+ fair share, then the fairness is improved. Alternatively, it could be TCP is friendly towards New Reno TCP; (2) Weswtood+ TCP noted that the setting cwnd = B×RTTmin sustains a transmission improves the fairness in bandwidth sharing with respect to New rate (cwnd/RTT) = (B×RTTmin)/RTT that is smaller than the Reno TCP; (3) Vegas TCP is not able to grab its own bandwidth bandwidth B estimated at the time of congestion: as a share when coexisting with New Reno TCP [11] or in the consequence, the Westwood+ TCP flow clears out its path presence of reverse traffic; (4) Westwood+ improves the backlog after the setting thus leaving room in the buffers for utilization of wireless (i.e. satellite) links with respect to both coexisting flows, which improves statistical multiplexing and Vegas and New Reno in the presence of uniform or bursty losses. fairness. The paper is organized as follows: Section 2 outlines the Westwood+ algorithm; Section 3 compares New Reno, Vegas and Westwood+ using the ns-2 simulator; Section 4 compares New 2.2 The end-to-end bandwidth estimate Reno and Westwood+ using live Internet experiments; finally, the The AIMD algorithm can be viewed as an end-to-end method to last section draws the conclusions. obtain a “rough” but robust measurement of the best effort bandwidth that is available along a TCP path. The first attempt to exploit ACK packets to improve bandwidth 2. WESTWOOD+ TCP estimation is the packet pair (PP) algorithm, which tries to infer This section describes the Westwood+ congestion control the bottleneck available bandwidth at the starting of a connection algorithm. In particular, Section 2.1 describes the control by measuring the interarrival time between the ACKs of two algorithm, Section 2.2 the bandwidth estimation algorithm used packets that are sent back to back [23]. Hoe proposes a refined PP by the control algorithm and section 2.3 summarizes some results method for estimating the available bandwidth in order to properly on mathematical evaluation of fairness and friendliness of Reno initialize the ssthresh: the bandwidth is calculated by using the and Westwood+ TCP. least-square estimation on the reception time of three ACKs corresponding to three closely-spaced packets [24]. Allman and 2.1 The algorithm Paxson evaluate the PP techniques and show that in practice they The Westwood+ algorithm is based on end-to-end estimation of perform less well than expected [25]. Lai and Baker propose an the bandwidth available along the TCP connection path [12],[14]. evolution of the PP algorithm for measuring the link bandwidth in The estimate is obtained by filtering the stream of returning ACK FIFO-queuing networks [26]. The method consumes less network packets and it is used to adaptively set the control windows when bandwidth while maintaining approximately the same accuracy of network congestion is experienced. In particular, when three other methods, which is poor for paths longer than few hops. The DUPACKs are received, both the congestion window (cwnd) and inaccuracy of the algorithms based on the packet pair approach is the slow start threshold (ssthresh) are set equal to the estimated due to the fact that the interarrival times between consecutive bandwidth (BWE) times the minimum measured round trip time segments at the receiver can be very different from the interarrival (RTTmin); when a coarse timeout expires the ssthresh is set as times between the corresponding ACKs at the sender. It will be before while the cwnd is set equal to one. shown in the following section that this effect is much more The pseudo code of the Westwood+ algorithm is reported below: significant in the presence of congestion along the ACK path. Jain a) On ACK reception: and Dovrolis propose to use streams of probing packets to cwnd is increased accordingly to the Reno algorithm; measure the end-to-end available bandwidth, which is defined as the end-to-end bandwidth estimate BWE is computed; the maximum rate that the path can provide to a flow, without b) When 3 DUPACKs are received: reducing the rate of the rest of the traffic. The estimate is ssthresh =max(2, (BWE* RTTmin) / seg_size); computed over an averaging interval [36]. Finally, they focus on cwnd = ssthresh; the relationship between the available bandwidth in a path they measure and the throughput of a persistent TCP connection. They c) When coarse timeout expires: show that the averaged throughput of a TCP connection is about ssthresh = max(2,(BWE* RTTmin) / seg_size); 20-30% more than the available bandwidth measured by their tool cwnd = 1; due to the fact that the TCP probing mechanism gets more From the pseudo-code reported above, it turns out that 2
  • 3. bandwidth than what was previously available in the path, 1.0E+09 grabbing part of the throughput of other connections. We notice that the latter result is not surprising being a consequence of the 1.0E+08 Bandwidth estimate (bps) fundamental property of the TCP congestion control algorithm for 1.0E+07 which a new joining TCP flow must get its bandwidth share from 1.0E+06 existing flows. 1.0E+05 A similar technique, based on using streams of probing packets, has been proposed by Melander et. al [37]. It uses sequences of 1.0E+04 packet pairs at increasing rates and estimates the available 1.0E+03 bandwidth by comparing input and output rates of different packet 1.0E+02 pairs. Fair share Westwood+ TCP proposes an end-to-end estimate of the “best- 1.0E+01 effort” available bandwidth by properly counting and filtering the 0 20 40 60 80 100 flow of returning ACKs [12]. A sample of available bandwidth s bk = d k / ∆ k is computed every RTT, where dk is the amount of (a) data acknowledged during the last RTT = ∆ k . The amount dk is 1.0E+09 determined by a proper counting procedure that considers delayed 1.0E+08 ACKs and duplicate ACKs: a duplicate ACK counts for one Bandwidth estimate (bps) delivered segment, a delayed ACK for two segments, a 1.0E+07 cumulative ACK counts for one segment or for the number of 1.0E+06 segments exceeding those already accounted for by previous 1.0E+05 duplicate acknowledgements (see [12] for more details on this). 1.0E+04 Bandwidth samples bk are low-pass filtered since congestion is due to low frequency components [31], and because of delayed 1.0E+03 ACK option [13,22]. In [42] the following time-invariant low- 1.0E+02 Fair share pass filter has been proposed as an alternative to the original time- 1.0E+01 varying filter of Westwood TCP [12]: 0 20 40 60 80 100 ˆ ˆ bk = α ⋅ bk −1 + (1 − α ) ⋅ bk (1) s where α is a constant set equal to 0.9. The filter (1) reveals to be particularly suited for kernel code implementation, where floating (b) point operations should be avoided [44]. Figure 1. Bandwidth estimates of 20 TCP flows in the presence It should be noted that bk are samples of used bandwidth that of ACK compression: (a) Westwood+; (b) Westwood. coincide with the “best-effort” available bandwidth when the connection hits network capacity and experiences losses. To conclude this section we report some considerations on the Measuring the actual rate a connection is achieving during the end-to-end bandwidth estimate of Westwood+ and on the backlog data transfer as done by Westwood+ TCP is a different and much estimate of Vegas. The bandwidth estimate employed by easier task than estimating the bandwidth available at the Westwood+ TCP measures the low frequency components of the beginning of a TCP connection going over a shared FIFO queuing used bandwidth samples bk, which coincides with the “best-effort” network. available bandwidth of a TCP sender at the end of the slow- To give an insight into the bandwidth estimation algorithm, Fig. 1 start/congestion-avoidance probing phase. We remark that the shows the bandwidth computed at congestion instants by 20 estimate (1) is different from measuring the low frequency Westwood+ or 20 Westwood flows sharing a 10Mbps bottleneck components of the sending rate cwnd/RTT, where cwnd/RTT is the in the presence of reverse traffic contributed by 10 TCP long lived measure of the instantaneous throughput employed by Vegas New Reno connections. Fig. 1(a) shows that all the 20 TCP. In fact, the Vegas actual rate cwnd/RTT is a measure of the Westwood+ connections estimate a best-effort available available bandwidth that is based on the number of sent packets bandwidth that reasonably approaches the fair share of 0.5Mbps. (cwnd) and not on the number of acknowledged packets dk. As a On the other hand, Fig. 1(b) highlights that Westwood consequence, Vegas samples do not take into account that a overestimates up to 100 times the fair share due to ACK fraction of sent packets could be lost thus leading to available compression. bandwidth overestimate. To illustrate this point, we simulate a single Westwood+ connection that sends data through a 1Mbps bottleneck link. Fig. 2 shows the bandwidth estimates obtained by filtering the bk samples or the cwnd/RTT samples using the filter (1). Results confirm that the bandwidth estimate obtained by filtering the ACKs is more accurate and less oscillating than the one obtained by filtering the input rate, which, in fact, provides an overestimate of the available bandwidth. 3
  • 4. 1.2E+06 based on RTT measurements [9, 22]. Packets are 1500 Bytes long BWE and buffer sizes are set equal to the link bandwidth times the Input Rate Bottleneck Capacity maximum round trip propagation time unless otherwise specified. 1.1E+06 The initial congestion window has been set equal to 3 segments [43]. The receiver window size and the initial ssthresh are set 1.0E+06 equal to a value greater than the pipe network capacity so that flows are always network constrained and grab the available bps 9.0E+05 bandwidth using slow-start when the connection starts. 8.0E+05 3.1 A simple single-connection scenario In order to analyze the fundamental dynamics of the considered TCP congestion control algorithms, we start by considering the 7.0E+05 single connection scenario depicted in Fig. 3. The TCP1 0 20 40 s 60 80 100 connection is persistent and sends data over a 2Mbps bottleneck Figure 2. Bandwidth estimate and input rate. link. The RTT is 250ms. Ten ON-OFF New Reno TCP senders inject traffic along the ACK path of the single TCP connection, 2.3 Mathematical evaluation of fairness and i.e. they generate reverse traffic for the TCP connection on the left side of Fig. 3. Reverse traffic aims at provoking congestion along friendliness the ACK path and at exciting ACK compression, which is This section investigates the intra-protocol fairness in bandwidth important to be considered since it exacerbates the bursty nature allocation provided by Reno and Westwood+ and their inter- of the TCP [41]. protocol friendliness using the mathematical models of the Reno and Westwood+ throughputs reported in [14,27,28]. In particular, it has been shown that when the average segment loss probability TCP1 TCP1 p is low, the Reno throughput is proportional to the inverse of the Sink source average round trip time RTT and to 1 / p as follows [28]: Forward Traffic 1 2( 1 − p ) T Reno = (2) RTT p Under the same assumptions, it has been shown that Westwood+ TCP provides the following steady state throughput [14]: Reverse Traffic 10 TCP 10 TCP 1 1− p sources T West = (3) Sinks RTT ⋅ Tq p Figure 3. Simulation scenario. where Tq is the average queuing time equal to the difference between RTT and the minimum round trip time RTTmin. The TCP1 connection starts at t=0. The 10 TCP connections on By comparing (2) and (3) it results that both throughputs of the backward path follow an OFF-ON-OFF-ON pattern in order to investigate the effect of reverse traffic. In particular, the reverse Westwood+ and Reno depend on 1 / p , that is Westwood+ and traffic is ON during the intervals [250s, 500s] and [750s, 1000s] Reno are friendly to each other. Moreover, since flows with and is silent during the intervals [0s, 250s] and [500s, 750s]. different RTTs experience the same mean queuing time Tq, Eq. (3) Tables I reports the goodputs that have been measured during shows that the throughput of Westwood+ depends on round trip each interval. time as 1 / RTT whereas the throughput of Reno as 1 / RTT , that is, Westwood+ increases fair sharing of network capacity Table I. Goodputs of a single TCP connection. between flows with different RTTs. Friendliness and fairness will [0s,250s] [250s,500s] [500,750s] [750s,1000s] (Mbps) (Mbps) (Mbps) (Mbps) be investigated through simulations in the next sections to confirm New Reno 1.86 1.62 1.99 1.64 these theoretical results. Vegas 1.97 0.48 1.97 0.51 Westwood+ 1.86 1.68 1.99 1.69 3. SIMULATION-BASED COMPARISON In this section we evaluate and compare Westwood+, New Reno and Vegas TCP using the ns-2 simulator [16,38]. Simple scenarios The Goodput has been computed as follows: are considered in order to illustrate the fundamental features of the considered protocol dynamics whereas more complex topologies ( sent _ data − retransmitted _ data ) are considered to test the protocols in more realistic settings. In Goodput = transfer _ time particular, single, multi-bottleneck and mixed wired/wireless scenarios are considered. Each considered scenario is particularly useful to highlight a particular feature of the dynamic behavior of When the reverse traffic is OFF, goodputs of all considered TCPs the protocols or to evaluate a specific metric. In all considered approaches the bottleneck link capacity. However, when the scenarios, unless otherwise specified, the timestamp option is reverse traffic is ON Vegas provides the worst goodput whereas enabled; destinations implement the delayed ACK option except Westwood+ obtains a slightly better goodput with respect to New when Vegas is used, since its congestion avoidance mechanism is Reno TCP. 4
  • 5. To get a further insight, Figs. 4-5 plot the cwnd and the ssthresh 100 dynamics. New Reno and Westwood+ TCP achieve a larger cwnd cwnd 90 ssthresh during the time intervals when the reverse traffic is off. The 80 reason is that, when the reverse traffic is silent, New Reno and Westwood+ TCP increase the cwnds up to the bandwidth delay 70 product, which is roughly 40 segments, plus the bottleneck queue 60 Segments size, which is 40 segments too, before experiencing congestion; 50 this is shown in Figs. 4 and 5 where the cwnds of New Reno and 40 Westwood+ systematically reach the value of 80 segments before 30 being reduced by following the Multiplicative Decrease or 20 Adaptive Decrease mechanism, respectively. On the other hand, 10 when the reverse traffic is turned on, both New Reno and 0 Westwood+ can experience a congestion episode as soon as the 0 100 200 300 400 500 600 700 800 900 1000 cwnd is larger than the buffer size because of the burstiness of the s TCP due to ACK compression. However, it should be noted that Figure 6. cwnd and ssthresh of a single Vegas TCP flow with the ssthresh dynamics of Westwood+ is less sensitive to reverse traffic contributed by 10 New Reno TCP flows congestion along the ACK path with respect to New Reno because of the bandwidth estimate used for its setting. As a consequence, the Vegas connection on the forward path Regarding Vegas, Fig. 6 shows that the cwnd is constant and additively shrinks the cwnd thus obtaining a poor utilization of the approaches the bandwidth delay product when the reverse traffic bottleneck link. We have investigated more this point to see if the is off, thus providing an efficient link utilization. On the other kind of used reverse traffic is of importance. Fig. 7 shows that hand, when the reverse traffic is on, the cwnd keeps very low Vegas is not able to grab the bottleneck link also when the reverse values thus preventing link utilization. The reason is that the traffic is contributed by 10 Vegas connections, that is, Vegas does reverse traffic provokes queuing delays along the backward path, not provide full link utilization whatever source type of reverse which increases the RTT. traffic is considered. To complete the comparison we report, for 100 this time only, the behavior of Reno TCP. Fig. 8 shows that Reno cwnd is heavily affected by the presence of reverse traffic. Therefore, 90 ssthresh 80 since now on, we will always consider the New Reno TCP. 70 100 cwnd 60 Segments 90 ssthresh 50 80 40 70 30 60 Segments 20 50 10 40 0 30 0 100 200 300 400 500 600 700 800 900 1000 20 s 10 Figure 4. cwnd and ssthresh of a single New Reno TCP flow 0 with reverse traffic contributed by 10 New Reno TCP flows. 0 100 200 300 400 500 600 700 800 900 1000 100 s cwnd 90 Figure 7. cwnd and ssthresh of a single Vegas TCP flow with ssthresh 80 reverse traffic contributed by 10 Vegas TCP flows. 70 60 Based on the simulations above reported and on others that we do Segments not report here, we can conclude that the reverse traffic heavily 50 affects protocol behaviors. Therefore, the reverse traffic will be 40 always active in all scenarios we will consider in the sequel. 30 Moreover, since we have seen that the effect of reverse traffic 20 does not depend significantly on the TCP control algorithm that 10 generates it, we will consider always New Reno type reverse 0 traffic because it is the more efficient and today used TCP. 0 100 200 300 400 500 600 700 800 900 1000 s Figure 5. cwnd and ssthresh of a single Westwood+ TCP flow with reverse traffic contributed by 10 New Reno TCP flows. 5
  • 6. 120 1.0E+07 cwnd ssthresh 9.0E+06 100 8.0E+06 7.0E+06 Total Goodput (bps) 80 6.0E+06 Segments 60 5.0E+06 4.0E+06 40 3.0E+06 New Reno 2.0E+06 Vegas 20 1.0E+06 Westwood+ 0.0E+00 0 0 20 40 60 80 100 120 140 160 180 200 0 100 200 300 400 500 600 700 800 900 1000 s M=No. of TCP connections Figure 8. cwnd and ssthresh of a single Reno TCP flow with Figure 12. Total goodput of M TCP connections reverse traffic contributed by 10 New Reno TCP flows. We conclude this section by noting that Fig. 4 and 5 show that Fig. 13 plots the Jain fairness index [17]: Westwood+ and Reno exhibit a cyclic behavior that continuously probes for network capacity. The main consequence of this J FI = (∑iM1bi )2 = 2 behavior is that, when Westwood+ and New Reno coexist, they M b M ∑ i =1 i are friendly to each other; on the other hand, when Vegas coexists where bi is the goodput of the ith connection and M are the with New Reno or Westwood+ it gives up bandwidth to New connections sharing the bottleneck. The Jain fairness index Reno or Westwood+ since Vegas lacks of this probing behavior belongs to the interval [0,1] and increases with fairness reaching [11]. the maximum value at one. Fig. 13 shows that Westwood+ improves fairness in bandwidth 3.2 Single bottleneck scenario sharing with respect to New Reno TCP when M<60. The reason is The scenario depicted in Fig. 11, where M TCP sources with that RTTs are uniformly spread over the interval [20+230/M, different RTTs share a 10Mbps bottleneck link, is particularly 250]ms so that, for smaller M, RTTs are more distant, which suited for evaluating goodput and fairness in bandwidth increases the unfair behavior of TCP as stated by Eqs. (2) and (3). allocation. The M TCP flows are persistent and controlled by the Regarding Vegas, it exhibits the best jain fairness index, but with same algorithm in order to evaluate the intra-protocol fairness. the lowest goodput (see Fig. 12). RTTs are uniformly spread in the interval [20+230/M, 250]ms, 1.00 with M ranging from 10 to 200, to investigate the fairness with 0.98 respect to the round trip time. Simulations last 1000s during which 0.96 all the TCP sources send data. In order to obtain ACK 0.94 Fairness Index compression, 10 TCP New Reno senders inject traffic along the 0.92 ACKs path of the M connections. 0.90 0.88 TCP1 TCP1 0.86 New Reno Sink source 0.84 Vegas 0.82 Westwood+ Forward Traffic TCPM 0.80 TCPM Sink 0 20 40 60 80 100 120 140 160 180 200 source M Reverse Traffic 10 TCP Figure. 13. Jain fairness index over a 10Mbps bottleneck. 10 TCP sources Sinks To provide a “visual” look into the fairness issue, the sequence Figure 11. Single bottleneck scenario. numbers of 20 New Reno or 20 Westwood+ or 20 Vegas connections sharing a 10Mbps bottleneck are shown in Figs. 14- Fig. 12 shows the total goodput, which is defined as the sum of 16, respectively. Figs. 14 and 15 show that the New Reno final the goodputs of all the M TCP connections on the forward path. In sequence numbers are spread in the interval [26693, 64238] particular, the figure shows that when M is larger than 40 the total whereas the Westwood+ ones are in the shorter interval [28822, goodput approaches the bottleneck link capacity. On the other 53423]. Fig. 16 shows that Vegas is fair but provides very low hand, when M is smaller than 40, Vegas provides a very low total goodput due to the presence of reverse traffic (see Fig. 12). goodput. Again this phenomenon is due to the TCP traffic on the To summarize simulation results of this section, we can say that backward path, which has a significant impact on Vegas TCP (see both Westwood+ and New Reno achieve full link utilization with also Figs. 6, 7). Westwood+ providing improved intraprotocol fairness with respect to New Reno. On the other hand, Vegas is fair but it is unable to utilize the network bandwidth. 6
  • 7. 70000 persistent and starts at time t = 10s. Notice that the described scenario is a ”worst case” scenario for the source C1 since: (1) C1 Sequence Numbers (Segments) 60000 starts data transmission when the network bandwidth has been 50000 grabbed by the cross traffic sources; (2) C1 has the longest RTT and experiences drops at each router it goes through. 40000 Sink3 C3 Sink5 C5 30000 20000 C1 Sink1 10000 R R R R 0 0 200 400 600 800 1000 s Figure 14. Sequence numbers of 20 New Reno TCP C2 Sink2 C4 Sink4 connections 70000 1th hop 2th hop Sequence Numbers (Segments) 60000 Figure 17. Multi bottleneck topology. 50000 40000 We will consider the following 4 scenarios: Scenario 1. The C2,C3,C4 … C2N+1 sources of cross traffic are 30000 controlled by New Reno TCP whereas the C1 connection is 20000 controlled by New Reno, Vegas or Westwood+, respectively. This scenario aims at comparing New Reno, Vegas or Westwood+, 10000 when going through an Internet dominated by New Reno traffic. 0 In other terms, this scenario allows us to investigate the capacity 0 200 400 600 800 1000 of New Reno, Vegas and Westwood+ to grab network bandwidth s when competing with New Reno cross traffic, which is the friendliness of New Reno TCP towards Vegas or Weswtood+ Figure 15. Sequence numbers of 20 Westwood+ TCP. TCP. 70000 Fig. 18 shows the goodput of the C1 connection as a function of the number of hops; the fair share is 5Mbps. The goodput of the Sequence Numbers (Segments) 60000 C1 connection monotonically decreases with the number of hops 50000 because of increased loss ratio and RTT. Westwood+ roughly achieves the same goodput as New Reno, whereas Vegas is again 40000 not able to grab its bandwidth share in a “New Reno 30000 environment”. Fig. 19 shows that the total goodput, which is now computed as 20000 the goodput of the C1 connection + average of the C2,C4..C2N 10000 connection goodputs, does not vary significantly with the number of hops. This is due to the fact that the total goodput mainly 0 depends on the behavior of the cross traffic connections 0 200 400 600 800 1000 C2,C4..C2N whereas the C1 connection has a negligible impact on s the total goodput. Figure 16. Sequence numbers of 20 Vegas TCP connections. 1.0E+07 Goodput of the C1 connection (bps) 3.3 Multi bottleneck scenario The multi-bottleneck scenario is particularly suited to investigate 1.0E+06 the inter-protocol friendliness of New Reno, Westwood+ and Vegas TCP. The topology depicted in Fig. 17 is characterized by: 1.0E+05 (a) N hops; (b) one persistent connection C1 going through all the N hops; (c) 2N persistent sources C2;C3;C4 … C2N+1 transmitting New Reno cross traffic data over every single hop. The capacity of the 1.0E+04 Vegas entry/exit links is 100Mbps with 20ms propagation delay. The Westwood+ capacity of the links connecting the routers is 10Mbps with 10ms Fair share propagation delay. Router queue sizes have been set equal to 125 1.0E+03 packets, which corresponds to the bandwidth delay product of a 1 2 3 4 5 6 7 8 9 10 typical RTT of 150ms. Simulation lasts 1000s during which the No. of traversed hops cross traffic sources are always active. The connection C1 is Figure 18. C1 Goodput vs. number of traversed hops in the presence of New Reno cross traffic. 7
  • 8. 1.0E+07 1.0E+07 9.0E+06 9.0E+06 Total Goodput (bps) Total Goodput (bps) 8.0E+06 8.0E+06 7.0E+06 7.0E+06 6.0E+06 6.0E+06 5.0E+06 5.0E+06 New Reno New Reno 4.0E+06 Vegas 4.0E+06 Vegas Westwood+ Westwood+ 3.0E+06 3.0E+06 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 No. of traversed hops No. of traversed hops Figure 19. Total Goodput vs. number of traversed hops in the Figure 21. Total Goodput vs. number of traversed hops in the presence of New Reno cross traffic. presence of Westwood+ cross traffic. Scenario 3. The C2;C3;C4 … C2N+1 sources of cross traffic are Scenario 2. The C2;C3;C4 … C2N+1 sources of cross traffic are controlled by Vegas TCP whereas the C1 connection is controlled by Westwood+ TCP whereas the C1 connection is alternatively controlled by Reno, Vegas or Westwood+. This alternatively controlled by New Reno, Vegas or Westwood+. This scenario investigates the friendliness of Vegas towards Reno and scenario allows us to investigate the friendliness of Westwood+ Westwood+ TCP. Fig. 22 shows that Reno and Westwood+ towards New Reno and Vegas TCP. Fig. 20 shows the goodput of basically achieve the same goodput, which is larger than the fair the C1 connection as a function of the number of traversed hops. share for any number of traversed hops: the reason is that the Again, it shows that Vegas is not able to grab its bandwidth share. Vegas cross traffic, differently from New Reno and Westwood+, Moreover, a comparison of the New Reno curves in Fig. 18 and avoids to systematically fill the queues thus leaving more room Fig. 20 shows that the C1 New Reno achieves a slightly greater for the C1 traffic. The total goodput is very low when the C1 goodput when going through Westwood+ cross-traffic than when connection is controlled by Vegas, whereas it approaches the going through New Reno cross-traffic, that is, Westwood+ is more 10Mbps link capacity when the C1 connection is controlled by than friendly towards New Reno. Also in this case, the total New Reno or Westwood+ (see Fig. 23). This result again shows goodput, which is reported in Fig. 21, does not vary significantly that the Vegas cross traffic connections C2,C4..C2N are far from with N. providing an efficient utilization of the network capacity. 1.0E+07 1.0E+07 Goodput of the C1 connection (bps) Goodput of the C1 connection (bps) 1.0E+06 1.0E+06 1.0E+05 1.0E+05 New Reno New Reno 1.0E+04 Vegas 1.0E+04 Vegas Westwood+ Westwood+ Fair share Fair share 1.0E+03 1.0E+03 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 No. of traversed hops No. of traversed hops Figure 20. C1 Goodput vs. number of traversed hops in the Figure 22. C1 Goodput vs. number of traversed hops in the presence of Westwood+ cross traffic. presence of Vegas cross traffic. 8
  • 9. presence of homogeneous cross-traffic. 1.0E+07 9.0E+06 To summarize, results of this section have shown the inter- protocol friendliness of New Reno and Westwood+ towards each Total Goodput (bps) 8.0E+06 other that is mainly due to the fact that they employs the same 7.0E+06 probing mechanism. On the other hand, the rtt-based congestion detection mechanism of Vegas TCP turns out an unfriendly 6.0E+06 behavior of New Reno or Westwood+ towards Vegas, which is 5.0E+06 not able to fully utilize the network bandwidth. For these reasons, New Reno we will not consider Vegas in the sequel of the paper. 4.0E+06 Vegas Westwood+ 3.0E+06 3.4 Wireless scenarios 1 2 3 4 5 6 7 8 9 10 This section aims at investigating the behavior of TCP over No. of traversed hops wireless links that are affected by losses not due to congestion. Figure 23. Total Goodput vs. number of traversed hops in the This case is particularly interesting since it is well known that presence of Vegas cross traffic. protocols that react to losses by multiplicatively decreasing the control windows do not provide efficient utilization of lossy Scenario 4. All traffic sources are controlled by the same control channels. In this scenario we consider Westwood+, New Reno and algorithm. This is a homogeneous scenario aiming at evaluating also TCP SACK to investigate the efficiency of SACK to recover New Reno, Westwood+ and Vegas TCP in absolute terms. from sporadic random losses. Since TCP SACK by default does Fig. 24 shows that New Reno and Westwood+ provide roughly not use delayed ACK, we consider Westwood+ and New Reno the same goodput, which monotonically decreases when the with delayed ACK (default case) and without delayed ACK in number of traversed hops increases, whereas Vegas achieves the order to get a fair comparison. highest goodputs when the number of traversed hops is larger than 3 because the Vegas cross traffic is not efficient as Reno or 3.4.1 Terrestrial scenario Westwood+. In fact, Fig. 25 shows that the total goodput obtained The first scenario we consider is the hybrid wired/wireless by using Vegas TCP scenario is much smaller than the total topology shown in Fig. 26. The TCP1 connection goes through a goodputs obtained by New Reno or Westwood+, which means wired path terminating with a last hop wireless link. The wireless that the Vegas cross traffic do not use the share of link capacity last hop models a mobile user accessing the Internet using a radio left unused by the C1 connection. link as in the case of a cellular telephone system. The one way delay of the TCP1 connection is 125ms with 20ms delay on the 1.0E+07 wireless link, which is a 2Mbps link [2]. RTTs of the 5 cross Goodput of the C1 connection (bps) traffic connections and of the 10 New Reno backward traffic 1.0E+06 connections are uniformly spread in the intervals [66ms,250ms] and [46ms,250ms], respectively. We consider a wireless link affected by bursty segment losses in both directions. We use the 1.0E+05 Gilbert two state Markov chain to model the loss process [10]. In particular, we assume a segment loss probability equal to 0, when New Reno the channel is in the Good state, and equal to 0.1 when the 1.0E+04 Vegas channel is in the Bad state. The permanence time in the Good Westwood+ state is assumed deterministic and equal to 1s whereas the Fair share 1.0E+03 permanence time in the Bad state is assumed also deterministic 1 2 3 4 5 6 7 8 9 10 but this time we consider values ranging from 0.1ms to 100 ms. No. of traversed hops When the permanence time in a state elapses, the state can transit to a Good or Bad state with a probability p=0.5. For each Figure 24. C1 Goodput vs. number of traversed hops in the considered case, we run 10 simulations by varying the seed of the presence of homogeneous cross-traffic. random loss process. For each value of the BAD state duration we 1.0E+07 report the maximum, minimum and average goodputs. 9.0E+06 5 TCP 5 TCP Total Goodput (bps) 8.0E+06 Sinks sources 7.0E+06 Wireless link Cross Traffic 6.0E+06 TCP1 5.0E+06 New Reno source 4.0E+06 Vegas TCP1 Westwood+ Reverse Traffic 10 TCP sink 3.0E+06 10 TCP sources 1 2 3 4 5 6 7 8 9 10 Sinks No. of traversed hops Figure 25. Total Goodput vs. number of traversed hops in the Figure 26. Mixed wired/wireless scenario. 9
  • 10. In order to analyze only the impact of bursty losses on the TCP 100 behavior, we have first turned off both the cross and reverse 90 traffic sources. This simple scenario is particularly useful to 80 investigate the effectiveness of the adaptive decrease paradigm when losses not due to congestion are experienced by the TCP. 70 Fig. 27 (a) shows the goodput of TCP1 Westwood+ and New 60 Segments Reno, when the delayed ACK is enabled (default case), as a 50 function of the duration of the BAD state. It turns out that 40 Westwood+ improves the goodput for a large set of channel 30 conditions. In particular, when the delayed ACK option is 20 cwnd enabled, Westwood+ increases the utilization of the wireless link 10 ssthresh from 70% to 230% with respect to New Reno. Fig. 27 (b) shows 0 the goodput of Westwood+, New Reno and TCP SACK when the 0 100 200 300 400 500 600 700 800 900 1000 delayed ACK is disabled (default case for TCP SACK). In this s case SACK TCP provides a goodput similar to that of New Reno, Figure 28. Cwnd and ssthresh of Westwood+ when the whereas Westwood+ improves the link utilization with respect to duration of the BAD state is 0.01s. SACK and New Reno from 34% up to 177%. The reason is that the adaptive setting of cwnd and ssthresh performed by 100 Westwood+ takes into account the bandwidth used at time of 90 cwnd congestion, so that the TCP sender does not lose ground when in ssthresh 80 the presence of losses not due to congestion. To get a further 70 insight into this feature, Figs. 28 and 29 report the cwnd dynamics 60 Segments of Westwood+ and New Reno, respectively, obtained when the 50 duration of the BAD state is 0.01s and the delayed ACK is 40 enabled. In this case, Westwood+ TCP provides a ssthresh 30 approaching the bandwidth-delay product of 40 packets, whereas 20 New Reno TCP provides a ssthresh that is smaller than one fourth the bandwidth-delay product. 10 0 2.0E+06 0 100 200 300 400 500 600 700 800 900 1000 1.8E+06 s 1.6E+06 Figure 29. Cwnd and ssthresh of New Reno when the duration 1.4E+06 of the BAD state is 0.01s. Goodput (bps) 1.2E+06 1.0E+06 One further point valuable of investigation is when Westwood+ 8.0E+05 shares the wired portion of the network with several TCP flows on 6.0E+05 the forward and backward paths. For that purpose, we turn on the 4.0E+05 Westwood+, DACK enabled cross and reverse traffic in Fig. 26 and we measure the goodput of the TCP1 connections for various values of the BAD state 2.0E+05 New Reno, DACK enabled duration. Fig. 30 shows that the delayed ACK option plays a 0.0E+00 major role in this scenario. In fact, protocols that do not employ 0.0001 0.001 0.01 0.1 the delayed ACK option provides goodputs that are roughly two Duration of the BAD state (s) times larger than those obtained when the delayed ACK option is (a) enabled. The reason is that the delayed ACK option slows down 2.0E+06 the TCP probing phase. In these scenarios Westwood+ TCP 1.8E+06 (DACK disabled) still improves the goodput with respect to New 1.6E+06 Reno (DACK disabled) and SACK TCP, but the improvement is 1.4E+06 now only up to roughly 20%. The reason is that in this case the Goodput (bps) 1.2E+06 TCP1 connection loses bandwidth in favor of the cross traffic that, 1.0E+06 being wired, is not penalized by losses not due to congestion. 8.0E+05 6.0E+05 4.0E+05 Westwood+, DACK disabled New Reno, DACK disabled 2.0E+05 SACK 0.0E+00 0.0001 0.001 0.01 0.1 Duration of the BAD state (s) (b) Figure 27. Goodput of the TCP1 connection without reverse traffic: (a) DACK enabled; (b) DACK disabled. 10
  • 11. 900000 Westwood+, DACK disabled 10000000 New Reno, DACK disabled 9000000 800000 SACK 8000000 700000 Westwood+, DACK enabled Total Goodput (bps) New Reno, DACK enabled 7000000 Goodput (bps) 600000 6000000 500000 5000000 4000000 Westwood+, DACK disabled 400000 3000000 Westwood+, DACK enabled 300000 2000000 SACK New Reno, DACK disabled 200000 1000000 New Reno, DACK enabled 100000 0 0.0001 0.001 0.01 0.1 0.0001 0.001 0.01 0.1 1 Duration of the BAD state (s) Duration of the BAD state (s) Figure 30. Goodput of the TCP1 connection in presence of Figure 32. Goodput over the GEO satellite scenario. cross and reverse traffic. 3.4.2 Satellite scenario 4. LIVE MEASUREMENTS This section investigates the performances of New Reno and When a new protocol is proposed, it is necessary to collect a large Westwood+ over a large leaky pipe such as in the case of a set of simulation and experimental results in order to assess its satellite scenario. For that purpose we consider the scenario in validity and the advantages of its deployment in the real Internet Fig. 31 where a 10mbps bottleneck link has a one-way delay equal [29]. In this section we test Linux implementations of New Reno to 275ms, which corresponds to a GEO satellite connection [32]. and Westwood+ [44] over the real Internet. More than 4000 files, with different sizes, have been uploaded via ftp from a host at the Laboratory of Communication and Control at Politecnico di Bari (South of Italy) to three remote servers, which are located at Parma (North of Italy), Uppsala University (Sweden) and 20 TCP 20 TCP University of California, Los Angeles (Ucla). For each upload we senders sinks have measured the goodput, the number of retransmitted segments, and important variables such as cwnd, sshtresh, RTT and bandwidth estimate. Each measurements session collects data of many file uploads, which are alternatively executed by using 10 TCP 10 TCP Westwood+ or New Reno. Table 2 summarizes the main sinks senders characteristics of the sessions from Bari to the FTP server at UCLA, whereas Fig. 33 shows the average goodputs achieved during data transfer. Figure 31. GEO satellite scenario. Table 2: FTP uploads from Politecnico of Bari, Italy to UCLA, We consider 20 TCP forward connections in the presence of Los Angeles. reverse traffic contributed by 10 long-lived New Reno connections. RTTs of the forward connections are equal to 590ms. date No. of Uploads Size of uploaded Simulations last 1000s. We assume the same error model used in files (MBytes) the previous sub-section except for the BAD state duration that Feb,21 2003 117 32 has been increased up to 1s. The bottleneck link experiences Feb,26 2003 197 3.2 segment losses in both directions. Fig. 32 shows the goodput Feb,28 2003 702 3.2 provided by the two considered TCP control algorithms. When Mar,14 2003 54 32 delayed ACK is enabled, Westwood+ TCP provides a goodput Mar,19 2003 79 32 improvement with respect to New Reno TCP that ranges from Mar,21 2003 47 32 20% to 160%, whereas, when the delayed ACK is disabled, the improvements of Westwood+ with respect to SACK TCP and New Reno are up to 80%. Again, the reason is that Westwood+ adaptively reduces the cwnd and sstresh by taking into account an The average goodput of a measurements session is obtained by estimate of the available bandwidth: this mitigates the impact of averaging the goodputs of the session uploads. It turns out that random losses not due to congestion that provokes multiplicative Westwood+ TCP provides goodput improvements ranging from reductions of New Reno control windows. 23% to 53% with respect to New Reno. It is also worth noting that improvements provided by Westwood+ TCP are not due to a more aggressive behaviour, since Fig. 34 shows that Westwood+ and New Reno TCP have similar retransmission ratios. Table 3 summarizes the characteristics of the measurement sessions from Bari to the server at the Uppsala University. Figs. 35 and 36 show the average goodputs and the average 11
  • 12. retransmission ratio in this case. Westwood+ TCP provides 12 goodput improvements ranging from 4% to 40% with respect to New Reno Westwood+ New Reno with similar retransmission ratios. 10 Retransmission ratio (%) Table 3: FTP from Politecnico of Bari, Italy to Uppsala 8 University (Sweden) date No. of Uploads Size of uploaded 6 files (MBytes) Dec,14 2002 253 32 4 Dec,17 2002 200 3.2 Jan,10 2003 100 3.2 2 Jan,12 2003 100 3.2 Jan,13 2003 150 32 0 21 Feb 2003 26 Feb 2003 28 Feb 2003 14 Mar 2003 19 Mar 2003 21 Mar 2003 Jan,17 2003 1000 3.2 32MB 3.2MB 3.2MB 32MB 32MB 32MB Feb,3 2003 278 32 Feb,7 2003 500 32 Figure 34. Retransmission ratio during ftp to UCLA. 250 Table 4 summarizes the characteristics of the measurement New Reno Average Goodput (KBytes/s) sessions from Bari to the FTP server located at Parma. Figs. 37 Westwood+ and 38 show the goodputs and the retransmission ratios that have 200 been measured during these data transfer. In this case the connection has a national extension and Westwood+ and New 150 Reno TCP provide similar goodputs. 100 Table 4: FTP uploads towards the server located at Parma (Italy) 50 date No. of Uploads Size of uploaded files (MBytes) 0 Mar,26 2003 100 32 14 Dec 17 Dec 10 Jan 12 Jan 13 Jan 17 Jan 3 Feb 7 Feb Mar,27 2003 98 3.2 2002 2002 2003 2003 2003 2003 2003 2003 Mar,28 2003 1000 3.2 32 MB 3.2MB 3.2MB 3.2MB 32MB 3.2MB 32MB 32MB Apr,4 2003 390 32 Apr,7 2003 200 3.2 Figure 35. Average goodput during ftp to Uppsala. Apr,9 2003 200 3.2 4.5 New Reno Apr,11 2003 1000 3.2 Retransmission ratio (%) 4 Westwood+ 3.5 3 50 New Reno 2.5 Westwood+ 45 2 40 Average Goodput (KBytes/s) 1.5 35 1 30 0.5 25 0 20 17 Dec 10 Jan 12 Jan 13 Jan 17 Jan 3 Feb 1003 7 Feb 2003 2002 2003 2003 2003 2003 32MB 32MB 15 3.2MB 3.2MB 3.2 MB 32MB 3.2MB 10 5 Figure 36. Retransmission ratio during ftp to Uppsala. 0 21 Feb 2003 26 Feb 2003 28 Feb 2003 14 Mar 2003 19 Mar 2003 21 Mar 2003 32MB 3.2MB 3.2 MB 32MB 32 MB 32 MB Figure 33. Average goodput during ftp to Los Angeles. 12