• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Modeling TCP Incast Issue in Data Center Networks and an Adaptive Application-Layer Solution

    2018-04-08 03:11:37JinTangLuoJieXuandJianSun

    Jin-Tang Luo, Jie Xu, and Jian Sun

    1.Introduction

    In data centers, an application often simultaneously requests data from numerous servers. This results in a many-to-one traffic pattern where multiple servers concurrently send data fragments to a single client via a bottleneck switch. For instance, in web search, a search query is partitioned and assigned to many workers, and then the workers’ responses are aggregated to generate the final result. As the number of concurrent senders increases, the bottleneck switch buffer which is usually shallow[1]can be easily overflowed, leading to massive packet losses and subsequent transmission control protocol (TCP) timeouts. As the minimum retransmission timeout (minRTO) is much greater than the round-trip delay in data centers, even one timeout remarkably prolongs the overall data transmission time. Hence, the client’s goodput decreases to be lower than the link capacity by one or even two orders of magnitude. Such catastrophic TCP goodput collapse for applications with many-to-one traffic pattern is referred to as TCP incast[1].

    Many solutions to the TCP incast issue have been proposed from the aspects of different layers. For instance, at the Ethernet layer, the fully-qualified class name (FQCN)[2]uses explicit network feedback to control congestion among switches;at Internet protocol (IP) layer, [3] explores the effectiveness of tuning explicit congestion notification (ECN) to mitigate incast,and the cyclic prefix (CP)[4]drops only the packet payload instead of the entire packet upon congestion to reduce timeout possibility; at the transport layer, data center TCP (DCTCP)[5],predicted rice interactome network (PRIN)[6], and pFabric[7]decrease the timeout possibility through the cooperation of end-hosts and network; by adding a “shim layer” to the data receiver, incast congestion control for TCP (ICTCP)[8], proxy auto config (PAC)[9]and deadline and incast aware TCP(DIATCP)[10]proactively adjust the in-flight data size to reduce the packet loss; and at the application layer, [1] and [11]-[13]restrict the number of concurrent connections to avert incast.Among these proposals, the application-layer solutions are most practical for their low deployment barrier and minimal impact on ordinary one-to-one applications. Indeed, Ethernetlayer solutions require hardware revisions not supported by current commodity switches, and transport-layer and shimlayer solutions may cause fairness issues to ordinary applications running on regular TCP. In contrary, application-layer solutions merely regulate the data transfer of applications with the manyto-one traffic pattern, so they are easy to deploy and pose no side effect on ordinary applications.

    Despite of the numerous application-layer solutions in literature, few works analytically study how application’s regulation on data transfer can affect incast. Currently, there are several analytical models for the TCP incast problem[1],[6],[11],[14]. However, most of them either ignore the possible existence of background TCP traffic[1],[11], or oversimplify the application layer as a dumb data source/sink without the ability to control data transfers[6],[14]. Therefore, the existing models provide few useful insights into addressing incast from the application layer, which explains why the current application-layer solutions can only avert incast in known and predefined network environments, e.g., the bottleneck link is the last-hop link to the receiver[12],[13], or the available bottleneck bandwidth is known by the receiver[1],[11]).But in real data centers, many network parameters may change drastically and cannot be known in prior. For instance, the bottleneck link often varies due to load balancing, and available bandwidth may fluctuate drastically in the presence of background TCP traffic. In these varying environments, the current solutions often fail to effectively prevent TCP incast, as demonstrated in Section 4.

    In this paper, we intend to understand and solve the TCP incast problem from the viewpoint of applications. Toward this goal, we first develop an analytical model to comprehensively investigate the impact of the application layer on incast.Then under the model’s enlightenment, we propose an adaptive application-layer control mechanism to eliminate incast.

    Since incast is essentially caused by TCP timeout, here we focus on modeling the relationship between the appearance of timeout and various factors related to applications, including the network environment and connection variables. Compared with the existing models[1],[6],[11],[14], our model is based on more general assumptions about TCP behaviors, and it considers the impact of background TCP connections on incurring timeout.In addition, integrated with the optimization theory, the model provides some insightful guidelines for dynamically tuning connection variables to minimize the incast probability in a wide range of network environments.

    According to the theoretical results, the crux of avoiding timeout is to adaptively adjust the number of concurrent connections and equally allocate the sending rate to connections. Following this idea, we design an applicationlayer incast probability minimization mechanism (AIPM),which only modifies receiver-side applications to avert timeout. With AIPM, a client (receiver) concurrently sets up a small number of connections, and assigns an equal TCP advertised window (awnd) to each connection. After a connection finishes data transferring, a new connection can be started. The regulation of awnd is based on network settings and the amount of concurrent connections. The concurrency amount is decided by a sliding connectionwindow mechanism similar to TCP’s congestion control. By this mechanism, the connection-window size grows gradually upon a successful data transmission and shrinks as a connection is “l(fā)ost”. A connection is considered as the lost connection when three new connections have finished since the last time it transmits new data. The lost connection is terminated at once and will be re-established when allowed by the connection-window.

    The simulations show that AIPM effectively eludes incast and consistently achieves high performance in various scenarios. Particularly, in a leaf-spine-topology network with dynamic background TCP traffic, AIPM’s goodput is above 68% of the bottleneck capacity, while the proposals in [1]and [11] both suffer from rather low goodput (< 30%) due to incast.

    The major contributions of this paper are two-folds:

    ● Build an analytical model to disclose the influence of network environment and connection variables on the occurrence of incast. The model provides some insightful guidelines for tuning connection variables to minimize incast possibility.

    ● Design an adaptive application-layer solution to the TCP incast problem. To our best knowledge, this solution is the first application-layer solution that can efficiently control incast in network environments with dynamic background TCP traffic and multiple bottleneck links.

    The rest of the paper is organized as follows. Section 2 describes our model for incast probability and provides some insightful guidelines for taming incast. Section 3 proposes an adaptive application-layer solution to incast, namely, AIPM. In Section 4 we evaluate AIPM in various scenarios using NS2 simulations. Section 5 concludes the paper.

    2.Modeling and Minimizing TCP Incast Probability

    As a data requestor, the receiver-side application (i.e., the client) can implicitly manage data transmission by adjusting connection variables (including the number of concurrent connections and the sending window size for each connection).This fact enlightens us to model the TCP incast probability as a function of connection variables. Based on this model, we derive the conditions on which the incast risk is minimal.

    2.1 Notations and Assumptions

    First of all, we put forward an important concept related to data transfers of concurrently living connections, called round.The first round starts from the beginning of data transmission from an endpoint, and lasts one round-trip time (RTT). The next round starts from the end of last round and lasts one RTT.

    Then consider the TCP incast scenario in Fig. 1. Let R be the number of rounds for transmitting data. During the kth round, there are n(k) servers sending data fragments, formally termed as a sever request unit (SRU) to a client via a common bottleneck link. The ith server’s sending rate is xi(k), and its sending window is wi(k), for 1≤i≤n. The bottleneck link is also shared by m(k) background connections from other applications,whose sending rates are yj(k) for. For the bottleneck link, its capacity is C packets per second, its buffer size is B packets, and the propagation RTT is D. The rest notations are summarized in Table 1.

    We also make some assumptions to facilitate modeling.Firstly, we assume that the queuing policy of the bottleneck link is drop-tail. Secondly, we assume the spare bottleneck buffer is ignorable compared with the sum of the sending windows. This assumption is reasonable due to the “buffer pressure” phenomenon and the fact that switches’ buffers are usually very shallow in data centers[1]. Thirdly, we assume that a connection sees timeout only if the entire window of packets are lost. Most researches[5],[14]have shown that the full-window loss is the dominating kind of timeout that causes TCP incast, thus other kinds of timeout are trivial while modeling incast. At last, we assume that minRTO =200 ms by default, which is significant for the overall transmission time of the requested data block, so even one timeout leads to TCP incast with drastic goodput degradation.

    Fig. 1. General scenario of TCP incast, where multiple servers concurrently transmit data fragraments (SRU) to a single client.

    Table 1: Meanings of the commonly used parameters

    2.2 Probability of TCP Incast

    Now we begin to model the probability of TCP incast as a function of current network condition (i.e., B, C, D, and yi(k))and connection context variables (i.e., n(k) and xi(k)).

    First, recall that the client will suffer from incast if it has at least one timeout during R rounds, so the incast probability can be expressed by

    where Pi(k) is the timeout probability of the ith connection in the kth round.

    The timeout possibility Pi(k) of a connection is determined jointly by the sending window wi(k) and the packet loss rate pi(k) of that connection. Specifically, we consider full window loss as the only cause to timeout, thus Pi(k) equals to

    The next task is to derive the packet drop rate pi(k). As we assumed previously, the spare bottleneck buffer is ignorable compared with the sum of the servers’ sending windows,which means that the bottleneck link would drop packets once the servers start transmitting data. We denoteas the sum of the servers’ sending rates,andas the sum of the background connections’ sending rates. During the kth round, packets are injected into the network at the total rate of, and are processed by the bottleneck link at the rate of C.Henceforth, the drop rate pi(k) is

    Moreover, a connection’s sending window wi(k) is related with its sending rate xi(k) and RTT D as

    Eventually, by substituting (2) to (4) into (1), we derive the TCP incast probability as an analytical function of network condition (i.e., B, C, D, and Y(k)) and connection variables (i.e.,n(k) and xi(k)) as follows:

    2.3 Minimizing Incast Probability

    As (5) indicates, to minimize the incast probability Pincast,we must minimize the timeout probability in every round. This is to maximize the timeout-free possibility that no connection ever incurs timeout in any round k as follows

    Here we explore how could the client maximize the timeout-free probability (6) through adjusting the sending ratesand the connection amount n(k).To reveal the individual impact of a parameter on timeout, we analyze it by keeping other parameters unchanged. Because our analysis below focuses on maximizing (6) in every single round, we simply omit the round number k in the notations, like n(k) to n.

    1) Adjusting sending sizes, x: We fix other parameters. Then the maximization of the timeout-free probability in (6) is to solve the optimization problem below:

    It is straightforward to check that the Hessian matrix of?ln[f(x)] is positive semi-definite over the region x≥0, which means that ln[f(x)] is a concave function. This gives our analysis a great convenience. Specifically, the method of the Lagrange multiplier states that ln[f(x)] is globally maximized by the sending rate allocationif and only if

    The unique solution of (8) that maximizes ln[f(x)] is

    which also minimizes the incast probability (6).

    Remark 1: To minimize the incast risk, the connections should always be given a same sending rate. This operation is feasible at the client application, as the client knows the concurrent connections’ number n, and it is able to implicitly control each connection’s sending rate xithrough modifying the TCP advertised window field in acknowledgement (ACK)packets.

    Remark 2: While the sending rates’ sum X is assumed as a constant for deriving (9), actually the client can change X by tuning the sending rates. But as proved in Appendix,the optimal X that can minimize the incast probability is dependent on the background traffic Y. Since the client does not know Y, it is unable to properly set X to prevent incast.

    2) Adjusting the concurrent connections’ number, n: We fix other parameters. According to (6) and (9),we optimally set the sending rates to be xi=X/n for 1≤i≤n and rewrite the timeout-free probability as

    which is always negative since p lies in (0, 1). This suggests that the timeout-free probability (10) decreases with the concurrent connection number n. Henceforth, the incast probability is an increasing function of n and will be minimized at n=1.

    Remark: The client should lower the number of concurrent connections n to reduce the incast risk. On the other hand, an excessively small n may cause a great waste of bandwidth in the cases where each connection has so little data to send SRU that it finishes sending before fully utilizing the bandwidth.How to properly set n will be discussed in the next section.

    3.Minimizing Incast Probability at Application Layer

    Based on the analyses of (9) and (11), we propose an AIPM scheme. The AIPM is implemented at the client application, and it minimizes the incast probability via equally allocating advertised windows of concurrent connections and dynamically adjusting concurrent connections’ amount.

    3.1 Allocate Equal Advertised Window to Connections

    As (9) indicates, the risk of TCP incast is minimal if the existing connections have an equal sending rate. However,AIPM is essentially a part of the client application, which means it cannot directly control each connection’s sending rate. Therefore, AIPM emulates the equal sending rate allocation by setting awnd at the client (e.g., via calling the setsokopt() application programming interface (API) in Berkeley software distribution (BSD) systems) as follows.

    First, according to (4) and (9), the ideal sending rate allocation is equivalent to the following sending window allocation:

    where X is the total sending rate of AIPM’s n connections,and D is RTT without queuing.

    Next, AIPM should let X equal the bottleneck capacity C,so that it can fully utilize the bottleneck link without selfinduced drops. From X=C and (12), we derive that

    Finally, since TCP’s sending window size is upperbounded by the advertised window size (awnd), AIPM can emulate the equal sending window assignment in (13) by assigning an identical awnd to its connections as

    where awndiis awnd of the connection i. Although such emulation is suboptimal, it avoids the polarized allocation of sending window sizes and thus decreases the timeout probability.

    Naturally, even if AIPM strictly follows (14) and equally allocates awnd to its connections, timeout may still happen due to the background traffic. Therefore, AIPM must further decrease timeout risk by adaptively tuning the connections’ amount. Besides, it must quickly recover data transmission from the timeout connections. These two demands can be met by the following two mechanisms,respectively.

    3.2 Determine the Proper Number of Concurrent Connections with Sliding Window Mechanism

    To reduce the timeout probability while keeping high goodput, AIPM must selectively connect to a subset of the servers at one time. The question is which servers AIPM should connect to. To find the answer, AIPM employs a sliding-window-like mechanism to maintain a window of the concurrently existing connections. In concept, we term this window as the connection window, or shortly, con_wnd. When the existing connections are less than the con_wnd size, AIPM establishes a new connection and admits it to con_wnd. When a connection finishes, AIPM removes it from con_wnd.

    AIPM uses an additive increase multiplicative decrease(AIMD) policy to decide the con_wnd size n. The initial value for n is n=1. Whenever a connection in con_wnd completes,AIPM infers that the bottleneck link is not jamming, thus it gradually increases n as

    The growth of the concurrent connections inevitably leads to timeout. Based on the fact that minRTO=200 ms, is much greater than a connection’s ordinary life period(mostly less than 1 ms). The AIPM deduces a connection to be timeout if three new connections have been admitted and finished since the last time the connection transmitted any data.

    After detecting timeout, AIPM realizes that the timeout probability right now is too high. According to (11), AIPM can reduce the timeout probability by lowering the number of concurrent connections n while fixing other parameters. Hence,AIPM halves con_wnd as follows:

    Because AIPM does not reduce the total awnd when performing (16), the live connections that are not timeout can quickly occupy the spare bandwidth and maintain relatively high link utilization.

    3.3 Fast Reconnection and Slow Withdrawal

    1) Fast reconnection: For a timeout-broken data connection, AIPM terminates that connection (by sending finish (FIN) to the connection’s server) and removes it from con_wnd. Then AIPM will reconnect to the data server as long as con_wnd allows. Since the SRU is so small(typically dozens of KB), the server can retransmit it within an ignorable period of time. This scheme is termed as fast reconnection. It enables AIPM to quickly recover data transmission from the timeout-broken servers rather than being slowed down by TCP’s sluggish timeout retransmission mechanism.

    2) Slow withdrawal: As timeout occurs to some connections,if AIPM naively follows (16) and instantly halves the number of the concurrent connections, it may have to close some live connections that are not timeout. Nevertheless, these live connections have already cut their sending windows after seeing packet losses and are unlikely to cause more timeout situation. Therefore, closing these live connections at one time will result in unnecessary data retransmissions and may even lead to link under-utilization.

    To solve the above two issues, AIPM adopts the slow withdrawal scheme in the presence of timeout. Specifically,upon detecting timeout, AIPM records the current con_wnd size n, and then slowly decreases con_wnd by one after each live connection finishes. This means that AIPM does not close any live connections that are still transmitting data. Moreover,this gives the live connections sufficient time to grow their congestion windows and to fully utilize the spare bandwidth left by the closed or timeout connections (since the delays in datacenters are so small, the live connections can grow their congestion windows very fast). Once con_wnd shrinks to n/2,AIPM ends slow withdrawal and resumes to the normal AI operation (15).

    3.4 Some Discussions about Design Issues

    1) How can AIPM know the bottleneck link’s capacity,C? Nowadays datacenters employ several mechanisms like offering the uniform high capacity between racks and load balancing technologies so that congestion only happens at the edge top-of-rack (TOR) switches[8]. This feature allows AIPM to conveniently set C to be the link capacity of TOR switch.

    2) How can AIPM know the round-trip propagation delay D? Data center’s network settings (i.e., hardware, framework,topology, etc.) are usually stable over a relatively long period.Thereby, the network administrator can measure D offline and feed the value to AIPM every time network settings change.

    3) How should AIPM react if it observes several timeout connections at one time? To avoid the link under- utilization,after detecting timeout-broken connections, AIPM starts slow withdrawal normally and halves the window only once. AIPM will not further halve con_wnd even if it detects more timeoutbroken connections during the slow withdrawal state.

    4.Empirical Validations of AIPM

    4.1 Simulation Settings

    We evaluate the performance of AIPM with NS2 simulation in two different network scenarios. The first one is a single 1 Gbps bottleneck link with a 64 KB buffer and 100 μs RTT. This scenario represents static network environments.The second scenario has the leaf-spine topology as in Fig. 2,which is commonly used in data centers[7]. The network has 144 end-hosts through 9 leaf (i.e., top-of-rack) switches that are connected to 4 spine switches in a full mesh. Each leaf switch has 16 downlinks of 1 Gbps to the hosts and four uplinks of 4 Gbps to the spine. RTT between two hosts connected to different leaf switches is 100 μs. Background TCP flows arrive following the Poisson process, where the source and the destination for each flow are chosen randomly, and each flow’s size satisfies the distribution observed in real-world data mining workloads[7]as shown in Fig. 3. The background flow arrival rate is set to obtain a data load of 0.5. This scenario represents realistic datacenter network environments that are complex and dynamic. Throughout our simulations, the client requests for a data block that is scattered over N servers, and it requests for the next data block after all N servers finish sending the current one. The data packet size is 1000 Byte, and the acknowledgement (ACK) size is 40 Byte.

    Fig. 2. Leaf-spine network topology used in the second simulation scenario.

    Fig. 3. Flow size distribution in the second simulation scenario is based on real-world measurements of data mining workloads[7].

    We compare AIPM with the Na?ve method (i.e., the client concurrently connects to all N servers), as well as two state-of-art application-layer solutions, namely, Oracle security developer tools (OSDT)[1]and OSM[11]. Similar to AIPM, these two solutions also try to mitigate incast by restricting the concurrent TCP connections’ number and sending rates. The major difference is that AIPM can correspondingly tune its own parameters for various network scenarios, whereas OSDT and OSM are only specified for predefined environments with single bottleneck link and no background TCP traffic.Due to this difference, AIPM is able to remarkably outperform OSDT and OSM in changing network environments, as we will see below.

    4.2 Fixed SRU Size

    In this subsection, we fix the SRU size of each server to 256 kB, and investigate the goodput of AIPM in the two aforementioned network scenarios, respectively.

    Fig. 4 shows that our solution AIPM achieves best goodput (0.68 Gbps to 0.88 Gbps) in both cases. The reason is that AIPM dynamically adjusts each connection’s advertised window size (awnd) and the connection window size (con_wnd) based on its estimation for the network state(i.e., if some connections incur timeout). Such adjustment enables AIPM to adaptively minimize incast risk and rapidly recover from timeout even in the network environment with multiple bottleneck links and varying background traffic(Fig. 4 (b)).

    Reversely, OSDT and OSM both restrict the concurrent connection number by predefined values. Although such fixed values are well suited for static conditions, they can hardly elude incast in dynamic network environments where both bottleneck links and available bandwidth can change drastically. This is why OSDT and OSM only have the goodput less than 0.3 Gbps in Fig. 4 (b).

    Fig. 4. Goodput with SRU=256 kB for (a) a single bottleneck link without background traffic and (b) leaf-spine topology with background TCP load of 0.5.

    4.3 Varying SRU Size

    Next, we fix the overall data block size to 2 MB, and set the SRU size to 2 MB/N, where N is the total number of the servers. We evaluate AIPM’s performance in terms of goodput and request completion time.

    Fig. 5. Goodput with block size=2 MB for (a) a single bottleneck without background traffic and (b) leaf-spine topology with background TCP load of 0.5.

    As Fig. 5 illustrates, AIPM achieves higher goodput than the alternative solutions in both network scenarios, which demonstrates again that AIPM can effectively address incast issue even in highly dynamic environments. Observe that AIPM, OSDT, and OSM all have descendant goodput as N grows, for a larger N reduces each server’s SRU and then decreases the average sending window size of the concurrent connections. However, AIPM’s goodput decreases more slowly than the other two’s due to its adaptive adjustment of the number of concurrent connections. Indeed, AIPM adapts to small SRU values by allowing more connections to concurrently send data (i.e., larger con_wnd), so that it can fully utilize the bottleneck link and keep high goodput regardless of the SRU value.

    Fig. 6 compares AIPM’s request completion time (RCT)with the other three schemes. As we can see, AIPM’s RCT keeps being the smallest in both scenarios. Particularly, in the leaf-spine scenario, AIPM’s RCT is less than 20% of OSM or OSDT’s RCT, and is less than 3% of Na?ve’s RCT.This result clearly demonstrates that AIPM effectively avoids incast by triggering no TCP retransmission timeout.With such small RCT, AIPM makes the client application respond more promptly to the upper-level user, and hence improves user experience.

    Fig. 6. Request completion time (RCT) with block size=2 MB for(a) a single bottleneck without background traffic and (b) leafspine topology with background TCP load of 0.5.

    4.4 Higher Bottleneck Capacity

    At last, we explore the scalability of AIPM in higher- speed data centers. In the single-bottleneck scenario, we increase the bottleneck capacity from 1 Gbps to 10 Gbps. In the leaf-spine scenario, we increase the edge link capacity from 1 Gbps to 10 Gbps and the core link capacity from 4 Gbps to 40 Gbps. Other settings remain unchanged, i.e., SRU=256 kB, buffer size=64 kB, and RTT=100 μs.

    As we can see in Fig. 7, the goodput of AIPM is generally higher than the goodput of other methods. In particular, AIPM maintains goodput up to 91% of the bottleneck capacity in the network with no background TCP (Fig. 7 (a)), and it achieves nearly two times higher goodput than the alternative solutions while coexisting with the background TCP traffic (Fig. 7 (b)).Such good performance shows that AIPM is readily scalable for higher-speed data centers in future.

    Fig. 7. Goodput with the bottleneck capacity C=10 Gbps for (a) a single bottleneck link without background traffic and (b) leaf-spine topology with background TCP load of 0.5.

    5.Conclusions

    We built an analytical model to reveal how TCP incast is affected by various factors related to applications. From this model, we derive two guidelines for minimizing the incast risk,including equally allocating the sending rate to connections and restricting the number of concurrent connections.

    Based on the analytical results, we designed an adaptive application-layer solution to incast, which allocates an equal advertised window to connections, and uses a slidingconnection-window mechanism to manage concurrent connections. Simulation results indicate that our solution effectively eliminates incast and achieves high goodput in various network scenarios.

    Appendix

    We adjust the sending rate sum X to maximize the timeoutfree probability in (6) while fixing other parametersAccording to (9), we let the sending rates xibe xi=X/n for 1≤i≤n, and express the timeout-free probability(6) as

    If the background traffic Y is much smaller than the sum of the connections’ sending rates X, the timeout-free probabilityreduces to, which is a decreasing function of X. Reversely, if Y is much greater than X, thenbecomes, which is an increasing function of X. As a result, the optimal X that maximizesis dependent on the background traffic Y.

    [1]S. Zhang, Y. Zhang, Y. Qin, Y. Han, Z. Zhao, and S. Ci,“OSDT: A scalable application-level scheduling scheme for TCP incast problem,” inProc. of IEEE Intl. Conf. on Communications, 2015, pp. 325-331.

    [2]Y. Zhang and N. Ansari, “On mitigating TCP incast in data center networks,” inProc. of IEEE Conf. on Computer Communications, 2011, pp. 51-55.

    [3]H. Wu, J. Ju, G. Lu, C. Guo, Y. Xiong, and Y. Zhang,“Tuning ECN for data center networks,” inProc. of the 8th ACM Intl. Conf. on Emerging Networking Experiments and Technologies, 2012, pp. 25-36.

    [4]P. Cheng, F. Ren, R. Shu, and C. Lin, “Catch the whole lot in an action: rapid precise packet loss notification in data centers,” inProc. of the 11th USENIX Symposium on Networked Systems Design and Implementation, 2014, pp.17-28.

    [5]M. Alizadeh, A. Greenberg, D. A. Maltz,et al., “Data center TCP,”ACM SIGCOMM Computer Communication Review,vol. 40, no. 4, pp. 63-74, Oct. 2011.

    [6]J. Zhang, F. Ren, L. Tang, and C. Lin, “Modeling and solving TCP incast problem in data center networks,”IEEE Trans. on Parallel and Distributed Systems, vol. 26, no. 2,pp. 478-491, Feb. 2015.

    [7]M. Alizadeh, S. Yang, M. Sharif,et al., “pFabric: Minimal near-optimal datacenter transport,”ACMSIGCOMM Computer Communication Review, vol. 43, no. 4, pp. 435-446, 2013.

    [8]H. Wu, Z. Feng, C. Guo, and Y. Zhang, “ICTCP: Incast congestion control for TCP in data-center networks,”IEEE/ACM Trans. on Networking, vol. 21, no. 2, pp. 345-358, 2013.

    [9]W. Bai, K. Chen, H. Wu, W. Lan, and Y. Zhao, “PAC:taming TCP incast congestion using proactive ACK control,”inProc. of IEEE the 22nd Intl. Conf. on Network Protocols,2014, pp. 385-396.

    [10]J. Hwang, J. Yoo, and N. Choi, “Deadline and incast aware tcp for cloud data center networks,”Computer Networks, vol.68, pp. 20-34, Feb. 2014.

    [11]K. Kajita, S. Osada, Y. Fukushima, and T. Yokohira,“Improvement of a TCP incast avoidance method for data center networks,” inProc. of IEEE Intl. Conf. on ICT Convergence, 2013, pp. 459-464.

    [12]H. Zheng and C. Qiao, “An effective approach to preventing TCP incast throughput collapse for data center networks,” inProc. of IEEE Global Telecommunications Conf., 2011, pp.1-6.

    [13]Y. Yang, H. Abe, K. Baba, and S. Shimojo, “A scalable approach to avoid incast problem from application layer,” inProc. of IEEE the 37th Annual Computer Software and Applications Conf. Workshops, 2013, pp. 713-718.

    [14]W. Chen, F. Ren, J. Xie, C. Lin, K. Yin, and F. Baker,“Comprehensive understanding of TCP incast problem,” inProc. of IEEE Conf. on Computer Communications, 2015,pp. 1688-1696.

    天堂av国产一区二区熟女人妻| 国产视频内射| 欧美成人a在线观看| 亚洲av免费在线观看| 天堂动漫精品| 亚洲一区二区三区色噜噜| a级一级毛片免费在线观看| 级片在线观看| 中文字幕免费在线视频6| 一本久久中文字幕| 三级毛片av免费| 一个人免费在线观看电影| a级毛片a级免费在线| 国产精品久久久久久精品电影| 亚洲国产欧洲综合997久久,| 国产一区二区在线观看日韩| 2021天堂中文幕一二区在线观| 日本熟妇午夜| 小说图片视频综合网站| 成人精品一区二区免费| 9191精品国产免费久久| 热99re8久久精品国产| 久久性视频一级片| 在线看三级毛片| 欧美国产日韩亚洲一区| 老司机午夜十八禁免费视频| 国产麻豆成人av免费视频| 极品教师在线视频| 日本在线视频免费播放| 国产欧美日韩一区二区三| 91久久精品电影网| 少妇被粗大猛烈的视频| 一区福利在线观看| 久久精品国产自在天天线| 婷婷六月久久综合丁香| 91字幕亚洲| 成人精品一区二区免费| 可以在线观看毛片的网站| 搡老熟女国产l中国老女人| 搡老熟女国产l中国老女人| 亚洲va日本ⅴa欧美va伊人久久| 熟妇人妻久久中文字幕3abv| 日韩精品中文字幕看吧| 精品不卡国产一区二区三区| 看片在线看免费视频| 精品国内亚洲2022精品成人| 亚洲男人的天堂狠狠| 中文字幕av在线有码专区| 国产av麻豆久久久久久久| 中文在线观看免费www的网站| 精品国内亚洲2022精品成人| 亚洲av电影不卡..在线观看| 欧美高清成人免费视频www| 中文字幕高清在线视频| 欧美激情国产日韩精品一区| 久久久久国产精品人妻aⅴ院| 亚洲av成人精品一区久久| 色5月婷婷丁香| 国产亚洲精品综合一区在线观看| 亚洲一区二区三区不卡视频| 真人做人爱边吃奶动态| 黄色配什么色好看| 在线观看66精品国产| 国产精品人妻久久久久久| 特级一级黄色大片| 亚洲色图av天堂| ponron亚洲| 免费av毛片视频| www.色视频.com| 床上黄色一级片| 免费av毛片视频| 俄罗斯特黄特色一大片| 色播亚洲综合网| 99riav亚洲国产免费| 日韩欧美精品免费久久 | 日韩欧美在线二视频| 久久久久性生活片| 欧美另类亚洲清纯唯美| 亚洲avbb在线观看| 免费黄网站久久成人精品 | 久久久久国产精品人妻aⅴ院| 12—13女人毛片做爰片一| 国内少妇人妻偷人精品xxx网站| 午夜a级毛片| 色噜噜av男人的天堂激情| 久久久久久大精品| 男女之事视频高清在线观看| 1024手机看黄色片| 国产高清有码在线观看视频| 桃红色精品国产亚洲av| 天天躁日日操中文字幕| 午夜精品在线福利| 一区二区三区免费毛片| 色视频www国产| 熟女人妻精品中文字幕| 在线播放国产精品三级| 亚洲av电影不卡..在线观看| 91在线观看av| 亚洲国产精品合色在线| 99久久久亚洲精品蜜臀av| 日韩成人在线观看一区二区三区| 精品久久久久久久久亚洲 | 亚洲,欧美精品.| 亚洲精品在线观看二区| 88av欧美| 舔av片在线| 亚洲精品久久国产高清桃花| 久久久久久久精品吃奶| aaaaa片日本免费| 精品欧美国产一区二区三| 久久亚洲真实| 性欧美人与动物交配| 国产精品嫩草影院av在线观看 | 亚洲黑人精品在线| 亚洲七黄色美女视频| 国产精品免费一区二区三区在线| 制服丝袜大香蕉在线| 麻豆国产av国片精品| 极品教师在线视频| 精品99又大又爽又粗少妇毛片 | 午夜免费男女啪啪视频观看 | 国产伦精品一区二区三区视频9| 中文字幕av在线有码专区| 成人精品一区二区免费| 国产精品国产高清国产av| 两个人的视频大全免费| 深夜a级毛片| 国产高清视频在线观看网站| 国产精品精品国产色婷婷| 深夜精品福利| 毛片一级片免费看久久久久 | 欧美日韩黄片免| 国产色爽女视频免费观看| 成熟少妇高潮喷水视频| 人妻久久中文字幕网| 韩国av一区二区三区四区| 三级国产精品欧美在线观看| 18禁黄网站禁片免费观看直播| 村上凉子中文字幕在线| 日本精品一区二区三区蜜桃| 亚洲国产欧美人成| 精品免费久久久久久久清纯| 99riav亚洲国产免费| 俺也久久电影网| 午夜久久久久精精品| 欧美日韩国产亚洲二区| 亚洲第一区二区三区不卡| 国产三级在线视频| 国产免费av片在线观看野外av| 噜噜噜噜噜久久久久久91| 国产精品av视频在线免费观看| 性色av乱码一区二区三区2| 69人妻影院| 一本一本综合久久| 国产久久久一区二区三区| 久久久国产成人精品二区| 看免费av毛片| 国内少妇人妻偷人精品xxx网站| 午夜免费男女啪啪视频观看 | 国产免费一级a男人的天堂| 18禁黄网站禁片免费观看直播| 精品人妻1区二区| 欧美日韩中文字幕国产精品一区二区三区| 在线观看一区二区三区| 舔av片在线| 免费看美女性在线毛片视频| 18美女黄网站色大片免费观看| 一个人免费在线观看电影| 非洲黑人性xxxx精品又粗又长| 黄色丝袜av网址大全| 天堂网av新在线| 日本黄色视频三级网站网址| 亚洲精品成人久久久久久| 久久婷婷人人爽人人干人人爱| 日日干狠狠操夜夜爽| 嫩草影院精品99| 国产精品嫩草影院av在线观看 | 男女做爰动态图高潮gif福利片| 国产午夜精品久久久久久一区二区三区 | 日韩欧美三级三区| 欧美在线一区亚洲| 18禁在线播放成人免费| 一本久久中文字幕| 成年免费大片在线观看| 美女高潮的动态| 别揉我奶头 嗯啊视频| 国产精品国产高清国产av| 精华霜和精华液先用哪个| 嫩草影视91久久| 亚洲av美国av| 久久久久九九精品影院| 久久久久久久精品吃奶| 午夜激情福利司机影院| 看片在线看免费视频| 精品福利观看| 国内精品美女久久久久久| 精品乱码久久久久久99久播| 色精品久久人妻99蜜桃| 美女 人体艺术 gogo| 看免费av毛片| 免费看光身美女| 美女大奶头视频| 我要搜黄色片| 有码 亚洲区| 成人无遮挡网站| 亚洲狠狠婷婷综合久久图片| 18禁黄网站禁片午夜丰满| bbb黄色大片| 欧美高清性xxxxhd video| 九色国产91popny在线| 亚洲精品一卡2卡三卡4卡5卡| 3wmmmm亚洲av在线观看| 亚洲最大成人av| av女优亚洲男人天堂| 中文字幕人妻熟人妻熟丝袜美| 国产成人啪精品午夜网站| 757午夜福利合集在线观看| 村上凉子中文字幕在线| 此物有八面人人有两片| 变态另类丝袜制服| av视频在线观看入口| 无人区码免费观看不卡| 国产精品亚洲av一区麻豆| 国产熟女xx| 成人永久免费在线观看视频| 国产野战对白在线观看| www日本黄色视频网| 国产精品,欧美在线| 亚洲av美国av| 一本一本综合久久| 怎么达到女性高潮| 女同久久另类99精品国产91| 国产精品一及| 精华霜和精华液先用哪个| 一个人免费在线观看电影| 97热精品久久久久久| 97超级碰碰碰精品色视频在线观看| 蜜桃亚洲精品一区二区三区| 好看av亚洲va欧美ⅴa在| 亚洲国产精品sss在线观看| 少妇裸体淫交视频免费看高清| 蜜桃亚洲精品一区二区三区| 欧美绝顶高潮抽搐喷水| 少妇熟女aⅴ在线视频| 精品一区二区免费观看| 91久久精品国产一区二区成人| 69av精品久久久久久| 欧洲精品卡2卡3卡4卡5卡区| 特级一级黄色大片| 人妻夜夜爽99麻豆av| 哪里可以看免费的av片| 久久久色成人| 中文字幕久久专区| 欧美区成人在线视频| 男女做爰动态图高潮gif福利片| 亚洲av.av天堂| 久久久色成人| 午夜日韩欧美国产| 国内少妇人妻偷人精品xxx网站| 国产高清视频在线观看网站| 一本一本综合久久| 淫妇啪啪啪对白视频| 欧美三级亚洲精品| 精品午夜福利在线看| a级毛片免费高清观看在线播放| 日韩亚洲欧美综合| 国产麻豆成人av免费视频| 看黄色毛片网站| 国产高清视频在线播放一区| 日韩欧美精品v在线| 99久久无色码亚洲精品果冻| 99精品久久久久人妻精品| av黄色大香蕉| 国产精品1区2区在线观看.| 好男人在线观看高清免费视频| 舔av片在线| 一卡2卡三卡四卡精品乱码亚洲| 美女被艹到高潮喷水动态| 人人妻,人人澡人人爽秒播| 美女xxoo啪啪120秒动态图 | 99久久99久久久精品蜜桃| 最好的美女福利视频网| 国产主播在线观看一区二区| 99久久无色码亚洲精品果冻| 夜夜躁狠狠躁天天躁| 欧美不卡视频在线免费观看| 高清日韩中文字幕在线| 简卡轻食公司| 欧美日韩亚洲国产一区二区在线观看| 少妇高潮的动态图| 91av网一区二区| 日本a在线网址| 天堂网av新在线| 看黄色毛片网站| 国产三级中文精品| 99国产综合亚洲精品| 久久久色成人| 亚洲一区二区三区不卡视频| 大型黄色视频在线免费观看| www.色视频.com| 麻豆成人午夜福利视频| 国产综合懂色| 老司机深夜福利视频在线观看| 午夜福利高清视频| 中文字幕熟女人妻在线| 青草久久国产| 欧美高清性xxxxhd video| 欧美区成人在线视频| 亚州av有码| 亚洲无线在线观看| 极品教师在线视频| 不卡一级毛片| 在线观看一区二区三区| 美女cb高潮喷水在线观看| 淫妇啪啪啪对白视频| 欧美日韩国产亚洲二区| 丁香欧美五月| 久久这里只有精品中国| 99精品久久久久人妻精品| 91麻豆av在线| 精品一区二区三区视频在线观看免费| 99久久精品热视频| 乱人视频在线观看| 午夜久久久久精精品| 我的老师免费观看完整版| 91麻豆精品激情在线观看国产| 日韩av在线大香蕉| 国产精品永久免费网站| 黄色一级大片看看| 亚洲乱码一区二区免费版| 日本a在线网址| av在线天堂中文字幕| 国产 一区 欧美 日韩| 亚洲一区二区三区色噜噜| 久久精品国产清高在天天线| 欧美在线一区亚洲| 免费人成在线观看视频色| 97人妻精品一区二区三区麻豆| 国模一区二区三区四区视频| 亚洲人成网站高清观看| 欧美三级亚洲精品| 国产精品一及| 中国美女看黄片| 天美传媒精品一区二区| 简卡轻食公司| 免费在线观看成人毛片| 国产毛片a区久久久久| 亚洲18禁久久av| 熟女人妻精品中文字幕| 老司机福利观看| 亚洲国产高清在线一区二区三| 色综合站精品国产| 亚洲欧美激情综合另类| 国产精品98久久久久久宅男小说| 国产黄a三级三级三级人| 国产精华一区二区三区| 天天一区二区日本电影三级| 他把我摸到了高潮在线观看| 午夜老司机福利剧场| 国产亚洲精品综合一区在线观看| 欧美在线黄色| 夜夜爽天天搞| 天堂网av新在线| 一a级毛片在线观看| 亚洲欧美日韩卡通动漫| 真人做人爱边吃奶动态| 国产一区二区在线观看日韩| 中文字幕熟女人妻在线| 老女人水多毛片| 国产极品精品免费视频能看的| 亚洲精品色激情综合| 成人性生交大片免费视频hd| 免费看光身美女| av在线观看视频网站免费| 久久久久九九精品影院| 国产 一区 欧美 日韩| 久久精品91蜜桃| av在线天堂中文字幕| 18禁黄网站禁片午夜丰满| 99视频精品全部免费 在线| 少妇的逼好多水| 在线免费观看不下载黄p国产 | 欧美一级a爱片免费观看看| 一区二区三区激情视频| 久久久精品欧美日韩精品| 久久久久性生活片| 久久精品国产亚洲av涩爱 | 午夜a级毛片| 青草久久国产| 欧美激情久久久久久爽电影| 不卡一级毛片| 偷拍熟女少妇极品色| 国产欧美日韩精品一区二区| 日本熟妇午夜| 丰满乱子伦码专区| 国产淫片久久久久久久久 | 国产蜜桃级精品一区二区三区| 97热精品久久久久久| 亚洲午夜理论影院| 91av网一区二区| 欧美日韩福利视频一区二区| 日韩大尺度精品在线看网址| 男女下面进入的视频免费午夜| 长腿黑丝高跟| 久久久久精品国产欧美久久久| 麻豆久久精品国产亚洲av| 色吧在线观看| 可以在线观看的亚洲视频| 在线十欧美十亚洲十日本专区| 一进一出好大好爽视频| 简卡轻食公司| 欧美区成人在线视频| 一二三四社区在线视频社区8| 亚洲精品乱码久久久v下载方式| 精品人妻1区二区| 嫩草影院新地址| 中文字幕久久专区| 亚洲av美国av| 国产久久久一区二区三区| 91久久精品电影网| 国产精品三级大全| 国产高清三级在线| 老熟妇仑乱视频hdxx| 老鸭窝网址在线观看| 看免费av毛片| 我的女老师完整版在线观看| 少妇裸体淫交视频免费看高清| 看片在线看免费视频| 久久亚洲精品不卡| 最新在线观看一区二区三区| 国产精品久久久久久久久免 | 十八禁网站免费在线| 一级av片app| 一区二区三区激情视频| 精品久久久久久久末码| 一进一出抽搐动态| 日韩精品中文字幕看吧| 欧美性感艳星| 内地一区二区视频在线| 久久精品国产亚洲av天美| 91午夜精品亚洲一区二区三区 | 俄罗斯特黄特色一大片| 国产精品人妻久久久久久| 国产免费av片在线观看野外av| 老司机福利观看| 欧美激情久久久久久爽电影| 欧美一区二区国产精品久久精品| 国产色爽女视频免费观看| 国产野战对白在线观看| 三级毛片av免费| 久久精品国产亚洲av涩爱 | 免费观看的影片在线观看| 免费av不卡在线播放| 有码 亚洲区| 好男人在线观看高清免费视频| 国产亚洲精品av在线| 淫秽高清视频在线观看| 亚洲美女搞黄在线观看 | 尤物成人国产欧美一区二区三区| 国产野战对白在线观看| 久久午夜福利片| 两性午夜刺激爽爽歪歪视频在线观看| 亚洲国产日韩欧美精品在线观看| 欧美区成人在线视频| 中亚洲国语对白在线视频| 久久久久国产精品人妻aⅴ院| 国产黄色小视频在线观看| 国产美女午夜福利| av在线蜜桃| 久9热在线精品视频| 中文亚洲av片在线观看爽| 亚洲精品影视一区二区三区av| 欧美日韩国产亚洲二区| 黄色视频,在线免费观看| 美女被艹到高潮喷水动态| 亚洲av成人av| 午夜精品一区二区三区免费看| 五月玫瑰六月丁香| 天堂网av新在线| 天堂√8在线中文| 久久人妻av系列| 亚洲激情在线av| 精品一区二区三区视频在线| 日韩中文字幕欧美一区二区| 国产精品精品国产色婷婷| 亚洲在线自拍视频| 国产私拍福利视频在线观看| 亚洲不卡免费看| 久久精品国产自在天天线| 国产成人欧美在线观看| 午夜福利18| 欧美中文日本在线观看视频| 精品久久久久久,| 无人区码免费观看不卡| 一区二区三区高清视频在线| 99在线视频只有这里精品首页| 亚洲一区二区三区色噜噜| 亚洲最大成人手机在线| 波野结衣二区三区在线| 女人十人毛片免费观看3o分钟| 国产精品乱码一区二三区的特点| 日本a在线网址| 亚洲国产精品合色在线| 亚洲七黄色美女视频| 国产精品99久久久久久久久| 欧美性感艳星| 亚洲国产日韩欧美精品在线观看| 噜噜噜噜噜久久久久久91| 脱女人内裤的视频| 丁香欧美五月| 校园春色视频在线观看| 午夜福利在线在线| 很黄的视频免费| 亚洲精品在线观看二区| 美女 人体艺术 gogo| 又粗又爽又猛毛片免费看| 国产成人av教育| 97碰自拍视频| 能在线免费观看的黄片| 日本免费一区二区三区高清不卡| 日韩 亚洲 欧美在线| 亚洲 国产 在线| x7x7x7水蜜桃| 婷婷精品国产亚洲av在线| 美女 人体艺术 gogo| 国产三级中文精品| 精品人妻熟女av久视频| 久久久久久久久久成人| 日本免费a在线| 久久久久久久亚洲中文字幕 | 啪啪无遮挡十八禁网站| 免费av不卡在线播放| 69av精品久久久久久| 成人永久免费在线观看视频| 最近中文字幕高清免费大全6 | 首页视频小说图片口味搜索| 无遮挡黄片免费观看| 国产精品久久久久久久久免 | 久久天躁狠狠躁夜夜2o2o| 亚洲精品一卡2卡三卡4卡5卡| a在线观看视频网站| 全区人妻精品视频| 久久久久亚洲av毛片大全| 免费在线观看成人毛片| 少妇人妻一区二区三区视频| 国产成人欧美在线观看| 久久精品久久久久久噜噜老黄 | 97热精品久久久久久| 日韩中文字幕欧美一区二区| 日韩欧美三级三区| 亚洲三级黄色毛片| 麻豆国产97在线/欧美| 啦啦啦韩国在线观看视频| 亚洲精品粉嫩美女一区| 久久久久精品国产欧美久久久| 天堂动漫精品| 可以在线观看毛片的网站| 午夜a级毛片| 男插女下体视频免费在线播放| 欧美区成人在线视频| 麻豆成人午夜福利视频| 亚洲精品在线观看二区| 麻豆国产97在线/欧美| 国产伦在线观看视频一区| 69av精品久久久久久| 久久久久国产精品人妻aⅴ院| 久久精品影院6| 亚洲欧美日韩高清在线视频| 变态另类丝袜制服| 国产精品99久久久久久久久| 欧美一区二区国产精品久久精品| 欧美高清成人免费视频www| 日本a在线网址| 麻豆成人av在线观看| 亚洲国产精品成人综合色| 麻豆成人av在线观看| 国产精华一区二区三区| 精品人妻偷拍中文字幕| 国产成人欧美在线观看| 亚洲片人在线观看| 级片在线观看| 亚州av有码| 免费一级毛片在线播放高清视频| 偷拍熟女少妇极品色| 亚洲av二区三区四区| 久久性视频一级片| 中文字幕av成人在线电影| 老司机午夜福利在线观看视频| 国产主播在线观看一区二区| 九色国产91popny在线| 国产三级黄色录像| 国产免费男女视频| 亚洲av第一区精品v没综合| 麻豆成人av在线观看| 两个人视频免费观看高清| 99久久99久久久精品蜜桃| 毛片一级片免费看久久久久 | av在线老鸭窝| 亚洲av电影在线进入| 精品人妻一区二区三区麻豆 | 国产欧美日韩精品一区二区| 久久欧美精品欧美久久欧美| 欧美日韩亚洲国产一区二区在线观看| 精品乱码久久久久久99久播| 久久久久久久久中文| www.色视频.com| 国产av麻豆久久久久久久| 日本a在线网址| 永久网站在线| 日本黄色片子视频| 婷婷色综合大香蕉| 色尼玛亚洲综合影院| 亚洲激情在线av| 天堂影院成人在线观看| 久久精品综合一区二区三区|