LIBRARY Mlchlgan State Unlverslty PLACE II RETURN BOX to roman IN. checkout from your mood. TO AVOID F INES mum on or baton dd. duo. DATE DUE DATE DUE DATE DUE A BURST-ORIENTED TRAFFIC CONTROL FRAMEWORK AND ASSOCIATED CALL ADMISSION CONTROL SCHEMES FOR ATM NETWORKS By Jose Roberto Santiano Fernandez A DISSERTATION Submitted to Michigan State University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY Computer Science Department 1995 ABSTRACT A BURST-ORIENTED TRAFFIC CONTROL FRAMEWORK AND ASSOCIATED CALL ADMISSION CONTROL SCHEMES FOR ATM NETWORKS By Jose Roberto Santiano Fernandez Asynchronous transfer mode (ATM) networks are intended to support integrated digital communications traffic. This means that they are expected to handle different traffic generation characteristics ranging from smooth continuous bit rate transmission to highly irregular bursty data generation. This dissertation proposes a traffic control approach that is tailored for the efficient handling of highly bursty traffic in which related bursts of cells may be treated as units. The approach is based on a connection- oriented transmission mode that allows “on-the-fly” allocation of bandwidth in the switches. The proposal includes a framework of ATM switch mechanisms that allow a bursty connection to acquire resources at a switch when a burst arrives and to release them when a burst leaves. The framework supports bursty connections that are silent between their bursty periods as well as connections that may transmit at a non-zero rate between bursts. Furthermore, it allows the bundling of these bursty connections into virtual paths. The task of controlling congestion in ATM networks primarily rests on a call ad- mission control (CAC) scheme that limits the number of established connections. In this dissertation, three different CAC schemes that are geared towards maintaining the burst blocking probabilities (BBPS) of connections are presented. The first scheme is a highly computationally efficient approach based on the Erlang loss system model. This is compared with a more general model (already proposed by other researchers) that requires more computation. The second scheme is an extension of this general model to handle burst-level (rather than cell-level) priorities. The third scheme is another extension of the general model whose purpose is to take into account quantifi- able correlations between related connections. This approach is useful for supporting distributed parallel computation over large ATM networks. The CAC schemes have been evaluated using analyticaland Simulation techniques. The results Show that the proposals contribute to utilizing bandwidth more efficiently while maintaining the BBPs of admitted connections. To my Creator. iv ACKNOWLEDGMENTS First, I thank my research adviser, Prof. Matt W. Mutka, for his guidance and direction in this work. I also thank the following for supporting me in one way or another during my years at MSU: Pepito, Cielo, Bolet, Billy, Joji, Katrina, my Guardian Angel, and the Department Of Computer Science. TABLE OF CONTENTS LIST OF TABLES LIST OF FIGURES 1 Introduction 2 Background: High-Speed Network Technologies 2.1 High-Speed Network Standards for Limited Areas ........... 2.1.1 HIPPI ............................... 2.1.2 FDDI ............................... 2L3 DQDB ............................... 2.2 Services Offered by High-Speed Networks ................ 2.2.1 ISDN ................................ 2.2.2 Frame-Relay ............................ 2.2.3 SMDS ............................... 2.2.4 BISDN ............................... 2.3 Asynchronous Transfer Mode Networks ................. 2.3.1 The ATM Layer .......................... 2.3.2 The ATM Adaptation Layer ................... 2.3.3 QOS Guarantees in ATM Networks ............... 3 Related Research: High-Speed Network Traffic Control 3.1 Traffic Characterization and Modeling ................. vi xi CDCDKIKIGU‘ 10 11 11 11 12 13 14 15 17 18 3.2 Connection Establishment Techniques .................. 21 3.3 Preventive vs. Reactive Congestion Control .............. 23 3.4 Traffic Control Schemes ......................... 25 3.4.1 Usage Parameter Control Schemes ............... 26 3.4.2 Call Admission Control Schemes ................. 30 3.4.3 Burst-Level Control Schemes ................... 33 3.4.4 Special Schemes for Best-Effort Traffic ............. 36 3.4.5 Special Schemes for Delay-Sensitive Traffic ........... 4O 4 FTamework for Burst-Level 'h'afiic Control 41 4.1 Burst-Oriented Services .......................... 43 4.2 Connection Establishment ........................ 45 4.3 Eager Transmission Detection and Handling .............. 47 4.4 In-Band Burst Control Cells ....................... 49 4.5 Thin Logical Connections in a Virtual Path .............. 51 4.6 Fat Logical Connections ......................... 52 4.7 Summary ................................. 55 5 A Simple CAC Scheme Based on the Erlang Loss System Model 57 5.1 5.2 5.3 5.4 5.5 The M/G/k/k Service Model ...................... 59 Basic BBP Computation Method .................... 61 Bandwidth Requirement of a Class of Sources ............. 62 Multi—Class Generalization of the Erlang Loss System ......... 66 Performance Evaluation ......................... 67 5.5.1 General Results .......................... 67 5.5.2 Network Scenarios ........................ 68 5.5.3 Comparison with a Non-Poisson Model of Sources ....... 76 5.5.4 Simulation Results ........................ 78 vii 5.6 Summary ................................. A Burst-Level Priority CAC Scheme for Bursty 'It'affic 6.1 Motivation ................................. 6.2 Traffic Model ............................... 6.3 CAC Scheme for Connections with Identical Peak Rates ....... 6.3.1 Burst Admission Priority Policy ................. 6.3.2 Bandwidth Demand Probability Tables ............. 6.3.3 Computing the BBP Bounds ................... 6.3.4 Summary of CAC Scheme and Computational Requirements . 6.3.5 Allowing Connections to Use Both Priorities .......... 6.4 CAC Scheme for Connections with Non-Identical Peak Rates ..... 6.5 Performance of the Priority Scheme ................... 6.6 Summary ................................. A Correlation-Aware CAC Scheme for Bursty 'h'affic 7.1 Scenario .................................. 7.2 Traffic Model ............................... 7.3 CAC Scheme for Applications with Intra-Application Correlation 7.3.1 Burst Admission Policy ...................... 7.3.2 Probability Tables ........................ 7.3.3 Computing the BBP Bounds ................... 7.3.4 Summary of CAC Scheme and Computational Requirements . 7.4 Illustration and Performance Evaluation ................ 7.5 Summary ................................. Conclusion 8.1 Summary of Contributions ........................ viii 81 83 84 85 87 87 88 91 93 95 95 97 105 108 109 112 113 113 114 116 117 119 128 130 130 8.2 Future Work ................................ 133 BIBLIOGRAPHY 136 ix LIST OF TABLES 5.1 Parameters and settings for the network scenarios .............. 70 7.1 Table of parameters for the given example application. .......... 121 2.1 2.2 2.3 3.1 4.1 4.2 4.3 4.4 LIST OF FIGURES BISDN protocol reference model. ...................... Cell header formats. ............................. AAL protocols and service types. ...................... Generalized leaky bucket with marker and spacer .............. Thin and fat logical connections. The shaded areas indicate the instanta- neous bit rates of the sources and the heavy lines indicate the band- width dynamically allocated to the connections. ............ Required head cell lead times. Without multicasting, the time it will take for the head cell to reserve the bandwidth at the last link, after it is transmitted at source S, is given by 2;,(1); + h,-) plus the queueing delays. With multicasting, this reduces to 21;”), plus the queueing delays. In contrast, the first data cell may take only ELI p.- to reach the last switch. Note that if each h; is less than the cell switching time, then the queueing delays are not a factor in the multicasting case. . . A cell routing algorithm that is safe for eager transmission in VPS. The VP table and routing table, except for the VCBT index, are modeled after the Fore Systems ASK-100 switch architecture. However, any ATM switch will need a table to contain the output VPI for each established VP. That same table may be modified to contain the VCBT index. In fact, the VCBT index could be stored in the output VCI field, which is not used for nonterminating VPS .................... Bursts in tandem logical connections. The total bandwidth consists of TLC and CBR components. Shaded bursts are sent eagerly using the TLC VCI; these may be dropped by the network. Unshaded bursts are sent using the CBR VCI, which has bandwidth preallocated to it. Note that the eager bursts in (b), (d), and (e) are correlated. This phenomenon has to be taken into account in the traffic description. xi 12 13 14 27 43 48 53 56 5.1 5.2 5.3 5.4 5.5 5.6 5.7 5.8 6.1 M / G / k / 1: modeling of a class of connections. All 5' sources at the left have the same peak rate, which is the bandwidth of each of the 1: servers at the right, where k S S. The burst arrivals from the sources are Poisson processes with rates A,- and the burst durations are given by the means b5. Arriving bursts for which no server is available are dropped in the bit bucket .................................. Bandwidth allocation procedures. Procedure A is a portion of the call admission procedure that checks the bandwidth availability for a new connection. It is based on the table lookup alternative and ignores the subtleties with pmin described in the text. Procedure B is a portion of the corresponding burst admission procedure. ............. Maximum utilizations for the burst-oriented reservation scheme. The x- axis is the capacity available to a class of bursty connections of the same peak rate. The data points are the maximum achievable utilizations for the different choices of BBP. These maximum bounds apply when the number of connections exceeds the number of capacity units (virtual servers) in a class. ............................ Maximum utilizations for the 64 Kbps class. ................ Maximum utilizations for the 2 Mbps class. ................ Maximum utilizations for the 10 Mbps class ................. Comparison of Erlang loss system to non-Poisson model. The BBP is set to 10". .................................. Simulation results using different burst source models. The target BBP was set to 10". The individual points plot the observed BBPS of individual runs, and the lines connect the corresponding averages of five runs per source type. The legend acronyms indicate the source types as follows: “PP” stands for Poisson process; “MM” stands for Markov modulated; “N” stands for normalized; “I” stands for interrupted; and “S” stands for serialized. These are described in the text ........ Burst admission priority policy. Each box represents a bandwidth unit. A box marked “10” has been allocated to an admitted low-priority burst and a box marked “hi” has been allocated to an admitted high-priority burst. The policy allows for only up to m bandwidth units to be used by low-priority bursts but up to n bandwidth units to be used by high- priority bursts at the same time ...................... xii 59 65 69 72 73 74 77 79 88 6.2 6.3 6.4 6.5 6.6 6.7 7.1 7.2 Simulation results using different on-ofl' traffic sources. The target low and high-priority BBPS were set to 10"2 and 10“, respectively. The upper band consists of the Observed low-priority BBPS and the lower band consists of the observed high-priority BBPS. Individual points plot the results of individual runs and the lines connect the corre- sponding averages. Interrupted Poisson process (IPP), interrupted Markov-modulated Poisson process (IMMPP), normalized IMMPP (NIMMPP), and Markov-modulated on-ofi process (MMOOP), de- scribed in the text, were the types of sources used ............ Results of the same simulations in the previous figure, but with bursts that were not dropped when blocked. In these simulations, blocked bursts were allowed to persist in the system and later use bandwidth released by departing bursts ........................ Utilizations for the three approaches for different total capacities. The fixed parameters are: low-priority BBP = 10“, high-priority BBP = 10’”, individual source load p = 0.1, and individual peak rate = l bandwidth unit. The high-priority sources constitute approximately 10% of the total load ............................ Utilizations for the three approaches for different burstiness values. The fixed parameters are: low-priority BBP = 10", high-priority BBP = 10‘“, and capacity = 100 bandwidth units. The high-priority sources constitute approximately 10% of the total load. ............ Utilizations for the three approaches for different relative amounts of high- priority traffic. The fixed parameters are: low-priority BBP = 10"“, high-priority BBP = 10'“, individual source load p = 0.1, and capacity = 100 bandwidth units. ......................... Utilizations for the three approaches for different low-priority BBP val- ues. The fixed parameters are: individual source load p = 0.1 and capacity = 100 bandwidth units. The high-priority sources constitute approximately 10% of the total load. .................. As network capacities increase, distributed or parallel computing over a wide-area network may become desirable and feasible .......... Example distributed application diagram. M is the master host, N is an intermediate node, and S is the set of slave hosts sending messages to M via N. The link of concern is marked L ................ xiii 99 101 102 105 109 120 7.3 7.4 7.5 7.6 7.7 7.8 7.9 Utilizations allowed by the three CAC approaches. The target BBP was set to 0.001. The “correlated assumption” approach uses the proposed scheme; the “optimistic” approach assumes that all sources are inde- pendent; and the “conservative” approach assumes maximal overlap among messages from the same application. .............. Calculated BBPS for the three CAC approaches assuming that the actual traffic follows the correlated model. ................... Measured BBPS from simulations using workloads admitted by the correlation-aware CAC scheme. The target BBP was set to 0.001 as in the previous figures. The points plot the results of individual runs and the lines connect the corresponding averages. “Leading” and “trailing” bursts refer to the first and last bursts of each application cycle. . . . Results of the same simulations in the previous figure, but with bursts that were not dropped when blocked. In these simulations, blocked bursts were allowed to persist in the system and later use bandwidth released by departing bursts ........................ BBPS resulting from Simulations where the traffic load was effectively in- creased beyond that specified to the CAC scheme. The capacity was 100 bandwidth units, the target BBP was 0.001, and the (unadjusted) workload was that admitted by the “correlated” approach in the pre- vious simulations .............................. BBPs resulting from simulations where the intra-application overlap was effectively increased beyond that specified to the CAC scheme. The capacity was 100 bandwidth units, the target BBP was 0.001, and the (unadjusted) workload was that admitted by the “correlated” approach in the previous simulations. ....................... Diagram of the message overlap profile of an application, with three sources using the same link of concern. Figure (b) shows the effect of decreasing the relative message delays in (a) by 50% ......... xiv 123 124 126 127 129 Chapter 1 Introduction The future of telecommunications is in high-speed integrated networks. The most promising candidate for the technology that will facilitate the integration Of data, telephony, and cable television networks is the asynchronous transfer mode (ATM) network [11]. In this network, all data is transported in streams of small protocol units called cells that are relayed at high speed by Special hardware called ATM switches [66]. ATM networks are designed to carry all types of communication traffic including continuous bit rate (CBR), variable bit rate (VBR), and “best-effort”1 traffic. The main advantage of the ATM approach over circuit-switched networks or traditional packet-Switched networks lies in its ability to support bursty connections with a guar- anteed quality of service. The statistical multiplexing of bursty connections allows a higher bandwidth utilization than that which the circuit-switching approach could achieve with comparable bursty traffic. In addition, with proper network design and management, a quality of service in terms of low loss rate and / or bounded delay can be guaranteed to ATM network connections. It is expected that in the future, a large portion of integrated network traffic will 1Alternatively, “available bit rate” (ABR) is a new term for non-guaranteed service. 2 be generated by bursty sources. Examples of bursty applications that are already beginning to be more widely used are resource discovery tools [53] such as the World- Wide Web and Gopher. In addition, multimedia applications [4, 67] including video teleconferencing and multimedia database retrieval will be more commonplace. Fi- nally, basic electronic communications services including electronic mail and network news (i.e., USENET) will be practically universal. In order for ATM networks to successfully support a wide range of bursty applica- tions, a new traffic control strategy must be developed. The new approach has to be tailored for highly bursty connections while still efficiently supporting the less bursty or well-behaved connections. In particular, it has to tolerate difficult connections with extra long bursts or prolonged quiescent periods. Furthermore, it has to achieve an acceptable bandwidth utilization while still guaranteeing the required level of service for the bursty as well as the non-bursty connections. Finally it is desirable that the scheme not require potentially expensive hardware such as large buffers in all the switching nodes. To this end, a comprehensive framework for burst-level handling Of cell traffic has been developed. The distinguishing characteristic of this proposal is that bursts of cells that are logically related are handled as a unit rather than individually as independent entities. While this general approach has already been suggested by Hui [35], the proposed framework builds on the idea, fills in some of the important details, addresses some important related issues, and provides features such as burst- level priorities and “correlation awareness.” The highlights of the current proposal include the following. It supports “burst- oriented” connections with “on-the-fly” bandwidth allocation and eager transmission of bursts, where “eager” means that the bursts are transmitted before bandwidth has been committed and acknowledged for them. Furthermore, it supports the bundling of burst-oriented connections into virtual paths (VPS). This feature is not covered in 3 any of the other published proposals of a similar nature. It supports two varieties of bursty connections, including one with a zero rate between bursts and another that allows the connections to have a non-zero nominal rate between bursts. The proposal also includes three different but related call admission control schemes that guarantee limits on the loss rate of bursts. The first admission scheme proposed is a computationally efficient approach based on the Erlang loss system model. Its main attraction is that the aggregate load of a set of sources can be obtained additively from the individual loads of the sources. The bandwidth require— ment of the set can then be obtained from the aggregate load using a simple recursion or a table lookup. This approach is compared with a more general model that does not assume any burst arrival process, but just assumes that the sources are indepen- dent. The second proposed scheme extends the aforementioned general model (that has been investigated by other researchers) by adding burst-level priorities to it. To be more specific, it allows two sets of sources to experience different levels of burst loss rates while sharing the same bandwidth resources. The third scheme also uses the same general model but allows subsets of the sources to be correlated. In this scheme, the correlated sources are grouped into independent applications that are to be modeled separately. The primary purpose of this scheme is to properly consider the strong correlation that may exist among traffic sources participating in distributed parallel computation over an ATM network. The rest of the dissertation is organized as follows. An overview of high-Speed network standards and technologies is given in Chapter 2. Chapter 3 gives a survey of related research results in the area of high-speed network communication, with emphasis on traffic control schemes. Chapter 4 presents the proposed framework of protocols and mechanisms designed to facilitate the burst-level handling of traffic. The call admission control scheme based on Erlang’s loss formula is described and evaluated in Chapter 5. Chapter 6 presents and evaluates the proposed burst-level 4 priority scheme. The “correlation-aware” admission control scheme is given and eval- uated in Chapter 7. Finally, Chapter 8 concludes this dissertation with a summary of the work accomplished and possible directions for further research in the burst- oriented approach to ATM traffic control. Chapter 2 Background: High-Speed Network Technologies The information superhighway of tomorrow will be the result of the interconnection of the telephone, cable television, and data networks. In the past, telephone net- works primarily carried voice conversations, cable TV networks distributed one-way TV signals in guided media, and data networks carried computer data. However, nowadays it is not uncommon for telephone lines intended for voice to be used for the transmission of digitized information using fax machines or modems. Similarly, com- puter networks such as the Internet and nationwide or international on-line services (e.g., CompuServe, Prodigy, and America Online) are being utilized for the collection and dissemination of all kinds of information including text, still or moving—image pictures, and audio recordings. It is therefore apparent that there is a demand for a network infrastructure that is capable of carrying all sorts of information and pro- viding all kinds of services. Note that some of the desirable services from such a network require continuous transmission (e.g., live audio or video) and that some of the services require a large available bandwidth (full-motion video). For this reason, new network technologies are being developed that satisfy both requirements. 6 In summary, here are some of the desirable properties of the network infrastructure for the information superhighway of the future: a high-speed (e.g., sufficient for full-motion video) a support for different traffic requirements (e.g., connection-oriented, connection- less, smooth, bursty), and o scalability over greater data rates and distances. Section 2.1 is an overview of competing network technologies intended for limited areas of coverage. Section 2.2 discusses some of the high-speed network services that are available or are being planned. Section 2.3 describes the ATM network, which is a viable all-purpose high-speed network with an unlimited area of coverage. 2.1 High-Speed Network Standards for Limited Areas In this section, some of the more popular standards for high-speed networks are described. These networks Offer high-speed networking solutions only for limited areas (i.e., local and metropolitan areas), but are worthwhile to mention in this work because they offer competing solutions for their intended areas of coverage. In addition, they share some characteristics with ATM networks: 0 HIPPI uses interconnection fabrics such as crossbar switches to interconnect multiple links. 0 Both FDDI-Il and DQDB support connection-oriented, connectionless, and iso- chronous (i.e., periodic or circuit-switched) traffic. 7 o DQDB cells are similar to ATM cells. Both DQDB and ATM networks can support SMDS (described in Section 2.2.3 below). 2.1.1 HIPPI The High-Performance Parallel Interface (HIPPI) [80] is a point-to—point interface for transferring data at high-speeds over short distances. The standard was developed by the ANSI Task Group X3T9.3 based on a proposal from the Los Alamos National Laboratory. The primary application of HIPPI is the interconnection of supercomput- ers and associated equipment such as visualization terminals and high-performance storage systems. The basic circuit of HIPPI is a simplex point-to-point link, a collection of which can be organized into a more sophisticated interconnection system such as a crossbar switch. The physical layer of HIPPI is a set of copper twisted-pair cables capable of parallel 32-bit or 64-bit data transmission at a total rate of 800 or 1600 Mbps and at a distance Of up to 25 meters (about 82 feet). Not part of the standard, but rather an “implementor’s agreement,” is a proposal for a serial version of HIPPI based on fiber for distances of up to 10 km (about 6 miles). 2.1.2 FDDI Fiber Distributed Data Interface (FDDI) [1, 42, 73, 74, 79, 84] is a set of standards for a high-speed local or metropolitan area network. It was developed by the ANSI X3T9.5 Task Group. Its main features include the following. o the use of fiber links at the rate of 100 Mbps 0 a dual-ring topology for fault tolerance 8 o a medium access control (MAC) protocol based on the IEEE 802.5 token-ring standard [42, 73, 74, 79, 84] but taking advantage of the token rotation time 0 dynamic bandwidth allocation with synchronous and asynchronous transmission 0 support of up to 1000 stations over distances Of up to 100 km1 (about 60 miles) 0 frame sizes Of up to 4500 octets F DDI was originally intended for data traffic and its synchronous transmission service suffers from having variable delay (even though it is bounded). F DDI—II is an upward—compatible extension to FDDI that supports isochronous service in addition to the synchronous and asynchronous packet-switched service. The FDDI standard has evolved to the point that its original name has become practically a misnomer. The latest standard allows for any transmission medium (including optical fiber and twisted-pair wire), distributed or centralized control, any traffic (including data, voice, and video), and may refer to an interface, LAN, or MAN [37]. aLs DQDB Distributed Queue Dual Bus (DQDB) [1, 42, 74, 84] is the standard proposed by the IEEE 802.6 committee for metropolitan area networks. It is based on the Queued Packet and Synchronous Circuit Exchange (QPSX) protocol developed at the Uni- versity of Western Australia. Its main features include the following. o a dual unidirectional bus topology 0 use of a variety of media including Optical fiber and coaxial cable, at data rates ranging from 34 to 155 Mbps and higher k 1More precisely, the standard specifies a maximum fiber length of 200 km. 9 a support of connection-oriented and connectionless packet—switched data services and isochronous service 0 support of over 500 nodes on a 160 km (about 100 miles) dual bus network 0 use of 53-Octet slots (48 octets Of which are the data payload) within the large frames Of the underlying transmission system In addition, the standard includes the Option of using a looped dual bus topology for fault tolerance. In order to support both packet-switched and circuit-switched services, DQDB uses two bus access control methods. The queue arbitrated (QA) access method is based on a distributed queue concept where each node keeps track of its position in the global distributed queue of each bus. This method is intended for data services that are not delay-critical. The pre-arbitrated (PA) access method is intended for isochronous applications such as voice transmission. It is based on reserving a portion of PA slots that are generated periodically by the head nodes of each bus. Research on performance aspects, fairness, and enhancements of DQDB are con- tinuously worked on. Sadiku and Arvind have compiled an annotated bibliography in [69]. 2.2 Services Offered by High-Speed Networks Traditional X.25 [72, 73, 75, 79] packet-switching service and the plain old telephone service (POTS) comprise the bulk of communication services today. These are low- speed services with X.25 limited to 64 Kbps and POTS limited to 4 KHz analog signals (64 Kbps digital). ISDN, frame-relay, SMDS, and BISDN are high-speed services being Offered or being planned for high-speed networks. ISDN and BISDN, in addition, provide integrated circuit-switched and packet-switched services. 10 2.2.1 ISDN The Integrated Services Digital Network (ISDN) [42, 72, 73, 75, 79, 84] is a small step toward replacing the POTS with a multipurpose communication facility capable of providing a variety of services such as voice, data transfer, fax, slow-scan video, and interactive terminal services. ISDN services are provided to customers at two levels. Basic rate interface (BRI) access is intended or residential customers and consists of two full—duplex 64 Kbps B-channels (B for bearer) and one full-duplex 16 Kbps D-channel (D for data) on a 192 Kbps link. Telephone, fax, slow-scan TV, and packet-switched data services are provided on the B-channels. The D-channel is primarily intended for signaling (e.g., call setup) but can also be used for telemetry and low-rate packet-switched data. Primary rate interface (PR1) access, which is intended for larger customers such as businesses, provides 232 B-channels and a 64 Kbps D-channel totaling 1.544 Mbps, which is the data rate of the SO-called Tl carrier service. The FBI may also be reconfigured to carry higher bit rate channels called H-channels up to the capacity Of the underlying carrier. ISDN, like most modern telephone systems, uses common channel signaling (CCS) [72, 73, 75, 79] in which all signaling information is transmitted over a separate network, typically an X.25 packet-switching network. However, in addition to having a circuit-switched network for supporting synchronous services like telephone service, it also has a packet-switched network for packet-switched data service separate from the CCS network. 230 for countries using the El rather than the T1 carrier link 11 2.2.2 flame-Relay Frame-relay [11, 42, 73, 75] is a connection-oriented fast packet-switching service intended for intermediate speeds from about 64 Kbps to 1.544 Mbps. Traditional packet switching based on X.25 has a large per-packet and per-node processing over- head that makes it unsuitable for rates higher than about 64 Kbps. Frame-relay reduces this overhead by allowing variable~sized frames to be forwarded as soon as possible, even before the whole frame is received at a node. This is practical in mod- ern transmission facilities that have very low bit error rates and high data rates, since only a small fraction of the bandwidth is wasted by forwarding frames that may be corrupt. 2.2.3 SMDS Switched Multimegabit Data Service (SMDS) [11, 42] is a connectionless packet- switching service specification appropriate for data rates higher than those provided by frame-relay. It is intended to provide features usually found in local area networks, such as high throughput and low delay, but for large metropolitan areas. Access to SMDS is provided through a three-layer SMDS Interface Protocol (SIP) in which data packets of up to 9188 octets are broken up to be transmitted in 53-Octet cells appropriate for DQDB, the underlying transport subnetwork. SMDS does not provide isochronous service. 2.2.4 BISDN Broadband ISDN (BISDN) [42, 72, 73, 75, 84] is the higher bandwidth version of ISDN that supports channels with rates higher than the ISDN primary rate. These rates ranging from under 50 Mbps to over 150 Mbps are capable of supporting full-motion video services of varying quality. These would make available broadband services 12 / Management Plane Control Plane User Plane Hi gher-Layer Protocols Higher-Layer Protocols ATM Adaptation Layer (AAL) Asynchronous Transfer Mode (ATM) Layer / Physical Layer Figure 2.1: BISDN protocol reference model. such as digital TV and full multimedia communication on a single network. The underlying network technology proposed for supporting BISDN is the Asynchronous Transfer Mode (ATM) network based on the cell-relay concept. The underlying high-speed transmission system is proposed to be based on the Syn- chronous Optical Network (SONET) standard [16, 42, 72]. The ATM network is covered in Section 2.3 below. 2.3 Asynchronous Transfer Mode Networks CCITT3 has proposed a protocol reference model for BISDN [7, 16, 60, 63, 73, 75], which is shown in Figure 2.1. The physical layer is based on SONET [16, 42, 72], which is a hierarchy of optical signaling rates starting at 51.84 Mbps and ascending to the multi-Gbps range. The ATM layer is in charge of the multiplexing and switching of 53-octet data units called cells. The AAL is an endpoint-to-endpoint layer that provides a variety of services on top of the ATM layer. 3International Telegraph and Telephone Consultative Committee, now known as the Interna- tional Telecommunications Union-Telecommunications Standardization Sector (ITU-TSS or ITU-T for short) 13 (a) Cell header format at the UNI. (b) Cell header format at the NNI. 87654321 bits 87654321 GFC VPI octet l VPI VPI VCI octet 2 VPI VCI VCI octet 3 VCI VCI PT! (11.? octet 4 VCI PTI CLP HEC octet 5 HEC Figure 2.2: Cell header formats. The higher-layer protocols are divided into a user plane and a control plane to pro- vide out-Of-band signaling over the same ATM subnetwork. Finally, the management plane manages all the planes and layers in the protocol stack. 2.3.1 The ATM Layer Asynchronous Transfer Mode (ATM) [7, 11, 16, 24, 42, 60, 72, 75] is a technique of transporting, multiplexing, and switching fixed-sized data units called cells using special hardware called ATM switches [66]. An ATM cell consists of a 5—octet header field and a 48—octet payload. The ATM layer processes a cell based solely on the information in the header field, the format of which is Shown in Figure 2.2. Figure 2.2(a) shows the format of the header at the User-Network Interface (UNI), which includes a generic flow control (GFC) field that is used to control flow or indicate priority levels at the edge of the network. The GFC field is not used inside the network and is not included in the header format at the Node-Network Interface (NNI) shown in Figure 2.2(b). The virtual path identifier (VPI) and virtual channel identifier (VCI) indicate the virtual connection to which the cell belongs. ATM switches contain routing tables and procedures for properly switching (i.e., forwarding) cells based on their VPI and AAL type AAL-1 AAL-2 AAL-3 AAL-4 AAL-5 Synchronization required not required Bit rate constant? vafiable Connection mode connwtion—ofiented (CO) connectionless CO Figure 2.3: AAL protocols and service types. VCI labels. If the cell belongs to a virtual path connection, then it is switched based on the VPI alone—the destination port and VPI are obtained from the routing table and the VCI is left unaltered. Otherwise, the cell is switched according to the VPI and VCI combined, and new VPI and VCI labels are obtained from the routing table. The payload type identifier (PTI) indicates whether a cell is a user cell or some other special-purpose cell. The cell loss priority (CLP) bit indicates whether the cell is recommended for dropping in case of congestion. This is used to distinguish between cells that are costly to lose from those that are less 80. Finally, the header error control (HEC) field contains a Cyclic Redundancy Check (CRC) code used to detect bit errors in the cell header. 2.3.2 The ATM Adaptation Layer The ATM adaptation layer (AAL) [7, 11, 16, 24, 42, 60, 75] provides a variety of services for applications with different traffic requirements. A list of different proposed AAL protocols (other than the null AAL) and the types of traffic characteristics they support are given in Figure 2.3. AAL-1 is essentially circuit emulation and is appropriate for constant bit rate (CBR) synchronized services such as CBR audio and video. AAL-2 is appropriate of variable bit rate (VBR) synchronized services such as the variable-rate (i.e., com- pressed) versions of audio and video. 15 AAL-3, AAL-4, and AAL-5 are intended for bursty data services. AAL—3 and AAL-4 used to be separate standards but these have been merged into one protocol called AAL-3/4 because of their similarities. AAL—3/4 provides two modes of data service: message mode service for framed data (compatible with SMDS) and streaming mode service for data streams. AAL—5 [77] is a simple and efficient AAL protocol that is optimized for connection-oriented data service. It has less processing and transmission overhead that the more general-purpose AAL-3 / 4. The AAL protocols are divided into two sublayers. The segmentation and reassem- bly (SAR) sublayer produces cells from the higher-layer data units (segmentation) at the sending end, and reconstructs the higher-layer data units (reassembly) from cells at the receiving end. The convergence sublayer (CS) performs adaptation functions other than segmentation and reassembly, such as synchronization for AAL-1 and AAL-2 and cell loss and error detection for AAL-3/ 4 and AAL-5. 2.3.3 QOS Guarantees in ATM Networks ATM virtual connections are provided with quality of service (QOS) guarantees [11, 46, 83]. This means that during the life of the connection, its traffic will experience cell losses and delays that are within specified ranges. Physical transmission systems are always subject to bit errors and hence will occasionally cause cells to be lost when the cell header is corrupted. ATM networks do not perform link-by-link error correction; therefore, these losses are not recovered within the network. Another cause of cell losses is buffer overflow within the ATM switches. Because of the stochastic and asynchronous nature of the traffic in ATM networks, buffer overflow is inevitable. Cell delay and delay jitter (variation in the delay) happen as a result of the competition that exists between active connections and the buffering of cells within 16 ATM switches. Again, the stochastic and asynchronous nature of the ATM approach contributes to the occurrence of cell delay and delay jitter. In order to control the cell losses and delays among the connections in an ATM network, before a new connection is admitted, the question has to be asked [46] “Can the requested call be accepted by the network at its requested QOS, without violating existing QOS guarantees made to on-going calls?” To answer this question, the network will require, among other things, descriptions of the source traffic characteristics and the Q03 requirements of the new connection. The traffic descriptors may include parameters such as the mean rate, peak rate, mean burst length, maximum burst length, and proportion of time in the active state. The QOS requirements may be specified in terms of throughput, cell loss rate, burst blocking probability, delay bounds, and delay jitter bounds. Guaranteeing the Q08 of connections is a matter of employing a variety of traffic control techniques. Some Of the available techniques are the following: 0 rate control and/or flow. control 0 traffic enforcement and shaping a bandwidth and buffer management These are covered in Chapter 3. F inalIy, it is instructive to explain what is meant by “best-effort” traffic or “best- efl'ort” service. In its common usage in the data communications field, “best-effort” is not to be interpreted as “highest quality” (i.e., no better service exists). Rather, it is to be considered as an abbreviation for “best but non-guaranteed delivery us- ing residual resources.” A new term, available bit rate (ABR), has been coined to distinguish it from (and to rhyme with) CBR and VBR. Chapter 3 Related Research: High-Speed Network Traffic Control This chapter covers research topics in high-speed network communication that are related to the author’s subject of interest, which is burst-level traffic control in ATM networks. Network traffic and source traffic characterization and modeling are rel- evant in the design and performance evaluation of traffic control schemes. Traffic characterization and modeling is covered in Section 3.1. In high-speed networks, fast connection establishment techniques are needed for minimizing the bandwidth wasted during connection setup, and for efficiently supporting connectionless services. Con- nection establishment techniques are discussed in Section 3.2. High-speed wide-area networks require preventive rather than reactive congestion control. This is explained in Section 3.3. Finally, a variety of traffic control schemes for high-speed networks are described in Section 3.4. (N.B. In this chapter, with a few exceptions, the related work reviewed in each (sub)section is presented in chronological order of publication.) 17 18 3.1 Traffic Characterization and Modeling In this section, various studies of network and source traffic characteristics, and mod- els for such, are discussed. Heffes and Lucantoni [32] present a performance study of a statistical multiplexer of packetized voice and data traffic. In the study, the superposition of packetized voice sources is modeled using a Markov-modulated Poisson process (MMPP). Each voice source is assumed to be a renewal process that alternates between talk spurts that are approximately exponentially distributed (actually geometric; i.e., discrete) and silence periods that are exponentially distributed (with a different mean). The superposition of these processes is a complex non-renewal process. A two-state MMPP is a doubly-stochastic Poisson process whose arrival rate is determined by a two-state continuous-time Markov chain. It is also a non-renewal process but is simpler than the superposition process. It has four parameters: the mean sojourn times of the two states of the Markov chain, and the arrival rates of the Poisson process at each state. Using sophisticated techniques, the four MMPP parameters are obtained to match four statistical characteristics (involving first, second, and third moments) of the superposition process. The MMPP approximation is compared with results from simulation and is shown to be more accurate than a comparable Poisson process model. In [38], Jain and Routhier propose a model for network traffic based on “packet trains.” In the packet train model, packets are assumed to travel in groups called trains. Trains are identified in terms of a maximum allowed inter-car gap (MAIG) parameter. Interarrival times less than the MAIG indicate-“cars” in the same “train” and those greater than the MAIG indicate gaps between distinct trains. The packet train model is therefore defined ‘ by specifying an inter-train arrival process and an inter-car arrival process. The authors analyze the packet arrivals in an actual 10 Mbps 19 token-ring network and Show that the packet train model fits the observed arrival process much better than the Poisson or compound-Poisson process models. The authors summarize the appropriate parameters (e.g., mean and coefficient of variation of the inter-car time) for that network but do not propose any well-fitting probability distributions. In [15], Caceres and others discuss the results of their study of the characteristics of wide-area TCP/IP conversations. In the study, conversations between application endpoints communicating via TCP/IP are analyzed in terms of conversation length (in bytes), conversation duration (in ms), number of packets transferred, packet sizes, and interarrival times. The results of the study include the following (among others): 0 Most classic bulk transfer applications transfer less than 10 Kbytes per connec— tion. 0 A large portion of bulk transfer applications show strongly bidirectional traffic flow. 0 While interarrival times for bulk transfers exhibit the packet-train phenomenon, the interarrival times for interactive applications closely resemble a constant plus exponential distribution. 0 Most interactive applications generate 10 times more data in one direction than the other. In the article, the authors also suggest a way to use the findings in the study to randomly generate source traffic for an internetwork traffic simulation model. Wirth and Meier-Hellstern [85] propose a canonical model for ISDN packet data traffic. The model is based on a three-state Markov chain packet generator. One state represents machine-generated packet arrivals, another one represents active typing, and the last one represents pauses such as host response time and user think time. i 20 Each state generates a run of packet arrivals. The interarrival time distribution, run length distribution, initial probability, and transition probabilities of each state are estimated from actual measurements Obtained from monitoring real ISDN user activity. The resulting model parameters and approximating distributions, which are quite complicated, are given in detail in the article. In [10], Bottomley and Nilsson present their study of the traffic characteristics of the CON Cert network, which is a wide-area network covering a great portion of the state of North Carolina. In their study, they observe the following. o The interarrival distribution does not appear to be exponential. In particular, the coefficients of variation of the interarrivals are observed to be between 1.3 and 1.7 (compared to 1.0 for the exponential distribution). 0 The packet length distribution is clearly bimodal with peaks at the minimum and maximum packet sizes. a Two of the three sampling locations appear to have correlated interarrival times. The other location, which happens to be the busiest, does not Show signifi- cant correlation. (Does this suggest that the combination of a large number of processes does eventually result in an uncorrelated process? This is an open question.) The article also summarizes the results of their modeling efforts using a semi-Markov process and two-state MMPP as in [32]. Both models seem to fit the interarrival distribution well. However, both models do not seem to capture the correlation process very well. In any case, the MMPP model is recommended in lieu of a renewal process model (e.g., a Poisson process). Leland and others evaluate Ethernet LAN (also known as a CSMA/ CD or IEEE Standard 802.3 LAN [79, 73]) traffic in [47]. The main proposition Of the article is 21 that Ethernet LAN traffic is a self-similar or fractal-like process. In particular, the “burstiness” of Ethernet packet arrivals exhibits itself across a wide range of time scales. This conclusion was drawn from rigorous statistical analysis of a very large set of high-precision Ethernet measurements. In the article, the authors contend that their findings have dire consequences on congestion control in future broadband networks. However, it is not clear how the traffic properties of a LAN using a largely undisciplined connectionless protocol would apply to a broadband network, such as an ATM network, that is based on connection-oriented controlled access, i.e., call admission control and usage parameter control (see Section 3.4). That is, it may well be the case that most of the self-similarity would be ironed out by these access control mechanisms. However, their observations would very likely be relevant in the context of ATM LAN emulation [50] or connectionless “best-effort” protocols in general (see Section 3.4.4). 3.2 Connection Establishment Techniques In this section, various proposed connection establishment techniques for high-speed networks, including ATM-like networks, are discussed. An adaptive dynamic technique of caching virtual circuits in connection-oriented packet-switching networks is proposed by J0mer in [39]. The proposal is about using an X.25 (connection—oriented packet switching) network1 to connect LANS. The LANS communicate in a connectionless fashion using the X.25 network as a virtual data link service. In the system, virtual circuits are cached in the sense that a virtual circuit that is set up to transfer a packet from one LAN to another is kept for some time after that packet is delivered so that it may be reused by a subsequent packet with 1Granted, an X.25 network is not a high-speed network; however, the ideas in the article seem to be applicable to faster networks that are also based on virtual circuit communication. 22 the same source and destination LAN. An unused virtual circuit is released only after a timeout expires. The author proposes a traffic adaptive timeout mechanism that varies the timeout value depending on the current traffic load in order to optimize the performance of the system. Cidon and others [17] give a fast algorithm for setting up connections in fast packet—switching networks. A key objective in their design is to make bandwidth reservation very fast in order to reduce the call blocking that is due to call establish- ment request processing overhead. Routing is done using Automatic Network Routing (ANR), also known as source routing, in which the route of a packet is predetermined at the source and prefixed to the data in the packet. The algorithm takes advantage of a selective copy feature that allows an establishment request packet to be forwarded to the next node while it is being delivered to the current node’s network control unit. In addition to the detailed protocol algorithms, proofs of correctness are provided in the article. Doeringer and others [21] propose a hierarchical approach to connection establish- ment in large-scale networks. In the proposal, the network is organized in a hierarchy so that in the global view of the network, the internals of the nodes are hidden. A node in the global view may be non-trivial in the sense that it may consist of several interconnected sub-nodes or a set of multiple internal paths in a switching fabric. The internal (intra-node) details of the connection establishment procedure (i.e., path computation and bandwidth reservation) are done in parallel among all the nodes after the path is computed on the network (inter-node) level. The authors argue that the parallel approach is more efficient than the traditional approach and that its execution time is independent of the network complexity but is essentially a function of the round-trip delay. Olsen and Landweber [55] present a method for fast virtual channel establishment in ATM-like networks. The method, called NOW (NO Waiting), uses in-band signaling 23 to allow the data cells to be sent immediately after the VC connection establishment message. Logically, the technique is equivalent to using preestablished virtual paths in the network so that setting up a connection becomes a matter of choosing available VCIS toward the destination. Scalability is achieved via partitioning of the network. The method does not concern itself with bandwidth allocation, only VC label man- agement. A software implementation described in the article requires the buffering of a few data cells while the connection is being set up. A hardware design that is supposed to be fast enough to require no buffering is also briefly described in the article. Finally, a short overview of routing in ATM networks is presented by Onvural and Nikolaidis in [56]. The paper addresses some of the issues in routing with virtual circuits and virtual paths including bandwidth reallocation and VC/ VP rerouting. The paper also discusses different routing methodologies from the ATM network per- spective. The same authors have also compiled a bibliography on performance issues in ATM networks, including routing, in [51]. 3.3 Preventive vs. Reactive Congestion Control Traditional low-speed networks control traffic congestion via reactive means such as the use of source quench messages [18, 73] or choice packets [72, 73, 79, 84]. Similarly, link-by-link or end-to—end flow control is based on feedback-based techniques such as stop-and-wait and sliding window protocols [72, 73, 84]. These techniques are practical for low-speed networks because in those networks, the time to send a message is dominated by the transmission time (“the time to put the bits on the medium”). The high-speed links in future networks will shrink the transmission time consid- erably, making it insignificant compared to the propagation delay (“the time for the bits to travel through the medium”) in wide-area networks. The large propagation 24 delay to transmission time ratio will make reactive and feedback-based traffic control techniques inappropriate for high-speed networks for reasons of efficiency and cost. On the one hand, if a conservative approach is used, such using stop-and-wait or reasonably-sized sliding windows, then the network will be under-utilized (see [73]). This is so because the links will tend to be idle for relatively long periods while the endpoints are waiting for acknowledgments for messages that are transmitted almost instantaneously. On the other hand, if an aggressive approach is used, then unless very large receiving-end buffers are available throughout the network, large amounts of data may be lost during periods of congestion. This is a consequence of the large propagation delay experienced by the congestion control messages themselves—by the time these messages reach the sources, megabits or gigabits of data may already be on the way (i.e., traveling on the medium). (See [43] for a more detailed discussion of the above issues). The preferred approach for high-speed networks in the future is therefore can- gestion avoidance or preventive congestion control. Preventive congestion control in ATM networks, in which data transmission is connection-oriented, can be broadly cat- egorized into call admission control and usage parameter control. Both of these types of controls are essential in guaranteeing quality Of service and work hand-in-hand. Call admission control (CAC) schemes control the traffic entering the network by limiting the number of VC and VP connections in the network. A CAC scheme accepts or rejects connection requests based on the traffic characteristics indicated in the request and the traffic load offered by the other preestablished connections. CAC schemes are concerned with high-level resource allocation including bandwidth and buffer allocation. These are discussed in Section 3.4.2 below. Usage parameter control (UPC) schemes are applied to the admitted connections to ensure that they comply with the accepted traffic description. Since they control the rate at which an application or site sends information into the network, they 25 are also known as rate control schemes. Similarly, they are sometimes called source policing schemes and their job bandwidth enforcement. Some UPC schemes change the behavior of the input traffic streams and are therefore called traffic shaping schemes. UPC schemes are covered in Section 3.4.1 below. 3.4 Traffic Control Schemes This section covers related research proposals that have been published in the area of traffic control in high-speed networks and especially ATM networks. To begin, here are a number of published surveys or overviews on traffic control issues and proposals .— - VT f¥€l for high-speed networks. 0 The survey of Bae and Suda [7] covers traffic source modeling (especially voice and video), admission control, bandwidth enforcement (especially leaky bucket), priority control, and ATM networks and protocols. o The overview of Vakil and Saito [83] covers traffic description, admission control (several approaches), usage parameter control (especially leaky bucket), cell loss priority (CLP), and other congestion control strategies. 0 In [30], Habib and Saadawi discuss congestion avoidance in high-speed networks at three levels including the network level (management of links and virtual paths to minimize virtual call blocking), the call level (admission control), and the cell level (leaky bucket and other schemes). 0 The overview of Giin and Guérin [29] discusses an integrated set of controls including, among others, bandwidth allocation based on equivalent capacity [28] and rate control based on a buffered leaky bucket and Spacer scheme [8]. 26 o Doshi and Johri [24] survey communication protocols for high-speed networks. Some relevant topics include ATM protocols, congestion control, admission con- trol, and dynamic bandwidth controls (e.g., in—call negotiation). o Nikolaidis and Onvural [51] provide an extensive bibliography on performance issues in ATM networks, including traffic source characterization, call admission, congestion control, source policing, and fairness. The rest of this section is organized as follows. Section 3.4.1 goes over vari- ous proposals for usage parameter control and Section 3.4.2 deals with proposals for call admission control. Burst-level traffic control proposals are covered in Sec- tion 3.4.3. Section 3.4.4 covers special schemes for best-effort traffic. Finally, a few special-purpose schemes for delay-sensitive traffic (e.g., audio and video) are covered in Section 3.4.5. 3.4.1 Usage Parameter Control Schemes Most likely, the most widely-known input rate control schemes are those based on the “leaky bucket” scheme which is described by Turner in [81]: “. . . A counter associated with each user transmitting on a connection is incremented whenever the user sends a packet and is decremented peri- odically. If the counter exceeds a threshold upon being incremented, the network discards the packet. The user specifies the rate at which the counter is decremented (this determines the average bandwidth) and the value of the threshold (a measure of burstiness). . . .” There are many variations and extensions of this scheme. One generalization proposed in [8] is a combination of a leaky bucket, marker, and spacer shown in Fig- ure 3.1. This particular version is for variable-sized packets that consume a variable 27 Token generation it Red tokens Waiting buffer Is I I I I I ATV—“Ii, i token bucket To network ; : empty? : ......... I I I I I I : Token : [ bucket O : : O ' I [ Spacer O : I ___________________________ _ _ _ ' Periodic leak Figure 3.1: Generalized leaky bucket with marker and spacer. 28 number of tokens proportional to their lengths (for fixed-sized cells, the bottom to- ken bucket will only have to hold at most one token). Colored tokens are obtained from the top buckets and dropped into the bottom bucket which leaks periodically at a fixed rate determined by the maximum rate allowed for the source. High-priority green tokens are generated according to the reserved average rate of the traffic source. Low-priority red tokens are generated to accommodate traffic exceeding the reserved rate as “best-effort” traffic. These are used whenever a packet cannot conveniently wait for the necessary green tokens because the buffer occupancy has reached a certain threshold. A packet waits in the buffer until the bottom bucket is empty and is then marked “green” or “red” as appropriate. It then drops its tokens in the token bucket before entering the network. Inside the network, green packets are given preferential treatment using a buffer-threshold policy. This is done by limiting the number of red packets in the node buffers, even when the buffers are not yet completely full, in order to leave room for additional green packets. Another generalization of the leaky bucket scheme is presented by Bovopoulos in [12]. The scheme, called traffic control mechanism (TCM), is essentially a buffered leaky bucket scheme with a more sophisticated token generation pattern. Instead of generating the tokens in a constant periodic rate, they are generated according to a multi-state, cyclic, and deterministic Markov chain. For each state of the Markov chain, a fixed number of tokens is generated. The time between state transitions of the Markov chain is fixed. It is claimed in the article that the more general token generation pattern will support a wider variety of traffic behaviors. The effectiveness of the (unbuffered) leaky bucket scheme is analyzed by Butto’ and others in [14]. Their results Show that the leaky bucket can be effectively used to control the peak rate of a source. However, for controlling the mean rate with- out introducing unacceptable cell loss rates, a long Observation time (high counter threshold) is required. Finally, they conclude that the leaky bucket is insensitive 29 to the burst and silence length distributions (of bursty on-ofi sources) and hence is inappropriate for controlling burst durations. Similarly, the performance of the buffered leaky bucket scheme is analyzed by Holtsinger and Perros in [33]. Using a two-state Markov Modulated Bernoulli Process (MMBP) model for the cell arrival process, they Show that the buffered leaky bucket can be configured to control the mean rate or the departure burstiness (expressed in terms of the squared coefficient of variation of the inter-departure time) but not both. However, the authors suggest that both configurations may be used together. The performance of the unbuffered leaky bucket is compared with that of the buffered version. The results suggest that the buffered scheme is more effective at controlling the burstiness Of the cell arrivals, but suffers from introducing a non-zero waiting time on the cells. Finally, the authors also show that there is a tradeoff between the arrival cell loss rate (at the buffer) and the departure burstiness. In [62], Rathgeb compares the performance of the (unbuffered) leaky bucket (LB) mechanism with several other policing mechanisms, including the jumping window (JW) mechanism, the triggered jumping window (TJW) mechanism, the exponentially weighted moving average (EWMA) mechanism and the moving win- dow (MW) mechanism. Using a combination of analysis and simulation methods, the policing schemes are evaluated in terms of violation probability for well-behaved sources, sensitivity to short-term and long-term overload, and worst-case traffic al- lowed. The LB and EWMA mechanisms are found to be the most promising ap- proaches by virtue Of their flexibility to handle short-term statistical traffic fluctua- tions. Shimokoshi [71] proposes the use of “pseudo cell buffering” to enhance the J W and MW mechanisms. In general, window-based mechanisms work by imposing a certain limit on the number of cells in a time window. The pseudo buffering mechanism allows this limit to be exceeded (up to the pseudo buffer limit) with the excess to be charged 30 in the next window. Using simulation techniques, the performance characteristics of the pseudo jumping window (PJW) and pseudo moving window (PMW) mechanisms are compared with the (unbuffered) LB, JW, and MW methods. The results Show that the LB and PJW mechanisms are the most effective at regulating the mean source rate. 3.4.2 Call Admission Control Schemes In this section, a variety of call admission control (CAC) schemes are discussed. CAC schemes typically manage bandwidth and buflers at the connection level, in order to guarantee a quality of service (QOS) to the admitted calls based on the expected cell loss rate (CLR). Hui’s monograph [35] discusses multi-level resource allocation for ATM / BISDN networks. Being a relatively early article, it has a few discrepancies with subsequent developments, including the persistent use of the word packet rather than cell, the absence of VP/VC routing (connections can have multiple routes set up), and the possibility of cells arriving out of order (which is not in line with the virtual circuit view of an ATM connection). In the proposal, traffic is to be managed at three levels: the packet or cell level, the burst level, and the call level. Blocking probabilities are to be maintained at each level. At the call level or burst level, a pilot packet is used to indicate the amount of bandwidth or change of bandwidth required by the source. This packet may also contain other information such as the estimated holding time for the bandwidth increase, but the author does not propose how such information is to be used except maybe in the computation of the call blocking probability which uses “holding time statistics.” At each level, the algorithm keeps routing and load information and makes decisions based on computations Of blocking probabilities. The computations are to be done according to the methods presented for the three levels. 31 In conclusion, the article presents simulation results with the following findings: 0 VBR traffic benefits more from statistical averaging than bursty (on-off) traffic. 0 Bursty traffic utilization is more sensitive to the tolerable blocking probability than VBR traffic utilization. In [27], Gallassi and others present a class related rule (CRR) for bandwidth assignment in ATM networks. Classes consist of homogeneous on-ofi' sources with the same peak rate, burstiness (peak rate to mean rate ratio), and mean burst length. With a predetermined buffer size (which determines the cell delay limit) and a target cell loss probability, the bandwidth requirements of each class (for different numbers of sources in the class) are determined via simulation (and stored). The CRR is then used to determine the required bandwidth of a set of classes of varying number of sources. The class related rule (CRR) from [27] is evaluated by Decina and others [20] using simulation. The performance gains obtained using the CRR is found to be sig- nificant for burstiness values greater than five. The CRR appears to be a conservative approach for the assumed traffic for a wide range of cases. The call-level blocking probability was also evaluated for CRR and was found to Offer significant gains. Suzuki and others [76] propose a bandwidth allocation strategy based on the concept of “virtual cell loss probability,” which is a quality estimation method based on a “logically buffer-less” model instead of a buffered queueing system. The method sacrifices link utilization for simplicity and robustness. It uses only the maximum and average rates of the calls (with on-ofl' sources) to estimate their virtual cell loss probability, i.e., no burst lengths or buffer size parameters are used. The connections are grouped into classes where each class has a maximum rate, average rate, and number of calls. The performance of the approach is analyzed and the authors suggest that traffic with high maximum rates (with respect to the link capacity) be allocated 32 the maximum rate, and that available capacity be separately reserved for such traffic and CBR traffic. Guérin and others [28] propose a method for computing the “equivalent capacity” of connections. This method uses a fluid-flow two-state Markov source model approx- imation. The traffic parameters are the peak rate, the source utilization (fraction of time the source is active), and the mean burst length for exponentially distributed burst and idle periods. The effective bandwidth is computed for a single source, or for a combination Of sources in which case the statistical gain is higher—i.e., the ef- fective bandwidth Of a combination is less than the sum of the effective bandwidths of each individual in the combination. The scheme uses approximations to make computations achievable in real—time. Two complementary formulas for equivalent capacity are used. A “fluid-flow approximation” is used for low numbers of sources and a “stationary approximation” is used for large numbers of sources. The authors suggest the use of standard grades of service for which the effective bandwidths have been precomputed. Extensions to handle non-exponential distributions are discussed. One major disadvantage of the approach mentioned in the article is the fact that it is not effective in those cases where the connections have long bursts or low source utilization. In [70], Saito proposes a call admission control scheme for ATM networks using output buffers. The scheme guarantees the cell loss probability and the author claims (perhaps misleadingly) that it does not have any assumptions on the cell arrival process. A buffer size is assumed that guarantees that the cell delays will remain within bounds. The traffic parameters required are the average and maximum number of cell arrivals in an interval equal to the time required to half-fill the buffer (note that the only case where this is not a restriction on the traffic process is when the interval is infinite—the smaller the interval, the more restrictive it is). The performance of the scheme was investigated and it was concluded that the use of only the maximum 33 and average number specification sacrifices some utilization loss compared to more detailed traffic specifications. 3.4.3 Burst-Level Control Schemes This section covers some proposals on handling bursty data traffic at the burst level. Already mentioned in Section 3.4.2 is Hui’s multi-layer control scheme [35] which uses pilot packets to reserve bandwidth for bursts. A burst-level approach are also briefly described by Ohnishi and others in [54] in the form of a short hold mode service. In the early 19803, researchers from GTE developed the “burst-switching” ap- proach [3] for supporting integrated voice and data traffic. The burst-switching con- cept is described in [31] as follows: “Information contained within the header of each voice or data burst establishes a path through the network. A burst consists of a header (containing routing, control information and error control bytes) , fol- lowed by the completely variable length information burst and ending with an end-of-message byte. . .. The header effectively latches up con- tiguous switch network channels for the duration of the burst.” Transmission on traditional digital carriers such as the T1 carrier is typically orga- nized into frames divided into separate channels. In burst switching, these channels are statistically multiplexed among the connections on the burst level. That is, rather than dedicating the channels to bursty connections for the duration of their lifetimes, the channels are assigned to currently active connections on demand. The scheme de- pends on Speech silence detection to take advantage of the fact that voice connections are active only 35% to 40% of the time. Both voice and data bursts are buffered in case of congestion, except that a buffered voice burst is limited to a few milliseconds worth of voice samples, after which speech clipping or freeze-out [49] is done. To 34 minimize freeze.out, voice bursts are to be given higher priority over data bursts by the resource allocation scheme. In [13], Boyer and Tranchier propose fast reservation protocols for bursty traf- fic. The article is primarily about FRP/DT (fast reservation protocol with delayed transmission) but also introduces its sibling FRP / IT (fast reservation protocol with immediate transmission). FRP/DT works best for highly bursty traffic where the transmission time for a burst is large compared to the round trip delay. The round trip delay (RTD) is required for reserving bandwidth. If the bursts are short, the effi- ciency (utilization) becomes compromised. F RP/ IT is tailored for applications with smaller bursts or which cannot tolerate the setup time (including the RTD). Their analysis shows that FRP/DT allowed for a statistical gain equal to or greater than that of ideal rate-based multiplexing which required more stringent traffic policing. The scheme uses the burst blocking probability (BBP) as a parameter analogous to the burst congestion probability (cell loss rate) in traditional call admission control schemes. FRP/DT is centralized in the sense that the user communicates with an FRP Control Unit which handles all the signaling within the network. It allows a user to change its bandwidth allocation in a stepwise fashion through negotiation such that unsuccessful negotiations do not prevent it from using its previously allocated bandwidth. F RP/ IT, which is based on ideas from [35], is only briefly described in the article. Adaptive in-call renegotiation schemes for burst-level resource allocation protocols such as F RP/ DT are proposed in [23] (for long file transfers) and [22] (for short intermittent file transfers). Turner proposes a fast buffer reservation scheme in [82]. The scheme allocates buffers to bursty virtual circuits on a burst-by-burst basis. The bursty traffic includes start and end cells to indicate the beginning and end of bursts. The arrivals of these cells and the middle cells, together with a timeout mechanism, affect a state machine associated with each bursty virtual circuit. The state machine has two 35 states representing the idle and active states of a connection. Because of the timeout mechanism (which is intended to protect against cell losses but at the same time remain tolerant to inter-cell delay jitter), the author proposes that the scheme be used only for virtual circuits with peak rates exceeding 1% to 2% of the link rate. In addition, there are computational reasons why high peak rate to link rate ratios are desirable in the scheme. The multiplexing efficiency of the scheme was analyzed and presented, albeit for relatively high blocking probabilities (0.01 and 0.001). In the analysis, the author did not address the issue of excessive overhead (i.e., high control cell to data cell ratio) for high burstiness (peak rate to mean rate ratio) or low peak rate cases. Finally, the scheme does not seem to be based on the equivalent bandwidth concept and requires the re-computation of all contention probabilities (burst loss probabilities) for all circuits whenever a new one is added. Crosby investigates the approach of in-call renegotiation in [19]. The most im- portant result from the study is that renegotiation offers the most benefit in those cases where the ratio of the peak rate to link rate is high. When the statistical multiplexing level is high (e.g., high speed backbone links), it Offers little gain. The analysis uses two types of source models: the familiar on-ofl source and a more so- phisticated source with a miniprocess to characterize an on state. The performance of the following approaches are compared: 0 static allocation (no renegotiation but with a bound on the burst blocking prob- ability) 0 simple renegotiation (renegotiation is performed before every state change), and o optimal renegotiation (the algorithm is unspecified; the performance is deter- mined using upper/ lower bounds analysis). With renegotiation, the traffic load is controlled with a parameter Pb, the probability of renegotiation failure. The results Show that the Simple renegotiation scheme offers 36 increased utilization when the ratio of the peak rate to the link rate is high. However, the results also show that the simple scheme is far from optimal. It should be pointed out that in the study, the author only considered the case of one ATM multiplexer and did not take into account a possibly large round-trip delay (RTD); more realistically, one would need to take the RTD from source to destination into account, especially in evaluating the utilization. 3.4.4 Special Schemes for Best-Efl'ort 'ITamc In this section, special schemes intended for best-effort or available bit rate (ABR) ser- vice are described. These schemes do not aim to provide a guaranteed Q08 in terms of throughput or delay.2 They are primarily intended for connectionless communication where the characteristics of the traffic generation cannot be adequately specified or where throughput and delay guarantees are just not necessary. These schemes, how- ever, aim to distribute the available unreserved bandwidth and buffers fairly among the ABR connections (for some appropriate definition of “fairness”). Moreover, they aim to totally prevent cell losses due to buffer overflow. These schemes do not de— pend on call blocking to control congestion.3 However, they have the property that the more ABR (and bandwidth-guaranteed) connections there are, the poorer the service each ABR connection will receive. Furthermore, they would generally require large buffers to avoid cell losses while keeping the bandwidth utilization high in the case of high-speed long-distance connections. There are two main camps in the ABR debate: the rate-based and the credit-based. The rate-based flow control approach for ABR service is surveyed in [9]. Con- ceptually, it is a feedback-based source rate throttling scheme where the destination 2However, some proposals such as EPRCA described in this section include support for a mini- mum cell rate, in which case that portion of the service is guaranteed. 3Naturally, a scheme supporting non-zero minimum cell rates will have to resort to call blocking if the minimum rate cannot be guaranteed. 37 and intermediate nodes participate in congestion detection. The culmination of the evolution of this concept is called Enhanced Proportional Rate Control Algorithm (EPRCA). EPRCA has three main components. 1. The first component is a source rate control mechanism that can raise or lower its sending rate based on feedback from backward Resource Management (RM) cells. A source periodically sends RM cells in the forward direction and pe— riodically decreases its sending rate (by default). It increases its sending rate whenever it receives a backward RM cell with a negative Congestion Indication (CI) flag. Backward RM cells also contain an Explicit Rate (ER) field. A source receiving an ER value less than its current sending rate lowers it rate to the ER value. 2. The second component consists of congestion detection and congestion indica- tion mechanisms at the intermediate switches. Basic congestion detection may be done by monitoring the instantaneous queue lengths in the switches. Con- gestion indication is done by the switches either by setting the Explicit Forward Congestion Indication (EFCI) flag in the forward data cells or by setting the CI flags in the forward RM cells. Switches may also optionally send backward RM cells to the source. 3. The third component is a destination mechanism that returns backward RM cells to the source to inform it Of any congestion detected in the forward direc- tion. The destination bounces back forward RM cells to the source and may Optionally set the ER or C1 fields based on congestion at the destination or on EFCI information from the forward data cells. The proposal also includes provisions for breaking up long end-tO-end feedback loops into shorter segments in order to improve performance. In this case, the source and 38 destination roles are also performed by Virtual Source and Destination nodes at the endpoints of each segment. Overall, the main attraction of the rate-based approach is its relative simplicity of implementation. At the minimum, participating switches only have to be able to raise the EFCI flags in data cells when appropriate. Furthermore, it does not require hop-by-hop per-VC queueing within the switches. The credit-based flow control approach for ABR service is described in [45]. It is based on the Flow-Controlled Virtual Channels proposal presented in [44]. The idea is to perform credit-based flow control on a link-by-link, per-VC basis. In general, credit-based flow control works by requiring a sending entity to get “credits” from the receiving entity before it can start forwarding data cells. The receiving entity gives credits to the sending entity based on buffer availability on the receiving end. This assures that the sent data cells will not be lost because of buffer unavailability on receipt. The receiving entity sends credit cells upstream at a rate proportional to the rate it can empty the buffers used by the virtual channel. Per-VC queueing together with round-robin scheduling is used to achieve fairness among the ABR connections. Fairness, high bandwidth utilization, and robustness under extreme conditions are the main strengths of the credit-based approach. Its main disadvantage is high switch complexity. This includes the per-VG queueing and credit management requirements of the scheme. Furthermore, the approach would generally require large buffers to keep the links highly utilized in the case of long-distance high-speed links. Finally, it also requires considerable bandwidth overhead for the credit cells. In addition to the rate-based and credit-based approaches, there are hybrid schemes that have been proposed. These schemes aim to combine the strengths of the two main approaches in an integrated system. In [61], the authors contend that the rate-based approach would be appropriate for a (public) WAN, while the credit-based approach would be appropriate for LAN 5. A WAN involves large propa- gation delays and very large numbers of connections (VCS). These conditions amplify 39 the disadvantages of the credit—based approach. However, the simplicity of the rate- based approach based on EFCI marking becomes specially attractive in this case. Furthermore, the rate-based approach appears to be more conducive to proper billing of ABR services. In contrast, a LAN would greatly benefit from the advantages of the credit-based approach with less penalties. The ABR service essentially emulates the behavior of traditional LAN technologies, and the credit-based scheme Offers high utilization, no congestion-based loss, and robustness in the presence of highly unpre- dictable traffic. Three levels Of integration are described by the authors: The basic scheme is to strictly use the credit-based approach in the LAN and the rate-based approach in the WAN. A more general scheme uses the rate—based approach as the default, but provides the credit-based approach as an option (when supported by all switches in the path). The proposed scheme is to fully integrate the two approaches so that rate-based and credit-based switches may Operate with each other. It in- volves using the RM cells to convey both rate information and credit information. Essentially, in each segment of adjacent rate-based switches, the (virtual) source is limited in rate by the rate-based switches and volume by the credit-based (virtual) destination. More details can be found in [61]. Another hybrid scheme is described in [36]. It combines the rate-based approach with link-by-link back-pressure control. The purpose of the link—by—link back-pressure control is to avoid cell losses due to sudden increases in queue length that may happen in spite of the efforts of the (not-tOO-conservative) rate—based control implementation. In the scheme, the rate—based congestion indication is performed whenever the queue length reaches a certain threshold. Back—pressure is applied whenever the queue length reaches an even higher threshold to give time for the rate adjustments to take effect. 40 3.4.5 Special Schemes for Delay-Sensitive 'Ifamc There are also specialized traffic control schemes intended for delay-sensitive (typ- ically audio and video) traffic. Since this dissertation is primarily concerned with bursty data traffic, only a few of these specialized schemes will be mentioned in pass- ing. Congestion control based on the selective discarding of packets or cells is presented and analyzed in [52, 86]. These techniques take advantage of the fact that the per- ceived performance of audio and video transmission is more sensitive to the integrity of some portions Of the encoded data than others. This allows the less important portions to be sent in low-priority cells for preferential discarding during congestion. The priority is to be indicated in the CLP bit of the cell header. A scheme for smoothing VBR video traffic is presented and analyzed by Ott and others in [57]. The scheme dynamically adjusts the rate of sending cells from an input buffer in order to keep the rate as close to constant as possible but without violating the delay constraints. The technique is based on using the recent behavior of the traffic stream to forecast the future traffic behavior. A variable bandwidth allocation scheme for VBR video is presented and analyzed by Pancha and Zarki in [59]. The scheme uses a simple prediction algorithm to estimate the bandwidth requirements of a video encoder. To further improve the performance, the scheme also takes advantage of layered coding where the encoded output is separated into different streams according to their relative importance. These streams are given different priorities with different loss characteristics. Chapter 4 Framework for Burst-Level Traffic Control Recall that in ATM networks, data are transferred in a connection-oriented mode with a guaranteed quality of service (QOS). The connections are statistically multiplexed in order to increase the number of concurrent connections and to keep the bandwidth utilization high. This approach is ideal for connections that carry traffic at a near constant rate. However, to accommodate highly bursty traffic, either large buffers have to be provided within the switches, or the bandwidth utilization has to be kept low. Otherwise traffic congestion due to bursts will cause frequent buffer overflows. Avoiding congestion at high utilization in the presence of bursty traffic is a major concern in the design of high-speed networks. The most conservative approach to congestion avoidance is to allocate bandwidth deterministically (rather than statistically) to each connection. This means that each connection will hold bandwidth equal to its peak rate (to be policed at the source) throughout its lifetime. This approach does not offer any statistical multiplexing gain and will result in low bandwidth utilizations in the presence of highly bursty traffic. Another approach to congestion avoidance is to have a call admission control 41 42 (CAC) procedure that will limit the number of concurrent connections so that the expected congestion is kept within acceptable limits. Based on parameters included in the connection establishment request, the CAC procedure determines if it has enough buffer space and bandwidth available to provide a guaranteed QOS (e.g., maximum cell loss rate) to the new connection. Depending on the approach used, the CAC procedure may also have to verify that the service guarantees to all previously established connections will continue to be met. Several proposals for allocating bandwidth and/or buffers to bursty connections have been proposed. Many of these approaches are based on guaranteeing a maximum cell loss rate (CLR) to the connections (see Section 3.4.2 above). These approaches typically treat cells indiscriminately without regard to their belonging to a burst. For many applications (e.g., file transfer), this is undesirable because losing a single cell in a burst may make the rest of the burst useless. A more appropriate approach is to manage the bursty traffic at the burst level and try to maintain a maximum expected blocking or loss rate of bursts. This has been proposed by Hui [35] and subsequently by Boyer and Tranchier [13] and Turner [82] (see Section 3.4.3 above). In this chapter, a framework for supporting highly bursty connections in ATM networks is presented. The proposal includes mechanisms designed to facilitate traffic control at the burst level. The framework is designed to support a wide variety of bursty connections, including connections that may have extremely long bursts and/or extremely long periods of inactivity. It also supports connections that are continuously active in a variable bursty fashion. Finally, it is capable of handling all of the above connection types on a burst-by-burst basis even when they are bundled into virtual paths. This last feature is an important capability that is taken for granted or overlooked by the other burst-level approaches that have been proposed. Section 4.1 presents the types of burst-oriented connections that the framework is designed to support. The connection establishment phase is briefly covered in 43 (a) Thin logical connection (TLC) for type- 1 source. (b) Fat logical connection (FLC) for type-2 source. W\VM‘ 7% MW V Ask ,- , . ‘3}; .t A\\—- NV,“- . “N30 Figure 4.1: Thin and fat logical connections. The shaded areas indicate the instanta- neous bit rates of the sources and the heavy lines indicate the bandwidth dynamically allocated to the connections. Section 4.2. The basic mechanisms for proper burst-level handling of the traffic are described in Section 4.3 and Section 4.4. Advanced mechanisms including the methods of bundling these bursty connections into virtual paths are given in Section 4.5 and Section 4.6. Section 4.7 gives a summary of the chapter. 4.1 Burst-Oriented Services We are interested in dealing with two types of bursty traffic sources, which we will call type-1 and type-2. A type-1 source, also known as an on-ofl' source, alternates between periods of burst and silence. During bursts, it generates cells at a rate not exceeding its peak rate. Between bursts, it does not generate cells. A type-2 source is a variable rate source that alternates between periods of high activity and low activity. It generates cells at a nominal non-zero rate between its high activity (burst) periods, which is unlike a type-1 source that is totally silent. The nominal rate is allowed to change but only in a time scale larger than the burst arrival time scale. To support these bursty sources, two kinds of burst-oriented connections are pro- posed; these are illustrated in Figure 4.1. These connections, called thin and fat logical connections, are essentially virtual channel connections with dynamically allocated 44 bandwidth and buffers. A thin logical connection (TLC) is a burst-oriented virtual channel connection that supports wait-free transmission of bursts with a guaranteed burst blocking probability (BBP). TLCs are intended for type-1 sources. A fat logical connection (FLC) is a burst-oriented virtual channel connection for type-2 sources (cf. [13]). It guarantees CBR-qualityl service on the nominal rate portion of its traf- fic, and it allows the wait-free transmission of bursts above the nominal rate with a guaranteed BBP. Furthermore, through renegotiation, it allows the nominal rate to be increased or decreased. In order for the network to handle the cell traffic from these connections on a burst-by-burst basis, a way has to be provided of recognizing the beginning and end of bursts. Each burst submitted by the source will be prepended with a reservation request cell (cf. [13, 35, 82]), which we will call the head cell and postpended with a reservation release cell, which we will call the tail cell. After an appropriate lead time, the data cells may be transmitted behind the head cell without waiting for acknowledgment. We will call this mode of transmitting the burst eager transmission, as Opposed to patient transmission in which transmission does not begin until the reservations along the whole path are acknowledged. All the data is sent through a TLC on a burst-by-burst basis. At the network interface (i.e., the ATM adaptation layer), the head and tail cells are added to each burst. Inside the network, the bursts are handled one by one. When a burst arrives at each switch, the connection is activated if bandwidth can be allocated to the burst at that time. Conversely, when a burst leaves, the connection is deactivated and the bandwidth is released. In comparison, data is sent through an FLC in two ways. All bursts that can be sent without exceeding the established nominal rate are sent continuously. No 1delay and loss rate equal to that of a continuous bit rate (CBR) connection 45 burst-level handling is required for these bursts in the network, so head and tail cells are not required for them. However, bursts that will cause the aggregate rate to exceed the nominal rate have to be sent on a burst-by-burst basis, complete with head and tail cells. The additional bandwidth required is allocated when a head cell arrives and is deallocated when the tail cell departs. In summary, these logical connections work as follows. First, they have to be established just like any virtual channel (or virtual path) connection before any data can be sent. This involves routing and virtual channel assignment. A TLC is not al- located any bandwidth or buffers initially. In contrast, an FLC is allocated resources sufficient for its nominal rate portion. The buffers and bandwidth reserved are equiv- alent to those required by a CBR connection of the same rate. After connection establishment, the connections are ready to carry bursty traffic as described above. Furthermore, an FLC is allowed to increase or decrease its nominal rate portion via renegotiation with the network. Finally, note that it is not necessarily assumed that the bit rate of each source has a perfect square wave-form pattern. Rather, it is assumed that the rate of each connection is limited to the declared peak rate2 at the source. Within the network, the fluctuations of the instantaneous bit rate due to jitter are absorbed by buffers appropriate for regular CBR connections subject to similar jitter. 4.2 Connection Establishment Data transmission in ATM networks is connection-oriented. This means that a virtual connection has to be established from the source to destination before any data can be transferred. Connection establishment involves the following important steps, among others: 2nominal peak and burst peak rates in the case of a type-2 source 46 1. finding a route from the source to destination and assigning a virtual path identifier (VPI) and a virtual channel identifier (VCI) to the connection at each switch in the connection’s path 2. allocating bandwidth and buffers to the connection to satisfy its quality of service requirements, and 3. setting up the routing table of the dedicated cell Switching. hardware at each node. Several fast approaches for step 1 are discussed in Section 3.2 above. Various proposals for step 2 are given in Sections 3.4.2 and 3.4.3. Finally, step 3 is hardware dependent but will be addressed in Section 4.5 below. This section will briefly describe a simple proposal for step 2 that will be covered in detail in Chapter 5 below. Chapters 6 and 7 cover additional proposals for step 2. A simple burst-oriented bandwidth reservation scheme for bursty connections has been developed. The proposed scheme groups TLCS into classes according to their peak rates (but the mean rates may vary among the connections in a class). For each class at each link, the scheme allocates a number of virtual servers so that the burst blocking probability (BBP) for each connection in the class is within the guaranteed value. A virtual server is simply a portion of the link bandwidth equal to the peak rate of the connections in the class. Assuming that the burst arrivals are Poisson, the set of virtual servers can be modeled as an M / G/ k/ k queueing system. Given the appropriate traffic parameters from each connection in the class, an upper bound for the BBP for the class can be obtained using Erlang’s loss formula (given in Section 5.1 below). Therefore, at connection establishment, each switch in the path will attempt to incorporate the new connection into an appropriate class of preexisting connec- tions. The expected BBP after adding the new connection is determined to see if 47 an additional virtual server is required for the class to maintain the BBP at its ac- ceptable level. If an additional virtual server is required, then additional bandwidth is allocated to the class (not the connection—bandwidth is reserved for the connec- tions on a burst-by—burst basis) if it is available; otherwise the connection request is rejected. More details on the bookkeeping involved can be found in Chapter 5 below. After accepting a new connection, a routing table entry must be provided for it at each switch. To use the burst-oriented approach, it is required that the switching hardware provide a method for turning a connection on and off as bursts come and go. This is covered in Section 4.3 below. 4.3 Eager Transmission Detection and Handling Inside the network, the switches need a way of detecting eagerly transmitted bursts in order to allocate bandwidth for the bursts as they arrive at the switch. One approach is to identify the head and tail cells with a special payload type identifier (PTI). The switches intercept and process the cells with the special PTI. This way, the head and tail cells can be sent in-band with the same cell headers as the data cells. Another approach is to send the head and tail cells on a separate signaling band, i.e., with a different virtual channel address from the data cells. A critical problem with this approach is the timing—it has to be ensured that the head cell always precedes the data cells by a sufficient margin to give the switches enough time to activate the connection before the data cells arrive. In addition, the tail cell must not be allowed to overtake any of the data cells. A useful feature for reducing the effects of the head cell processing delays is to use a limited form of multicasting similar to the selective copy technique described in [17]. This will allow the head cell to be forwarded to the next node downstream at the same time it is automatically delivered to the switch control processor (SCP) at a 48 pi = propagation delay h, = head cell processing delay for connection activation q, = queueing delay Figure 4.2: Required head cell lead times. Without multicasting, the time it will take for the head cell to reserve the bandwidth at the last link, after it is transmitted at source S, is given by Z?=l(p,- + h.-) plus the queueing delays. With multicasting, this reduces to 2;, p.- plus the queueing delays. In contrast, the first data cell may take only 213;, p.- to reach the last switch. Note that if each h.- is less than the cell switching time, then the queueing delays are not a factor in the multicasting case. node. General multicasting may also be used to simultaneously forward the head cells downstream and to the SCP. However, the head cells will have to be transmitted on a separate virtual channel/ path from the data cells. Without any of the multicasting features above, the head cell processing delays at the nodes will accumulate and the lead time for the head cells has to be adjusted accordingly. This is illustrated in Figure 4.2. A required feature to allow the safe introduction of eager transmission into the network is a method of dropping the cells of unadmitted bursts. These unadmitted bursts may be either bursts for which no bandwidth is available or bursts whose head cells are corrupted or lost. These bursts must be prevented from consuming bandwidth and buffers that have not been reserved for them. One straightforward approach is to use an active/inactive bit flag in the switch routing table.3 ‘For an activated TLC (i.e., with bandwidth currently reserved), the bit is set to ‘1’, otherwise it is set to ‘0’. The switch checks this flag whenever it consults the routing table to forward a cell. However, this approach is not sufficient for logical connections routed 3The Fore Systems ASX-100 switch [25] has a “valid route” bit in its routing table that could serve this purpose. 49 by virtual path. A solution for virtual path connections is presented in Section 4.5. 4.4 In-Band Burst Control Cells In this section, we discuss in more detail the function and handling of the in—band burst control cells used by this burst-oriented framework. There are two basic types of such cells, the head cell and the tail cell. The head cell is prepended at the beginning of each burst to activate the connec- tion, i.e., to turn on the routing table entry of the connection if bandwidth is available for it at that time. The head cell contains the following: 1. special PTI to identify the cell as a burst control cell 2. logical connection identifier (i.e., incoming port, VPI, and VCI) 3. control cell type (head cell) 4. connection type (e.g., thin logical connection) 5. burst rate (connection class), and 6. estimated burst duration. The first two items above are located in the cell header, while the rest are stored in the cell payload. The third item is stored in the cell payload so as not to use up precious bits in the cell header. The fourth and fifth items are not necessary but may speed up table searches. The last item, which is optional, is useful for setting appropriate timeout durations. When a switch identifies a head cell, it checks to see if the associated connection is activated. If not, then it checks if bandwidth is available for the burst.“ If it is, 4In the case of the reservation scheme presented in Chapter 5 below, this is a matter of determining if the class to which the connection belongs has a currently unused virtual server. 50 then the switch activates the connection and reserves the bandwidth for it. Once the connection is activated, the data cells following the head cell will be switched properly. If the head cell is lost or corrupted, or if bandwidth is not available for an arriving burst, then the connection will not be activated and the data cells will be dropped and will not use up resources reserved for other connections. The tail cell is appended to the end of each burst to deactivate the connection. It contains the same information as the head cell except for a different control cell type. When a switch identifies a tail cell, it checks to see if the connection is activated. It it is, then the switch deactivates it and deallocates the bandwidth temporarily allocated to it. If the tail cell is lost or corrupted, then the connection will not be deactivated and will continue to hold resources it is not using, possibly causing bursts from other connections to block unnecessarily. To avoid this, a timeout mechanism has to be used to deactivate connections automatically. The timer for a connection is set when the connection is activated and is reset when the connection is deactivated. The duration is set according to information in the head cell payload or according to connection or switch parameters. In any case, an additional burst control cell may be used to extend the timer in the middle of a burst. We will call that cell a refresh cell. A refresh cell is similar to a head cell except that it does not activate a deactivated connection. It only adjusts the timer associated with the connection. Therefore, if the head cell is not successful in acquiring bandwidth for a burst, then the whole burst up to the tail cell will be dropped even though the burst contains refresh cells. A refresh cell contains the same information as the head cell except for a different control cell type. Finally, if the timer for a connection fires, then the switch assumes that a tail cell has been lost. It then deactivates the connection and deallocates its temporary bandwidth reservation. timer" "".- ., "r"?! 51 4.5 Thin Logical Connections in a Virtual‘Path The handling of a TLC at switches where it is routed by virtual channel, i.e., by looking at its VPI and VCI header fields, is straightforward since each connection will have a separate routing table entry. The mechanisms described in Section 4.3 and Section 4.4 above are sufficient. This section deals with how thin logical connections are to be handled when they are bundled in a nonterminating virtual path in a switch. A nonterminating virtual path is a virtual path (VP) in which the cells are routed by VPI alone—the VCIS are ignored in the switching process. The main problem that has to be solved is that of properly dropping cells from an eager burst that is rejected or whose head cell is lost. Since the cells in the VP are routed via the VPI alone (i.e., there is only one routing table entry for all of them) and since each virtual channel (VC) in the VP has to be allowed to transmit bursts independently of the other VCS in the VP, the switch needs a way to distinguish between cells with reserved bandwidth and cells without reserved bandwidth in the VP. It is proposed that, in addition to the routing table, another table be provided to contain bitmaps of the activated VCS in the VP8. Virtual paths with thin logical connections will have an entry in this table. The entry is a bit vector indicating which VCS in the VP are currently activated. For convenience, the first entry in this table may be a special bit vector containing all ‘1’s that will be used for all virtual paths not carrying any burst-oriented connections. The other bit vectors will be allocated to the VPS containing thin logical connections, with one bit for each VCI.5 When the switch receives a cell belonging to a nonterminal VP, it will have to check the bit vector for the VP to see if the VC is activated. An activated VC will 5To support all possible VCI values, a bit vector length of 216 is necessary. However, it is straight forward to allow only a subrange of possible burst-oriented VCIS to decrease the length of the bit vectors. -—'-4n a "'3 fi' ‘ _ V .- m ' E: 52 have a ‘1’ in its position and a deactivated VC will have a ‘0’. This is illustrated in Figure 4.3. The size of the VC bitmap table will be the product of the maximum number of VCS per burst-oriented VP and the maximum number of such VPS to be concurrently supported. In summary, it is proposed that the routing table that a switch uses to determine the output VPI label of cells be supplemented with virtual channel bitmap table. This table of bit vectors is used to distinguish between activated and deactivated bursty virtual channels in a virtual path. Cells for activated virtual channels are" to be forwarded and cells for deactivated virtual channels are to be dropped. 4.6 Fat Logical Connections In this section, a way of supporting fat logical connections using the mechanisms already described above is presented. First, we assume that we have at our disposal a type Of continuous bit rate (CBR) connection that supports renegotiation. This type of connection has bandwidth deterministically allocated to it that it can fully utilize during its lifetime. Furthermore, during its lifetime, the connection is allowed to negotiate with the network long-term stepwise changes in its bandwidth allocation. The renegotiation process is patient in the sense that the connection is required to receive an acknowledgment from the network before it can start using any additional bandwidth requested. The FRP/DT proposal in [13] is an example of a scheme that supports this type of connection. We reduce the problem of handling fat logical connections with eager bursts to the problem of handling thin logical connections with eager bursts by using tandem logical connections. We use two virtual channel connections on the same virtual path to carry the traffic for a single fat logical connection. One of the virtual channels is a CBR connection with renegotiation as described above, and the other is a TLC. 53 bit vector. VP Table—indexed by input port and VPI Step 1: Check if the virtual channel is activated. From the VP table, get the VCBT index to select a bit vector from the VC bitmap table. Use the input VCI to select the bit :I: from the input [port [ VPI] RT base mode bit VCBT index VC Bitmap Table—indexed by VCBT index v 000... ...OOO... ...000 111... ...111... ...111 from VP table . 3 WCBT index ] bbb. .. ..bxb. .. . .bbb [VCI] indicates virtual path switching. Routing Table—indexed as indicated Step 2: If x = 1 from above, then the virtual channel is activated and has bandwidth reserved for it. Therefore, get the output address from the routing table. Note that a mode bit of zero from VP table and input VCI [RTbase+modebithCI [ output port output VPI output VCI Figure 4.3: A cell routing algorithm that is safe for eager transmission in VPS. The VP table and routing table, except for the VCBT index, are modeled after the Fore Systems ASX-100 switch architecture. However, any ATM switch will need a table to contain the output VPI for each established VP. That same table may be modified to contain the VCBT index. In fact, the VCBT index could be stored in the output VCI field, which is not used for nonterminating VPS. 54 Putting the two connections in the same virtual path means that at each switch along the way, the two connections will have the same incoming port and VPI and same outgoing port and VPI. However, the two components will have different VCIS. We will assume that cells in a virtual path, even if they belong to different virtual channels, are not allowed to overtake each other.6 This way, there will be no unwanted overtaking between the cells sent with the TLC VCI and the cells sent with the CBR VCI. If it is necessary for the burst to persist (i.e., the connection requires long-term bandwidth increase rather than a short temporary increase for a short burst), then the CBR renegotiation request and the eager burst can be sent at the same time. This way, when the renegotiation is approved, all the subsequent traffic can be Sent on the upgraded CBR component. This subsequent traffic could include the still-tO-be-sent portion of the eager burst. After transferring the latter portion of an eager burst to the CBR component, then the TLC component can be used for a new burst. In general, when a fat logical connection initiates a burst with a head cell, then it has at its disposal an amount of bandwidth equal to the sum of the CBR bandwidth and the TLC peak rate. Also, it has two VCIS to work with. It can send data cells using the two VCIS with the following restrictions: 1. The total cell traffic sent using the two VCIS must not exceed the total band- width of the two components. 2. The cell traffic sent using the CBR VCI must not exceed the CBR bandwidth because the head cell for the TLC portion might be rejected. 3. The cell traffic sent using the TLC VCI can be more than the TLC peak rate but less than the total bandwidth Of the two components. The TLC VCI is used with the understanding that if the head cell is rejected, then all the cells 6Alternatively, we assume that the ATM network supports a special type of virtual path with the explicit feature that all cells in that type of virtual path are kept in order. 55 sent with that VCI will be dropped even though a portion Of the burst may be using bandwidth from the CBR portion. Various approaches for splitting the F LC data cells among the two components are possible. Some of these are illustrated in Figure 4.4. Finally, if thin logical connections with eager bursts can be bundled in virtual paths, then fat logical connections with eager bursts, implemented as described above, can be bundled in virtual paths as well. 4.7 Summary A traffic control framework for ATM networks has been presented. The framework is designed to handle bursty data traffic in a burst-by-burst fashion. That is, a burst Of logically related data cells is treated as a unit by the protocol framework. Bandwidth is allocated and deallocated to bursts as they come and go by dynamic activation and deactivation of virtual channels at the individual switches in the connection path. Mechanisms are provided to allow sources to transmit bursts without requiring that bandwidth be reserved throughout the whole connection path. Two kinds Of bursty connections are supported: one with zero bandwidth between bursts and another with non-zero bandwidth between bursts. Mechanisms are provided to allow such connections to be bundled into virtual paths. 56 (a) Total bandwidth used for TLC VCI. TLC bandwidth 3 l - «VCI \5\'-- '-\~: '- 4 ,U'. .113 (b) Same as (a) but for train of bursts. Figure 4.4: Bursts in tandem logical connections. The total bandwidth consists of TLC and CBR components. Shaded bursts are sent eagerly using the TLC VCI; these may be dropped by the network. Unshaded bursts are sent using the CBR VCI, which has bandwidth preallocated to it. Note that the eager bursts in (b), (d), and (e) are correlated. This phenomenon has to be taken into account in the traffic description. Chapter 5 A Simple CAC Scheme Based on the Erlang LOSS System Model Recall that in ATM networks, connections that carry bursty traffic are statistically multiplexed in order to increase the number of concurrent connections and the band- width utilization. In order to prevent the deterioration of the network performance, a call admission control (CAC) strategy limits the number of concurrent connections so that a specific quality of service is guaranteed. For each new connection request, the CAC strategy determines if it can accept the connection and continue to meet the service guarantees. Research has been conducted on the problem of determining the bandwidth and/ or buffer requirements of bursty connections for this purpose. Many of the proposed approaches are based on using the cell loss rate (CLR) as the measure of the quality of service (see Section 3.4.2 above). A new approach for allocating bandwidth to highly bursty connections in ATM networks is presented in this chapter. Following the lead Of Hui [35] (see Section 3.4.3 above), it is proposed that bandwidth be requested and allocated for bursts on a burst- by-burst basis. Each burst is to be preceded by a bandwidth reservation request and then transmitted. If the requested bandwidth is available, then the burst is forwarded; 57 qm—‘“' 1 58 otherwise the burst is dropped. The new scheme is based on modeling the burst-by- burst traffic control system as an M / G / k / k queueing system. On this model, Erlang’s loss formula1 [68, 78] is applied to guarantee a maximum burst blocking probability (BBP). This approach exhibits several desirable characteristics. First, the determination of the availability of bandwidth for a new connection can be performed within a few microseconds per link. This is faster than all other published proposals known to the author. In addition, the determination of whether a particular burst can be accommodated is straightforward, such that its evaluation can be implemented easily in hardware. Furthermore, the model does not impose a constraint on the burst duration distribution other than its average length must be specified. The scheme is burst-oriented in the sense that bursts are treated as units. Bursts will be delivered or dropped completely—the only isolated cell loss that will occur will be the same as that experienced by a generic continuous bit rate (CBR) connection. Moreover, bursts are not buffered as such—the only cell buffers required are those that are allocated to any CBR connection (to handle jitter). Finally, the burst-oriented approach allows the application of traflic policing (UPC), priorities, and preemption at the burst level. Performance studies to compare the new approach with the other proposals cited above have been performed. From the results, it appears that the new approach achieves comparable maximum utilization levels of within a few percent over a wide range of input traffic parameters. The results are presented in Section 5.5. Section 5.1 describes the proposed service model for the connection types described in Section 4.1 above. Section 5.2 derives the basic BBP computation method for a set of homogeneous sources from Erlang’s loss formula. The method is then extended 1The formula, given in Section 5.1 below, is widely known to apply to the M /M / k/ 1: system. However, it has also been shown to apply to the more general M / G/ k / 1: system. See [78] for more information. 59 “2": i=2 1, ‘ A3b > b=(z"ibi)n‘ ® 1.51:5 V - Figure 5.1: M / G/ k/ k modeling of a class of connections. All S sources at the left have the same peak rate, which is the bandwidth of each of the 1: servers at the right, where k S S. The burst arrivals from the sources are Poisson processes with rates A,- and the burst durations are given by the means b5. Arriving bursts for which no server is available are dropped in the bit bucket. to the computation of traffic bandwidth requirements of a class of connections in Section 5.3. The multi-class generalization of the Erlang loss system is discussed briefly in Section 5.4. The performance of the single-class system, including discussion of the impact of non-Poisson burst arrivals, is covered in Section 5.5. Finally, a summary for the chapter is given in Section 5.6. 5.1 The M/G/k/k Service Model As mentioned briefly in Section 4.2 above, the model that is being considered for the burst-by—burst service scheme is the M / G / k/ k queueing model that is illustrated in Figure 5.1. Recall that in the notation M / G / k/ k, M indicates Markovian interarrival intervals (or a Poisson arrival process) and G indicates a general service time distri- bution. The constant k is the number of identical servers which is also the maximum 60 capacity of the bufferless system. The advantages of the model are: 0 It has no restrictive assumption on the service time distribution. In particular, the burst durations do not have to be exponentially distributed as is the case in [13, 20, 27, 2s, 40]. o It has been shown that for this system, for any service time distribution, the probability that i servers are being used is given by Erlang ’3 loss formula [78] (AE[X])‘/i! P" = ’= comm/j! j: where A is the arrival rate and E [X] is the mean service (or holding) time. The limitations of using this model are: 0 It requires the grouping of connections into classes to achieve utilization gains. However, the peak rate is the only common characteristic required for a class. In this scheme, connections in the same class are allowed to have different mean rates. This is in contrast to the other proposals that require connections to have the same peak rate and mean rate to be combined into a class [13, 20, 27, 40, 76]. o The burst arrivals are assumed to be Poisson. Real traffic may have undesirably correlated arrivals. However, the effects of these non—Poisson behavior may be minimized by appropriate traflric shaping at the burst level. Note that the Poisson assumption is also made in many of the previous proposals [13, 20, 27, 28, 35, 40]. Also, though the proposals of Saito [70] and Suzuki et al. [76] do not use this assumption, they are based on guaranteeing the CLR. Furthermore, Saito’s proposal has a restriction on the maximum and average rates over a given time interval. Finally, the proposal of Turner [82] does not require the Poisson assumption; however, its computational requirements are significantly greater 61 and (for other reasons) is only prOposed for traffic whose peak rates are at least an order of 1% to 2% of the link rate. 5.2 Basic BBP Computation Method Recall that a burst is a sequence of cells sent consecutively at a certain peak rate. The head cell at the beginning of a burst is treated by each node as a reservation request for that burst. If the node has the requested bandwidth, then the whole burst is switched to the next node via appropriate manipulation of the switch routing table. Otherwise, the burst is effectively dropped. The probability that the latter happens is called the burst blocking probability (BBP). Using the M / G/k/k model, we first show how the BBP is determined for 50 identical sources. Let the arrival rate from each source be A0, the mean burst duration be 6, and the peak rate be R. Then, the aggregate Poisson arrival rate is given by A = .9vo (by property of Poisson processes) and the aggregate mean service time is simply I) (ignoring overhead). Now, conceptually divide the available capacity C into discrete virtual servers of capacity equal to the peak rate R, giving a number of discrete servers equal to k = [C/RJ. 'The probability that the system is full, given by Erlang’s loss formula, is _ (Ab)"/k! " ‘ saw/2'! This is an upper bound of the BBP. A slightly tighter bound for the BBP can be (5.1) obtained by taking into account the fact that a burst from a source does not arrive while that source still has a burst in the system.2 That is, a burst competes only with the bursts from other sources. Therefore, for any particular source, its BBP is equal to the probability that the system is filled with bursts from the other sources. This 2Recall that the bursts are not queued. 62 can be obtained by replacing A with :\ = (50 “-1)/\o in Equation 5.1 above. Note that for large 50 (or small R), this will make little difference. This technique of combining the Poisson arrival processes can be extended to a set of non-identical sources as long as the sources have the same peak rate. Therefore, it is proposed that the bursty sources be classified according to their peak rates. One advantage of doing this is that the peak rate is relatively easy to enforce at the entrance to the network (see Section 3.4.1 above). The peak rates supported can be optimized for widely used applications. For example, 64 Kbps can support voice, low-speed data transfers, and fax; 2 Mbps can support color image and high-speed data transfers; and 10 Mbps can support video applications (cf. [20, 27]). 5.3 Bandwidth Requirement of a Class of Sources The total capacity available for bursty connections is divided among the different classes dynamically as connections are set up and torn down. This section shows how the BBP computation method from the previous section is extended to compute the bandwidth required by a class of bursty connections. Let 59 indicate the number of bursty sources in class q of peak rate R9. Let the burst arrival rates for the individual sources in the class be A1, A2, . . . , A59 and their corresponding mean burst durations be b1,b2, . . . ,bsq.3 For i = 1,2, . . . , Sq, let p, = A;(b,- + h), where h is the burstwise overhead (e.g., for bandwidth reservation and release). The aggregate Poisson burst arrival rate is therefore given by A=ZA,- 3The following relationship holds for bursty sources B.- = l/p; = l/(A;b.-) where p,- is the nor- malized load (fraction of time at peak rate) and B.- is the burstiness (peak rate to mean rate ratio) of source 3'. Therefore for a source i, instead of specifying both A.- and b5, either one of them in combination with one of p.- and B,- could specify the traffic. 63 and the aggregate mean burst serving time by 1 Sq Let Then, to determine the required number of servers for the class, find the minimum 1: such that Pk S BBP where pk/k! [9 = —____ 5.2 " mow/2°! ( ) The required bandwidth is therefore qu. Note that Equation 5.2 above overestimates the blocking probability for a single source as already explained in Section 5.1. To be more precise, we have to compute the maximum probability that bursts from only (5'9 — 1) of the sources occupy all 1:: servers. Note that this will be experienced by a source j with p, = pm," = min{p1, p2, . . . ,psq}. Therefore, we can simply replace p with p = (p - pm;,,) in the formula above. In practice, this might be done only for small 59, because for large Sq, the cost of finding pmin might be too great relative to the additional utilization gained. An efficient way of finding the required number of virtual servers is needed for fast connection establishment. Note that this is not absolutely necessary for supporting eager bursts as long as we allow the preestablishment of the logical connections. However, if the same processor will be handling the connection admission and burst admission requests, then it would be imperative to have an efficient CAC that will not impede the burst admission processing. Note that a class will require at most one additional virtual server when adding a new connection. Therefore, if the current number of virtual servers allocated to the class is k, we need only compute Pk with the additional traffic included and check if 64 it is still less than the desired BBP. If it is, then the current number of virtual servers is adequate. Otherwise, an additional virtual server will be required to admit the connection. The straightforward method for directly computing P], from p requires two divisions, one multiplication, and one addition per step for 1: steps. Instead of com- puting Pk, one can compute its reciprocal (1 / P1,) = rk obtained recursively from r0 = l r, = i(1/p)rg-1 +1 This approach requires only two multiplications and one addition per step for k steps to compute (1 / Pk) from (1 / p). Furthermore, it is computationally more accurate. A better method is to use a table lookup to find the required number of virtual servers. For a fixed BBP, precompute the maximum possible p for each value of k and store the values in a table,4 e.g., pmu[1:10000]. Then, given a particular p, simply search the table for the minimum pmu[k] that is greater than or equal to the given p. The index k is the required number of servers. In practice, the CAC can simply keep the current pmu[k] in a variable. Then, when adding a new connection, it only has to check if the resulting p exceeds the current pmuUc]. In summary, the CAC procedure is as follows: Compute the new bandwidth re- quirements of the class to which the call belongs if it is added. This amounts to determining if the class needs an additional virtual server. Then check if the addi- tional bandwidth required is available from the residual bandwidth not allocated to any class. The burst admission procedure is simply to check if the bandwidth required is available from the bandwidth allocated to the class but not currently reserved for any burst. These procedures are sketched in Figure 5.2. . ‘A sparser table containing the utilizations p/k for staggered k can be used if the large table size 18 a problem or if large values of k are anticipated. Values of p/k not represented in the table can be obtained via linear interpolation. 65 [Relevant variables] For switch: HadeTbl table of maximum loads (pmu[l:10000]) BurstOv burstwise overhead (h) For link: BIAvail bandwidth not allocated to any class For class: BwServ bandwidth of virtual server (R) NunServ number of virtual servers (1:) ServInUse number of servers in use by bursts TotflorlLd total normalized load for class (2,- pg) HuAdnLd maximum admissible load (pmu[k]) For call: Arr-Rate burst arrival rate (Ac) HeanDur mean burst duration (be) NonLd normalized load [Portions of interest from admission procedurg] A. ALLOCATE BANDWIDTH FOR CALL: NonLd = Arrltate x (HeanDur + BurstOv) IF TotNoruLd + Norah! s HuAdnLd THEN TotNornLd = TotNornLd + 110de ACCEPT CALL ELSE IF Butvail 2 3135011 THEN BIAvail = BsAvail — BIServ NunServ = NunServ + 1 Maxim = HazLd‘l'lelumServ] TotflorlLd = TotNornLd + 110de ACCEPT CALL ELSE REJECT CALL B. RESERVE BANDWIDTH FOR BURST: IF ServInUse < Nuns." THEN ServInUse = ServInUse + 1 ACCEPT BURST ELSE REJECT BURST Figure 5.2: Bandwidth allocation procedures. Procedure A is a portion of the call admission procedure that checks the bandwidth availability for a new connection. It is based on the table lookup alternative and ignores the subtleties with pmin described in the text. Procedure B is a portion of the corresponding burst admission procedure. 66 5 -4 Multi-Class Generalization of the Erlang Loss System A number of researchers have investigated the generalization of the Erlang loss system that allowed resource sharing among different classes of customers [2, 6, 41, 64]. Like the classical Erlang loss system, this system assumes that the arrival process is Poisson and that an arrival that cannot be allocated its required resources is blocked (i.e., dropped from the system). Furthermore, the service time distribution may or may not be exponential; only its mean has to be known. However, the multi-class system allows resource sharing among different classes of customers with different resource demands. The multi-class model can be summarized as follows. 0 There are Q independent classes for a constant but arbitrary Q. 0 Class q has a Poisson arrival rate of Ag. 0 Class q customers have a “spatial” resource demand of R, integral resource units. 0 Class q customers have a mean service time of 8,, time units. In short, each class has its own arrival rate, resource demand, and mean service time. All the customers from all classes are to share the same resource pool according to the resource sharing policy. A variety of resource sharing approaches are possible. For example, in the full sharing approach, all customers have full access to the resources such that no customer is blocked as long as its resource demand is available when it arrives. In the dedicated access approach, the resources are partitioned among the classes, effectively eliminating dynamic sharing of resources among the different classes. In between these two approaches are the partial sharing approaches in which some of the resources are dedicated and the rest are shared. The partial sharing 67 approaches may also be viewed as limiting the access of some classes, and may be used as a priority scheme to control the blocking probabilities of the various classes. The general formulas for the state probability distributions and the blocking prob- abilities are given in [2, 6, 41, 64]. However, in their natural forms or their multi- dimensional recursion equivalents, these solutions are not computationally practical for reasonably large systems. Only in the full sharing approach is there a practically useful one-dimensional recursion [41]. However, various approximations and numeri— cal techniques are available. See [2, 6, 41, 64] for more information. 5 - 5 Performance Evaluation In this section, we will evaluate the burst-oriented approach for handling bursty traffic in ATM networks. In particular, we will study the utilizations achievable by the call admission scheme based on the Erlang loss system given in this chapter. We will also c(Dmpare the eager transmission approach to patient approaches over a wide range of Iletwork scenarios. In addition, we will compare the scheme given in this chapter to another scheme based on a more general traffic model. Finally, simulation results are given that validate the effectivity of the scheme in controlling the BBP of the system. 5.5.1 General Results Here, we briefly examine the maximum utilizations achievable using the admission scheme presented in Section 5.3 for the thin logical connections (TLCS) presented in Section 4.1. Recall that in the given scheme, TLCs are grouped into classes according to their peak rates. Bandwidth is allocated to whole classes by way of virtual server units. We assume that the peak rates are small relative to the total available band- width. For example, classes of 64 Kbps, 2 Mbps, and 10 Mbps can support a wide variety of applications (cf. [20, 27]). In comparison, ATM link bandwidths are in the 68 range of hundreds of Mbps to several Gbps. Figure 5.3 plots the maximum achievable utilizations for a wide range of capacities for different choices of the burst blocking probability (BBP). The utilizations were obtained using Equation (5.2) by computing, for each k, the maximum aggregate p for which P1, is no greater than the BBP. The maximum p divided by I: gives the maximum utilization for the 1: servers. Note that these maximum utilizations apply to all cases where there are more connections than virtual servers in a class.5 Furthermore, if we ignore the burstwise overhead or consider it as part of the aggregate p, then the utilizations do not depend on the burst durations or the burstiness (peak to mean rate) ratios of the connections; they only depend on the aggregate p. The results in Figure 5.3 clearly show that the burst-oriented reservation scheme with eager transmission is able to achieve moderately high utilizations for a wide range of bursty traffic. Note that to obtain the corresponding statistical gains, simply multiply the utilization value by the burstiness ratio. For example, if the utilization is 50% and the burstiness is 20, then the statistical gain is 10; this means that with the same bandwidth, the network is able to support 10 times as many connections with the burst-oriented scheme than with deterministic (peak-rate) allocation. 5-5.2 Network Scenarios In this section, we analyze the difference in the achievable utilizations between the eager and patient transmission schemes in different network scenarios. Recall that the eager transmission scheme, unlike the patient transmission scheme, does not wait for the required bandwidth to be reserved (and acknowledged) before transmitting a burst. Therefore, to compare the performance of eager transmission with the patient \ 5If the number of virtual servers is equal to the number of connections, then the maximum I3<>ssible utilization with no burst blocking is 100%, which is the case when all the connections are a(itive all of the time. 69 009 I I I I I I T I 0.8 0.7 c 0.6 O '3 s 0.5 .. g t BBP =1e-02 . 0.4 , BBP =1e-03 + . ‘ BBP =1e-04 a ‘ BBP =1e-05 x 0-3 - BBP=1e-06 . * 0.2 " ~ 0.1 A M L 1 10 20 30 40 5060708090100 Capacity (1 unit = peak rate of single bursty source) Figure 5.3: Maximum utilizations for the burst-oriented reservation scheme. The x‘a-Xis is the capacity available to a class of bursty connections of the same peak rate. The data points are the maximum achievable utilizations for the different choices of B BP. These maximum bounds apply when the number of connections exceeds the number of capacity units (virtual servers) in a class. 70 Table 5.1: Parameters and settings for the network scenarios. Parameter Setting traffic burstiness 10 available bandwidth 100 Mbps 10-4 BBP (per hop) number of hops LAN: 4 switches MAN: 6 switches WAN: 10 switches LAN: 5 miles (8 km) MAN: 30 miles (48 km) WAN: 3,000 miles (4,800 km) 186,000 miles/s (300,000 km /s) 0.1 ms/switch total link distance propagation speed (speed of light) average switch queueing delay average connection setup delay and average connection tear down delay (excluding bandwidth reservation / release) average bandwidth reservation delay and average bandwidth release delay (with preestablished virtual channel) 0.5 ms/switch 0.01 ms/switch approaches, we simply have to increase the overhead portion of the mean service time in the formula to include the round trip delay and any additional overhead. We analyze the utilizations for three networks scenarios and three bandwidth classes. The network scenarios are the local area (LAN), metropolitan area (MAN), and wide area, (WAN) networks. The bandwidth classes are 64 Kbps, 2 Mbps, and 10 Mbps. The various parameters used in the performance studies are summarized in Table 5.1. The maximum achievable utilizations are plotted in Figures 5.4 to 5.6, where each Curve represents one of the following approaches: O eager—thin logical connections with eager transmission. . fast—fast reservation: thin logical connections with patient transmission. O reconn—reconnection without preestablished logical connections. Here, a con- nection is established and disconnected for each burst. The offered load is limited so that the BBP is maintained. 71 o noGOS—no guarantee of service. Same as reconn but with an unlimited offered load and no guaranteed BBP. In this case, the only limiting factor is the connection setup and tear down overhead for each burst. For theoretical comparison only. 0 PVCs—permanent virtual circuits with permanently reserved bandwidth. For comparison. Note that all the above approaches except noGOS maintains the BBP. In Figure 5.4, we see that for low bandwidth classes with many connections, any burst-oriented approach (eager, fast, or reconn) is able to achieve high bandwidth utilizations. However, in the WAN scenario (Figure 5.4(c)), the eager approach has a. slight advantage. This is because of the greater round trip delay that is effectively an additional overhead for the patient approaches (fast and reconn). For the medium-sized class of 2 Mbps, we see a different story. First, note that Since the total class bandwidth is set to 100 Mbps in this study,6 there are fewer virtual serVers available for this class compared to the 64 Kbps class. This is the reason why the utilizations are in general not as high as in the 64 Kbps class. In any case, we see in Figure 5.5, that the eager approach has a significant utilization advantage over the Patient approaches. In the LAN and MAN scenarios, the advantage is significant for short mean burst lengths. However, in the WAN scenario, the advantage is significant even for long mean burst lengths. Note also that in the WAN scenario, the eager a'I)I>l.‘oach is better than noGOS for short to moderate mean burst lengths, even thongh noGOS does not guarantee any level of service (and is totally impractical be(Rinse of the uncontrolled congestion). \s In no way does this imply that the class bandwidth allocation scheme is static. In the scheme, t . . . the bandwndth allocated to each class 18 increased and decreased as connections are established and orh down. 72 (a) LAN Scenario 1 ‘ we - v hm; we : 0.8 - noGOS -°—-- eager -..m- c fast we" ,9 0.6 - reconn -*-~‘ ‘6 PVCs «~- g 5 0.4 - - 0.2 - - o .-.-l . . - .-...1 . - - -...-1 100 1000 10000 Mean burst length in number of cells (b) MAN Scenario W: : i‘ i i 9“." __ .. _ _ — i — 0.8 - noGOS -o—‘ eager --+--- c last ------e g 0.6 ' . $038" —n—-- g s ..... 5 0.4 - ~ 0.2 - ~ 0 - -...--1 - . -....A1 A - .....n 100 1000 10000 Mean burst length in number of cells (c) WAN Scenario 1 -....-, +~ - : - -:...-; ¢ 4' K - _ _ _ noGOS .._. eager ----+ fast -----a g reconn 4....- g PVCs 4-.. g 0.4 ~ « 0.2 ~ ~ 0 . . l.i..1 A . a. l J . . n 100 1000 10000 Mean burst length in number of cells Figure 5.4: Maximum utilizations for the 64 Kbps class. 73 Utilization A i 1 ‘l noGOS eager fast reconn PVCs A L A AAAAl l L 100 1000 10000 Mean burst length in number of cells (b) MAN Scenario vvv' - '*“'-°;jf;jjg.tttrr.-.-. Utilization h -2g;¢~A.—.——+—I——-1 I noGOS -+— eager "rm-u fast reconn PVCs A A l A A A A I 100 A AAAAA 10000 Mean burst length in number of cells (c) WAN Scenario vvv 'I 0.8 0.6 Utilization 0.4 0.2 vwwvt A r u 7.0" . . 1' A A — *WI noGOS -— eager "rm-- last reconn 4*" PVCs «W1 A A . “l’ "n’ .'/’ ’_/' , . Xx . . s . ./ / . u o l' ,. .C . - A M — _ i v A A A A A I A A A l A 100 1000 10000 Mean burst length in number of cells Figure 5.5: Maximum utilizations for the 2 Mbps class. 74 (a) LAN Scenario 0.8 - g 0.6 I'IOGOS 4—_ == eager -..m 3 fast 2: 0,4 reconn -~—-- 3 PVCs -.-... l 0.2 +. m 0 A AAAAAAL A A AAAAAAl A A A .AAAA1 100 1000 10000 Mean burst length in number of cells (b) MAN Scenario C .2 a: s S 0 AArAAi A A A AAAAAn A A A AAAAAA 100 1000 10000 Mean burst length in number of cells (c) WAN Scenario 1 ' """I fi' '*""‘I ‘ ' "" ‘I 0.8 - 8 0.6 . noGOS -°—. 3; eager --+--- g fast we g 0.4 - reconn +~~ PVCs «~- ,__.-.+--+-----+--- s a T r a ”" l 0'2 .................. - {ff 0‘ A AAAAAAJ A AAAAAAAr A A AAAAAAn 100 1000 10000 Mean burst length in number of cells Figure 5.6: Maximum utilizations for the 10 Mbps class. 75 Figure 5.6 presents the results for the 10 Mbps class. Note that there are only 10 virtual servers in this case. However, for the most part the eager approach still achieves a utilization of about 23%. This is equivalent to a statistical gain of 2.3 since the burstiness is equal to 10 (2.3 is also the ratio of the utilization of eager to the utilization of PVCs). All the observations for the 2 Mbps class apply except that the range in which the eager approach is better extends to longer mean burst lengths. Note that in the case of Figure 5.6(c), when obtaining the results for low mean burst lengths, the arrival rate was deliberately decreased so that there is only one burst on the path per source per round trip time.7 This is to simulate a protocol (such as stop-and—wait) that requires an acknowledgment between bursts. The performance of the eager approach was not affected by this. However, it further increased the overhead and decreased the utilization of the patient approaches. Overall, we see that the eager approach is better than the patient approaches in the cases where either the mean burst length is relatively short, or the round trip delay is long, or both. That the eager approach is better for shorter mean burst lengths can be explained by the fact that the overhead per cell incurred by waiting for the bandwidth to be reserved becomes larger when the number of cells per burst becomes fewer. As already explained, the round trip delay is additional overhead to the patient approaches, but this is not the case for the eager approach. Furthermore, note that the performance gain of the eager approach is more sig- nificant for traffic classes of higher peak rates. This is due to the fact that the same number of cells corresponds to a shorter burst duration (i.e., holding time) when transmitted at a higher peak rate. In general, the amount of benefit of the eager approach over the patient approaches depends on the round trip delay to burst du- ration ratio. An analysis of this relationship was briefly presented in [13]. It is most 7To obtain the results without this adjustment, note that the utilizations will just bottom out at 10%. 76 interesting to note that the performance of the eager approach is only a little affected by the burst duration as long as the reservation overhead is small. The most impor- tant thing that aflects the performance of the eager approach is the number of sources or virtual servers in the class—the more, the better. Finally, note that the burstiness used in the studies was equal to 10, and that the gains of the eager burst-oriented approach will be even greater for higher burstiness values. 5.5.3 Comparison with a Non-Poisson Model of Sources A concern that must be evaluated is how well does a model with an assumption of Poisson sources limit the utilization of the network safely within target BBP re- quirements even if the sources themselves do not follow a Poisson generation process. Although the computation of BBP and the determination of accepting new requests in the Erlang loss system are simple and efficient, it must be determined whether the utilization of the network using the Erlang loss system will exceed the amount that can be safely accepted under a more general model of burst sources. Therefore, to provide a comparison to the Erlang loss system model, we investigate a more general model that only assumes that the sources are independent. The model does not re— quire the arrivals from the sources to be Poisson. What it requires is that the state of each source (i.e., whether it is active or inactive) is independent of states of the other sources. In the basic form of the general model to be used in this section, we will assume that the burst sources are homogeneous. For each source, we assume that p is the fraction of time that the source is active at its peak rate, and the 1 — p is the fraction of time that it is silent. We assume that there are k virtual servers at the network link of concern, and that each virtual server has bandwidth equal to the peak rate of 0.75 - 0.65 r 0.6 - 0.55 ~ Erlang loss system _._. Source utilization p = 0,25 -...--- Source utilization p = 0.10 ------"a Source utilization p = 0,05 4...... tion 0 01 Ila 0.45 Ut ‘ d 090 298's 0.25 ’ 0.2 I I I I 102030405060 7080 90100 Capacity (Virtual Servers) Figure 5.7: Comparison of Erlang loss system to non-Poisson model. The BBP is set to 10“. each source. The number of sources is s. For this general model, which does not assume any (inter)arrival process, the probability that a burst from some source will be blocked will be bounded above by the probability that all the servers are being used by bursts from the competing s -- 1 sources. Therefore, the BBP for the source is given by (cf. the stationary approximation for homogeneous sources used in [28]) s-l _ BBP s 2 p‘(1 — p)“"" (5-3) i=k i The offered load to the servers will be p a: s, which is the individual source utilization of each source (normalized with respect to its peak rate) times the number of sources. We now compare the utilizations allowable by this non-Poisson model with the utilizations allowable by the Erlang loss system. Figure 5.7 plots the achievable utilizations using the Erlang loss system model and the non-Poisson model for selected 78 values of the individual source utilization p. For a small number of sources, or a relatively large value of the source utilization p, the Erlang loss system model safely constrains the number of sources. For p = 0.25, the Erlang loss formula is safe for all numbers of servers shown. For p = 0.1, it appears to be completely safe up to about 55 to 60 servers. However, as the number of independent sources increases, it is possible that the aggregate load of the sources behaves more like a Poisson process, and the Erlang loss formula may be safe even at larger numbers of servers (and sources).8 These results show that as long as the loads of the sources operate within an acceptable degree of burstiness, the simplicity and efficiency of the Erlang loss formula makes it attractive for use in a CAC strategy, even in the cases of non- Poisson sources. 5.5.4 Simulation Results Simulations have been performed to validate the effectivity of the CAC scheme based on the Erlang loss system in controlling the BBP experienced by the admitted connec— tions. Figure 5.8 shows the measured BBPs obtained from simulations using different types of burst source processes. The target BBP was set to 10". The simulation runs were set up as follows. Each run consisted of 107 bursts from independent sources of the same type, each with an individual source load (i.e., mean burst arrival rate multiplied by the mean burst length) equal to 0.1. The number of sources for each capacity value of interest was determined by dividing the maximum utilization allowed by the CAC scheme (i.e., from Figure 5.3) by the individual source load. Seven different types of source processes were used: Poisson process (PP), Markov-modulated PP (MMPP), ¥ 8Note that in the case of heterogeneous sources (i.e., same peak bandwidth but varying p) with the same aggregate average p, the general model will actually allow a higher utilization for reasonably small enough BBP values. This is the result of a demand density curve that will tend to have a higher peak and a flatter tail. 79 0.0001 ................................................. BBP PP +— MMPP -----+ « NMMPP~am IPP ------ IMMPP -..-.-. NIMMPP 1 SMMPP . 1e—05 I I I I 10 20 30 40 50 60 70 80 90100 Capacity (in bandwidth units) Figure 5.8: Simulation results using different burst source models. The target BBP was set to 10". The individual points plot the observed BBPs of individual runs, and the lines connect the corresponding averages of five runs per source type. The legend acronyms indicate the source types as follows: “PP” stands for Poisson process; “MM” stands for Markov modulated; “N” stands for normalized; “1” stands for interrupted; and “S” stands for serialized. These are described in the text. 80 normalized MMPP (N MMPP), interrupted PP (IPP), interrupted MMPP (IMMPP), normalized IMMPP (NIMMPP), and serialized MMPP (SMMPP). The burst lengths (i.e., service times) were exponentially distributed. The PP was simply a homogeneous Poisson process. The MMPP was a non- homogeneous Poisson process whose arrival rate was determined by a two-state continuous-time Markov chain. The IPP was a Poisson process whose arrival rate temporarily dropped to zero (i.e., got interrupted) while the process still had a burst in the system. The IMMPP was the similarly interrupted version of the MMPP. The NMMPP was an MMPP whose burst length distribution was normalized with respect to the interarrival time distribution dictated by the Markov chain so as to maintain the same offered load in the two Markov chain states. The NIMMPP was the analo— gously normalized version of the IMMPP. Finally, the SMMPP was an MMPP whose bursts were serialized through a single server queue before they entered the multi- server system. Note that in the simulations, these processes were used to model single sources, and that as many such independent processes were simulated as was allowed by the CAC scheme. None of the above processes were used to model the aggregate traffic from multiple sources. The IPP, IMMPP, NIMMPP, and SMMPP are on-ofi' processes which can have at most one burst in the system. In contrast, the PP, MMPP, and NMMPP are not on-ofl sources since they can have more than one burst in the system. Strictly speaking, they are inappropriate for the given burst-oriented framework which allows only one burst at a time. However, they are included for comparison. The interarrival times of the PP and IPP are independent. On the contrary, the interarrival times of the Markov-modulated processes are autocorrelated. From Figure 5.8, it is apparent that in this system which does not queue bursts as such, the introduction of autocorrelation through Markov chain modulation does not adversely affect the BBP as long as one of two conditions holds: either the process is of the 81 on-ofl' type, or the process has a relatively stable source load. ’ The first condition is demonstrated by the relative closeness of the observed BBPS of the different on-ofl processes, which include the SMMPP. Note that without se- rialization, the MMPP simulations resulted in BBPs that were higher than allowed. However, after serializing the bursts from each source, which can be viewed as a form of traffic policing, the SMMPP simulations resulted in BBPS that were below the target BBP and were in the neighborhood of the BBPS of the other simulations with on-ofl' sources. That the BBPs were‘significantly below the target BBP can be attributed to the fact that the CAC scheme was derived from a model of the ag- gregate traffic of non-on-ofl' PP sources, effectively making the BBP computations conservative for on-ofi sources (with equivalent effective source loads). The second condition is demonstrated by the NMMPP simulations. Note that as is, the MMPP effectively resulted in an individual source load that fluctuated together with the arrival rate that varied according to the modulating Markov chain. In the NMMPP, the burst length distribution also varied according to the Markov chain state so that the source load was effectively the same for both Markov chain states. This resulted in BBPS that were in the neighborhood of the BBPS of the PP simulations. Naturally, since the CAC scheme used was based on the Erlang loss system which assumed PP sources,9 the BBPS of the PP simulations were in the neighborhood of the target BBP of 10". 5.6 Summary A simple call admission control scheme to control the burst blocking probability has been developed for the framework presented in Chapter 4. It is based on Erlang ’3 loss 9Technically, it assumes a single Poisson process source. However, as already pointed out, com- bining Poisson processes results in an aggregate Poisson process. 82 formula and is shown to be computationally efficient. The scheme allows connections to have arbitrary burst length distributions. It groups the burst-oriented connections into classes of the same peak rate and is shown to achieve high utilizations when used with eager transmission and when there are many connections in a class. The scheme has been evaluated for a wide range of parameters in three hypothetical network scenarios and compared with a more general approach using a non-Poisson source model. Furthermore, simulations have validated the effectivity of the scheme in controlling the BBP. For the most part, the CAC scheme appears to be a viable alternative to computationally more complex approaches even for non-Poisson traffic. Chapter 6 A Burst-Level Priority CAC Scheme for Bursty Traffic Statistical gain is achieved in ATM networks by making bursty connections share resources stochastically. A CAC scheme is used to limit the number of admitted connections to guarantee their QOS requirements. When connections with different QOS requirements share the same resources, the highest QOS requirements would typically be the limiting factor in determining the admissible load at a link. This may lead to connections with low QOS requirements getting better service than they require, leading to an underutilization of the resources. To alleviate this problem, a burst-level priority scheme is proposed. In the proposal, priorities are associated with connections and bursts rather than individual cells. This chapter presents the details and the evaluation of a two-level priority CAC scheme for controlling the burst-level blocking rates of independent heterogeneous on-off sources. 83 84 6.1 Motivation One of the primary attractions of the ATM network is its statistical multiplexing capability. It allows VBR and bursty connections to share the same limited resources (e.g., bandwidth and buffers) stochastically. This, in turn, allows more connections to be supported simultaneously than in a circuit-switched or a synchronous transfer mode (ST M) network. If all traffic from different connections with different QOS requirements are treated identically in the network, then the connections with lower QOS requirements will tend to receive better service than they require. This is the case because the higher QOS requirements of the other connections would tend to limit the available competi- tion (i.e., cause the CAC scheme to admit fewer competing connections). Therefore, without priority handling of traffic within the network, the resources may become underutilized. Different connections may have different levels of tolerance for data loss. Fur- thermore, a connection may tolerate different levels of loss rates depending on the nature of the data it is currently transmitting. On the ATM cell level, the CLP bit is provided in the header to distinguish between the more loss-sensitive portion of the traffic and the less sensitive. In this chapter, an approach is proposed to provide priority handling of bursts of cells (rather than individual cells) inside the network. The proposal consists of a burst admission priority policy and a CAC procedure that controls the BBPS of the two priority levels. The proposal provides a way for bursty connections with different loss requirements to more efficiently share resources, specifically bandwidth. The correctness of the CAC procedure (i.e., that it controls the BBPs to the proper levels) has been verified using simulation techniques. Using analytical techniques, the performance of the two-priority scheme is com- pared with that of the alternative approach where connections with different loss 85 requirements do not share resources. The results demonstrate that a significant in- crease in utilization is achieved by using the priority scheme in a wide variety of configurations. The rest of the chapter is organized as follows. Section 6.2 defines the type of traffic sources that the priority CAC scheme is intended to handle. Section 6.3 describes the burst-level admission procedure and the CAC scheme for connections with identical peak rates. The relevant portions of the CAC scheme are then generalized to handle connections with non-identical peak rates in Section 6.4. The performance benefit of the priority scheme is evaluated in Section 6.5. A summary of the chapter is given in Section 6.6. 6.2 ’I‘raflic Model The traffic sources are assumed to satisfy the following conditions: 0 Each source is of the on-ofi' type that transmits at its peak rate between periods of silence. The distributions of the on and of periods are unspecified. Note that the peak cell transmission rate can be easily enforced with a leaky bucket mechanism (see Section 3.4.1). Furthermore, when a source transmits at a non- zero rate less than the peak rate, it would be treated as if it were transmitting at the peak rate.1 0 The long-term bandwidth demand of a source is characterized by the probability that the source is active. This probability, which we will call the source load, may be different for each source. Note that for random on-ofl sources that always transmit at the peak rate during the on periods, this probability is 1Appropriate buffering at the source may be employed to reduce the underutilization resulting from this condition. 86 equivalent to the mean cell transmission rate of the source normalized to the peak rate, i.e., the mean rate divided by the peak rate. Like the peak rate, the mean rate can also be enforced using a leaky bucket mechanism [33] (separate from the leaky bucket mechanism enforcing the peak rate). However, for maximum effectivity, this leaky bucket mechanism may have to be enhanced so as to take into account the burst groupings of cells. 0 The sources are independent in the sense that at any instant, the state of a source (i.e., whether it is active or inactive) is independent of the states of the others. This is a reasonable assumption that is commonly made for unre- lated connections whose transmissions are typically initiated by non-conspiring parties.2 Note that the independence of the transmissions of admitted connec- tions is the issue here, not the independence of the connection establishment request arrivals. Observe that no assumption is made on the burst arrival process. Furthermore, note that if the sources are homogeneous (i.e., with the same peak rate and source load), then this model reduces to the general non-Poisson model discussed briefly in Section 5.5.3. Here, we are concerned with heterogeneous sources. As required by the burst-oriented traffic control framework, the beginning and end of bursts should be marked as such to allow the “on-the-fly” allocation and deallocatiOn of bandwidth to connections as bursts come and go. When a burst arrives at a switch, bandwidth equal to the peak rate of the source, if available, is allocated to the connection. This bandwidth is deallocated when the end of the burst is detected. If the required bandwidth is not available when a burst arrives, then the whole burst is dropped. 2'The case of (known) cooperating sources is handled in Chapter 7. 87 6.3 CAC Scheme for Connections with Identical Peak Rates In this section, a call admission control scheme for connections with identical peak rates is presented. The CAC scheme guarantees two levels of service: a low burst blocking probability (BBP) for high-priority bursts and a higher BBP for low-priority bursts. For now, we will assume that all bursts belonging to the same connection have the same priority (i.e., each connection has a fixed BBP requirement). However, allow- ing connections to carry bursts of different priorities is also possible and is discussed in Section 6.3.5. 6.3.1 Burst Admission Priority Policy Connections using the same outgoing link at a switch are to partially share a set of bandwidth units each equal to the peak rate of the individual connections. This collection of bandwidth units is allocated to the whole set of connections—individual connections will “hold” the bandwidth units only during their admitted bursts. A high-priority burst is dropped only when there is no bandwidth unit available when it arrives. In contrast, a low-priority burst is also dropped whenever the number of low-priority bursts in the system has reached a (lower) threshold, even if there may be an unused bandwidth unit available. More formally, we define the following variables: 0 Let n = total number of bandwidth units allocated to the set of connections. 0 Let m = low-priority threshold, m S n. 0 Let h = number of bandwidth units currently used by high-priority bursts. 0 Let Z = number of bandwidth units currently used by low—priority bursts. 88 (a) Example where low and high-priority bursts would be admitted. l m n [lollollollollollo]lo]hi]hi[hiLhi[hi] r] L] (b) Example where only high-priority bursts would be admitted. 1 m n [1011:4101lolfllollolloholmlhilhilmlhilJ ] (c) Example where low and high-priority bursts would be rejected. imnolno]wimlmmmlhmnilmlmlmlwrit] Figure 6.1: Burst admission priority policy. Each box represents a bandwidth unit. A box marked “10” has been allocated to an admitted low-priority burst and a box marked “hi” has been allocated to an admitted high-priority burst. The policy allows for only up to m bandwidth units to be used by low-priority bursts but up to n bandwidth units to be used by high-priority bursts at the same time. A high-priority burst is admitted if h +5 < it when it arrives; otherwise, it is dropped. In contrast, a low-priority burst is admitted if both I < m and h + l < n when it arrives; otherwise, it is dropped. The burst admission priority policy is illustrated in Figure 6.1. It is the objective of the CAC scheme to find the minimum 12 (and the corre- sponding m) that would achieve the required BBPS for the low and high-priority connections. In practice, connections will be incrementally added and removed from the set. The CAC scheme then dynamically adjusts m and n as connections are established and torn-down. 6.3.2 Bandwidth Demand Probability Tables To calculate the BBPs for the set of low-priority and the set of high-priority con- nections, two tables are maintained. These tables will store the discrete probability mass functions of the simultaneous bandwidth demands of each set of connections. 89 In theory, this will require a maximum table size equal to the maximum number of connections; but in practice, a table size equal to the maximum number of bandwidth units will suffice for the BBP computations. First, we define the following parameters and notation: e Let N = maximum number of bandwidth units that could be allocated to the set of bursty connections. 0 Let the source loads of the 3 connections, c1, c2, . . . , c,, be p1, p2, . . . , p,, respec- tively. For convenience, let q.- = 1 - pg, for all i. We will assume that 0 < p; < 1 for all i. e Let .C = set of low-priority connections. 0 Let ’H = set of high-priority connections. 0 Let Pc(k) = probability that exactly 10 connections in some set C (e.g., AC or 'H) are simultaneously active. This is given by Pc(k)= Z (Hp.- H Qj) (6.1) SQCJSI=k C; 65 CJ’ EC—S where the product in the summation is simply the probability that all and only the connections in S are active. The summation, in turn, is over all the subsets of C that are of size It. The values of Pc(k) for k = 0, 1, . . . , N —1 may be stored in a table and computed incrementally as connections are added to or removed from an arbitrary set C by using the following recurrence relations (cf. [82]). 90 o For an empty set of connections, we simply have 1 ifk=0 P009): 0 otherwise 0 When adding a connection to a set C, we have, for c,- ¢ C qucUC) if k = 0 Peu{c.-}(k) = (6.2) PiPcUt — 1) + qucUc) if k > 0 Note that Pc(k) vanishes for k > [C]. From the equations above, computing the new table of probabilities,3 when adding a new connection to a set C, will require 2|C I + 1 multiplications and [C I + 1 additions. If we limit the table to N entries (i.e., for k = 0,1, ...,N — 1), then this will require at most 2N — 1 multiplications and N — 1 additions. 0 When removing a connection from a set C, we have, for c, 6 C an Hk=0 P- c k = q’ . C {J}( ) Pcfk)‘PjPC-{fi}(k-1) if k > 0 (6 3) 9: which can be obtained from Equation 6.2 using a “backward substitution” ap- proach. From Equation 6.3 above, computing the new table of probabilities,‘ when removing a connection from a set C, will require 1 division (to compute 1 /q,), 2|C| — 1 multiplications, and |C| — 1 subtractions. If we limit the table to N entries (i.e., for k = 0,1, ..., N — 1), then this will require at most 2N — 1 multiplications and N — 1 subtractions. 3which can be done in place, if necessary or convenient, by computing the new entries from right to left 4which can be done in place by computing the new entries from left to right 91 Note that in general, Pc(k) is non-zero for values of I: up to [C | However, the CAC scheme will not use values of Pc(k) or Py(k) for k 2 N. 6.3.3 Computing the BBP Bounds For convenience, the following notation is introduced: Let Pc'f(k) = probability that k or more connections in set C are simultaneously active. This is given by k-l PM) = 1 — X: Pea) (6.4) i=0 Using this notation, the BBP for a low-priority connection c,-, given the parameters m and n (as defined in Section 6.3.1 above), is bounded above by m-l BBP..- = (Z PA-{..}(k)P;:(n - a) + Pz.{...}(m) (6.5) k=0 where the summation covers the cases where there are less than m active low-priority connections but there are sufficient active high-priority connections to make up a total of n or more active connections of both type. The remaining term covers the cases where there are m or more active low-priority connections. Similarly, for a high-priority connection cj, the BBP is bounded above by m-l BBPC, = (Z Pc(k)P+_{cJ}(n —- k)) + P§(m)P§_{cJ}(n — m) (6.6) k=0 where the summation covers the cases where there are less than m active low-priority connections but there are sufficient active high-priority connections to make up a total of n or more active connections of both type. The remaining term covers the cases where there are m or more active low-priority connections (of which at most m could be activated and holding bandwidth) and sufficient active high-priority connections to make up a total of n or more active connections (when only m activated connections 92 are counted among the active low-priority connections). Normally, we are just concerned with the worst-case BBP experienced by each set of connections. For the low-priority connections, this will be experienced by the connection c,- with the minimum p; in L. For the high-priority connections, this will be experienced by the connection c,- with the minimum p,- in 71. Therefore, it would be sufficient to compute the BBP bounds for those connections rather than all connections. For large Ill] and I’Hl, including or excluding c; and c; from the computations will make little difference, in which case it is preferable just to include them in the (conservative) approximations of the BBP bounds. Therefore, for large [1:] and [71], when determining c; and C; can be costly, we can use the following upper bound for the BBP of the low-priority connections m-l Bl(£,’H,m, n) = (z: Pc(k)P,'['(n — k)) + PZ(m) k=0 where the right hand side is identical to that in Equation 6.5 above except that c: is left in .C in the computation. Similarly, we can use the following upper bound for the BBP of the high-priority connections m-l Bg(£,'H,m,n) = (Z P£(k)P,‘[’(n — k)) + PE(m)P{['(n — m) k=0 where the right hand side is identical to that in Equation 6.6 above except that c, is left in ’H in the computation. Using Equation 6.4, we can compute 81(£,H,m,n) or B;(£,’H,m,n) from P; and PH with m multiplications, m additions, and m + n subtractions. Finally, note that we are only interested in computing Bl(£, ’H, m, n) and Bg(£, 'H, m, n) for values of m and n satisfying m S n S N. 93 6.3.4 Summary of CAC Scheme and Computational Require- ments Let B; and B; be the maximum tolerable BBPS for the low-priority and high-priority connections, respectively, where HI Z 3;. Given C, ’H, m, and n, the procedure for accepting a new connection c, is as follows. 1. If c, is a low-priority connection, then let L, = .C U {c,} and let it, = ’H. Otherwise, let .C, = E and let ’Hz = ’HU{cz}. In other words, add the connection to the appropriate set. 2. If |£z| is small (e.g., if |£,| is less than some small multiple of N), then find the c;, that has the smallest pa, in £2, and let L; = £2 — {C5}. Otherwise, let E’, = L, That is, only if the set of low-priority connections is small will we bother to find the particular low-priority connection that is expected to experience the highest BBP. 3. Similarly, if [71,] is small, then find the Cg that has the smallest pg in ’Hz, and let H; = ’H, — {Cg}. Otherwise, let 71’, = ’Hz. 4. Find the required m and n. (a) If BI(C’,,’Hz,m,n) S BI and Bg(£z,7‘t’z,m,n) S 32, then let m’ = m and n’ = n. In this case, adding the new connection does not cause any of the BBPs to exceed the targeted values with the current m and n. (b) Otherwise, if Bl(£’,,’H,,m +1,n) S BI and B;(£,,’H’z,m + 1,n) S 32, then let m’ = m + 1 and n’ = n. In this case, increasing the low-priority threshold m (without increasing the size of the bandwidth pool n) is suf- ficient to keep the BBPS in check. #1 1A._. ‘. 94 (c) Otherwise, if Bl(L’,,’l-t,,m,n + 1) S B; and Bg(L,,’H',,m,n + 1) S 32, then let m’ = m and n’ = n + 1. In this case, increasing the bandwidth pool (without increasing the low-priority threshold) is sufficient. (d) Otherwise, let m’ = m + 1 and n' = n + 1. Since the new connection can use up at most one bandwidth unit, then this should suffice. 5. If n’ — n bandwidth units can be added to the collection of bandwidth units for L U ’H, then accept the connection c2 and let L = L,, 'H = ’H,, m = m’, and n = n’. Otherwise, reject the connection (at the link in question) and leave L, ’H, m, and n as they were. As long as we have m S n S N, then all the above steps can be done in 0(N) time. The corresponding procedure to release a connection cu, is given below. Note the similarities with the steps given above. 1. If cw is a low-priority connection, then let Lg, = L — {cw} and let Hm = ’H. Otherwise, let Lg, = L and let ’Hq, = ’H — {cw}. 2. If [L5,] is small, then find the c; that has the smallest p, in L0, and let L; = La, — {65,}. Otherwise, let L1,, = Li. 3. If [He] is small, then find the C“; that has the smallest pf, in HQ, and let it}, = Hm — {Cg}. Otherwise, let 711;, = HQ. 4. Find the required m and n. (a) If B,(£;,H,,m — 1,n — 1) g B, and 32(Lz,'H;,m — 1,n — 1) g 32, then letm’zm—landn’zn—l. (b) Otherwise, if Bl(L’z,’l-t,,m,n — 1) S Bl and B;(L,,’H;,m,n — 1) S 3;, thenletm’=mandn’=n—1. 95 (c) Otherwise, if BI(L’,,’H,,m — 1,n) S 31 and B;(L,,’H;,m — 1,n) S 32, thenletm’zm-landn’zn. ((1) Otherwise, let m’ = m and n’ = n. 5. Release n —n’ bandwidth units from the collection of bandwidth units for LU'H, andletL=Lu—,,’H=’Hm,m=m’,andn=n’. Again, all the above steps can be done in 0(N) time. 6.3.5 Allowing Connections to Use Both Priorities This section briefly discusses a technique for admitting a connection that transmits bursts of different priorities. The idea is simply to consider such a connection as having two burst streams, one consisting of the low-priority bursts and the other consisting of the high-priority bursts. The low-priority component is incorporated into L (and P5) and the high-priority component is incorporated into 'H (and Pu). Assuming that the two components may be considered as independent with each of them having its own source load, then the CAC scheme will apply. If it is the case that such connections transmit at most one burst at a time, then the CAC scheme would effectively be a little conservative. 6.4 CAC Scheme for Connections with Non- Identical Peak Rates For simplicity, we will assume that the peak rates of the sources during their on state are multiples of a basic bandwidth unit (e.g., 64 Kbps). For connection q, we will use the symbol R,- to indicate its peak rate in number of bandwidth units. We now generalize the definition of Pc(k) to be based on total simultaneous bandwidth 96 demand. That is, we let Pc(k) = probability that the total bandwidth demand of set C (e.g., L or ’H) is exactly I: bandwidth units. This is given by Pc(k) = 2: (H Pi H €11) {SQCIECiesR,-=k} CiES gee-5 where the product in the summation is simply the probability that all and only the connections in S are active. The summation, in turn, is over all subsets 5 whose total bandwidth demand is I: when all connections in 8 are simultaneously active (cf. Equation 6.1). The recurrence relation for adding a new connection c,- to set C is now ,‘P k if 0 S k < R. PCU{c.-}(k) = q C( ) (6.7) PiPC(k - Rt) + QiPCUc) if k 2 Hi (cf. Equation 6.2). Finally, the corresponding recurrence relation for removing a connection c, from the set C is ”-6051 ifOSk 0 which can be solved from equation (7.1). From equation (7.2) above, computing the new table of probabilities,2 when removing a connection from a set A, will require 1 division (to compute 1/Q.-(0)), less than M x (R,- + 1) multiplications, and fewer subtractions. If we limit the table to M entries, then this will require less than M x (R,- + 1) multiplications and fewer subtractions. Note that in general, PA(k) is non-zero for values of I: up to M. However, the CAC scheme will only use values of P.4(k) for k < N S M. 7 .3.3 Computing the BBP Bounds For convenience, we introduce the following notation: Let Pj(k) = probability that the connections in set A simultaneously require (demand) It or more bandwidth units. This is given by PHI“) = 1_kE:1PA(j) i=0 Using this notation, an upper bound for the BBP of a source belonging to an appli- cation a,- requiring at most R,- bandwidth units is given by 2which can be done in place by computing the new entries from left to right 117 N-R, Pj_{a.,}(N - R.- +1) = 1 — Z PA_{...}(J') i=0 The above formula is derived as follows. Given that we are given only the aggregate bandwidth probability table of application a,-, and given that we are interested in maintaining a source-wise BBP (for fairness), we have to consider the worst-case situation for a source within the application. Without loss of generality, select a source 3 in application a,-. Assume that it produces bursts requiring r bandwidth units. Since application a; as a whole uses up at most R,- bandwidth units simultaneously, then the other sources in a,- (besides s) can use up at most R,— — r bandwidth units at the same time, leaving N — (R,- — r) free for source 5 and the other applications. In the worst-case but perfectly conceivable scenario, it is possible that whenever a burst from source 3 arrives, the other sources in a,- are already using their R,- - r bandwidth units. This means that the burst from source 3 will be blocked unless the other applications in A — {0;} are using less than N — (R,- — r) — r + 1 = N — R,- + 1 bandwidth units. The probability of this situation is given by Pj_{a'}(N — R,- + 1). Computing this value for each a,- in A can be computationally expensive when IA] is large. Therefore, when [AI is large, it is recommended that the upper bound PI(N — 12+ 1) be used for all applications instead. Here, R is the maximum number of bandwidth units required by any single application in A. 7 .3.4 Summary of CAC Scheme and Computational Require- ments Let B be the maximum tolerable BBP. Given A, M, and N, the procedure to accept a new application a,- Q’ A with the bandwidth demand table, Q,(r) for r = 0, 1, 2, . . . , 12", 118 is as follows. 1. Let A’ = A U {a,-} and let R be the maximum number of bandwidth units required by any single application in A’. 2. Compute the new aggregate probability table, PA:(k) for k = 0,1, . . . , M — 1. 3. Find the minimum N’ such that P;,(N' — 13: + 1) s B. 4. If N’ — N additional bandwidth units are available for allocation to the set A, then increase N to N’ and accept the application. Otherwise, reject it. Step 2 above requires less than M x (R;+1) multiplications and fewer subtractions3 as long as we limit the table P to M entries, which is sufficient. Step 3 requires at most M subtractions and R; comparisons, since we are only interested in values of N’ not exceeding M, and since N’ can be at most R,- more than N. As long as the value of R,- tends to be much less than M, then the algorithmic complexity is approximately linear with respect to M. However, if a value of R.- approaching M is quite common, then we suggest that the size (number of entries) of the Q,- table be scaled down appropriately as follows. 1. Determine an appropriate scale factor 3,- that would make the size of the new table, [Rt/3;] + l, acceptable for computational purposes. 2. Construct a new table Q: consisting of the entries 1!ng Q20) = Z Q,(k)forj=1,2,...,[R,/s.] k=s.‘ x(j-l)+l 3Note that we are only counting the required floating point operations. The additions and comparisons associated with loop control are ignored. 119 3. Equation (7.1) now becomes min([k/s.‘J,[Rt‘/8il) . . P.4u{a.-}(k) = Z Qi(J)PA(k — Si X J) i=0 Finally, note that each application can have its own scale factor, and that only “large” applications need to be scaled. 7 .4 Illustration and Performance Evaluation In this section, a simple example demonstrating the use of the proposed model is presented. In the example, we consider the traffic load that could be offered by a hypothetical application at some link. For simplicity in the analysis, we will assume a periodic traffic behavior and ignore traffic jitter and computation-time variations at the hosts. These issues will be addressed later. The scenario is diagrammed in Figure 7.2. In the figure, host M periodically sends a computation request to the slave hosts S. The request passes through the intermediate node N which multicasts the message to the slave hosts. After the slave hosts finish their computations (which we will assume to take uniform time), they send their identically-sized result messages to M via N. The link of concern is the first link emanating from N heading toward M. Obviously, the arrivals of the result messages at N and their demands for bandwidth at L will be strongly correlated. The complete characterization of the traffic pattern from the slaves S to M involves many parameters. However, we can construct the bandwidth demand “probabilities” from the parameters given in Table 7.1. From the given parameters, the resulting “probability” table is the following: r: 0 l 2 3 4 5 H 0‘ p... 0‘ . 2660 25 2 QM 73.66 30—06 m 73.65 mil £7 120 ll \ P 2 in I 0.:- Figure 7.2: Example distributed application diagram. M is the master host, N is an intermediate node, and S is the set of slave hosts sending messages to M via N. The link of concern is marked L. Even though the above table was obtained with the assumption of perfect synchro- nization (in the network and among the hosts), we believe that it is an adequate example for illustration purposes. The variations in network delay and computing time would more likely decrease the “perceived burstiness” by the link because the arrivals will be more scattered in time.4 To evaluate the CAC scheme, we compare its performance in terms of maximum utilization gained to the performance of two alternative approaches. The first alterna- tive approach is the “optimistic assumption” approach, where the individual sources within an application are assumed to be independent of each other, in addition to the applications being independent among themselves. The second alternative approach is the “conservative assumption” approach, which assumes the maximum possible 4In contrast, consider the opposite extreme in which the timings are perfectly “random.” In this case, the variations in the network delay and computing time are likely to reduce the “randomness” and increase overlapping, and thus be harmful to the network (if not taken into account). 121 Table 7.1: Table of parameters for the given example application. Parameter Value number of slaves 5 cycle length (period) 3000 time units S-to—M message length 300 time units number of bandwidth units required 1 per message time between earliest and second arrival 5 time units time between second and third arrival 5 time units time between third and fourth arrival 10 time units time between fourth and last arrival 20 time units overlap among the messages belonging to the same application. To compare the approaches, heterogeneous applications similar to the given exam- ple application (i.e., Figure 7.2) were generated. The applications were given different numbers of sources (ranging from 2 to 10), different degrees of message overlap, and different cycle lengths. The amount of overlap ranged from almost complete overlap (where the messages arrived almost simultaneously) to minimal all-sources overlap (where the last message barely overlapped with the first). For simplicity, each source was assumed to be active 10% of the time (i.e., each cycle). For each approach, the appropriate CAC procedure was applied to determine the maximum number of in- dependent applications admissible given a certain number of bandwidth units and a target BBP of 0.001. The resulting bandwidth utilizations are summarized in Fig- ure 7.3. From the figure, observe that the “correlated” approach results in a utilization be- tween the “conservative” and “optimistic” approaches, as expected. In theory, larger utilizations closer to the optimistic approach would be attainable by the correlated approach if the applications had lower correlation (i.e., less expected overlap). Con- versely, the utilizations would be lower for the correlated approach if the applications had higher correlation. 122 o o o . a"' ...‘ o a " 0' 6 p " .- . C 0. ..v o .. O .6 ‘ '5 -‘ t 8' I I’ a. o- l’ - ' I- D I c 'a .- I' ‘. . . I- ’l’ ,- d o I ,0 I’ ' ’1 I o I optimistic assumption ~— oorrelated assumption -+--- 0.2 of.» conservative assumption ------a - o . A - A l - A A . . . . . j 100 1000 Capacity (in bandwidth units) Figure 7.3: Utilizations allowed by the three CAC approaches. The target BBP was set to 0.001. The “correlated assumption” approach uses the proposed scheme; the “optimistic” approach assumes that all sources are independent; and the “conserva- tive” approach assumes maximal overlap among messages from the same application. Figure 7.3 gives only half the story. One benefit of the “correlated” approach is that it allows us to achieve a higher utilization than the simpler “conservative” approach would allow. The other benefit is that it is more likely to maintain the correct BBP compared to the optimistic approach that assumes that all the sources are independent. This is illustrated in Figure 7.4 which shows that the optimistic approach results in much higher BBPs if the correlation is indeed there. Note that the correlation of concern in the model is that on the likelihood that one source is active given that another source (in the same application) is also active. In the burst-oriented traffic control approach, where bursts from “on-off” sources are not queued as such, the (intra—source) correlation affecting the interarrival times (from the same source) does not appear to be relevant. Using simulations with a wide range of parameters, it has been demonstrated that 123 1rf\'fil\; ' m r J 0.1 f . - f '1 E t 0.01 ? optimistic assumption -0— 3 % : correlated assumption -+--- ‘ m I conservative assumption ------a i 0.001 t : Ar 4 , i : li- ------------------- «a ------------------- .9 -------------------- a -------------------- a -------------------- (n 0.0001 - * 100 1000 Capacity (in bandwidth units) Figure 7.4: Calculated BBPs for the three CAC approaches assuming that the actual traffic follows the correlated model. the proposed CAC scheme does indeed control the BBP for the given traffic model. Figure 7.5 shows the simulation results for a set of synthetic workloads consisting of heterogeneous applications similar to the given example application (i.e., Figure 7.2). To facilitate the data collection, the applications in the simulations had five sources each. This way, the trailing burst from each application cycle experienced the same highest BBP which was supposed to be controlled by the CAC scheme. The applica- tions, however, had different cycle lengths and message overlap patterns. For each capacity value of interest, a set of admitted applications was generated and then ten independent runs (with different initial states/starting points) were executed from that workload. Three BBPs were monitored in each run: the BBP of the leading burst of each application cycle, the BBP of the trailing burst of each application cycle, and the BBP of all bursts. The individual results and the averages are plotted in the figure, which shows that the BBPs were kept below the target BBP of 0.001 as intended. The figure also shows that the BBPs of the trailing bursts were 124 BBP oftrailin bursts 4— BBP ofal bursts -+--- 0-001 BBP of leading bursts ----a- 3 'U BBP Capacity (in bandwidth units) Figure 7.5: Measured BBPs from simulations using workloads admitted by the correlation-aware CAC scheme. The target BBP was set to 0.001 as in the previ- ous figures. The points plot the results of individual runs and the lines connect the corresponding averages. “Leading” and “trailing” bursts refer to the first and last bursts of each application cycle. 125 generally higher than the BBPs of the other bursts. Note that the measured BBPs were all below the target BBP and that the BBPs tended to decrease as the capacity increased. These observations were also made from the simulation results described in Section 6.5. These phenomena can be attributed to the conservative nature of the CAC scheme, which does not take burst blocking into account. Burst blocking tends to occur in bunches, and dropping blocked bursts helps to reduce the likelihood of burst blocking in the near future. In a higher capacity system, this effect becomes more pronounced because of the larger number of states (i.e., number of bandwidth units in use by bursts) that it has. The large number of states results in long periods without threat of burst blocking between short periods of high likelihood of burst blocking. Dropping blocked bursts effectively decreases the load during those periods where it would help the most. Just like in Section 6.5, additional simulations were performed in order to provide evidence that the lower-than-expected BBP and the decreasing trend of the BBPs were mainly due to the dropping of blocked bursts. Figure 7.6 shows the results using the same workloads as in Figure 7.5, except for the use of “persistent” bursts. That is, in the new simulations, blocked bursts were not necessarily dropped completely but instead allowed to stay in the system and later use bandwidth released by departing bursts. This does not mean that the blocked bursts were queued; rather, only the front part of a blocked burst was effectively dropped until a bandwidth unit was released by a departing burst. Note that the measured BBPs of the trailing bursts were closer to the target BBP in the new simulations. Finally, we examine two sets of simulation results evaluating the sensitivity of the CAC scheme to the accuracy of the input application demand probability tables. Figure 7.7 shows the effect of increasing the traffic load given to the system beyond that “admitted” by the CAC scheme. The results were obtained by simply increasing all the message lengths by the given percentage (without increasing the total cycle 126 04-00. BBP BBP of trailin bursts _._.. - BBP of al bursts ...--- BBP of leading bursts ------a 0.0001 ' I 100 1000 Capacity (in bandwidth units) Figure 7.6: Results of the same simulations in the previous figure, but with bursts that were not dropped when blocked. In these simulations, blocked bursts were allowed to persist in the system and later use bandwidth released by departing bursts. time) in the simulations. The “load” seems to be the parameter to which the BBP is most sensitive. Fortunately, the offered load per source is relatively easy to control with appropriate traffic policing. Figure 7.8 shows the effect of increasing the amount of message overlap within applications without increasing the load. The BBP appears to be sensitive to such increase in overlap but not much. The results were obtained by reducing the de- lays between messages (in the same cycle) by the given percentage, making their arrival times closer together. This has the effect of increasing the probability of more messages overlapping without increasing the load. For example, if the cycle time for the scenario in Figure 7.9(a) is 100 time units, then its bandwidth demand probability table is r: 0 1 2 3 Q(r): 0.84 0.06 0.06 0.04 127 0.004 . w v t v . . v . °°°°35 * BBP of trailin bursts 4— BBP ofal bursts -+—-- 0.003 - BBP of leading bursts -a---- . O 2 4 6 8101214161820 Percent load increase Figure 7.7: BBPs resulting from simulations where the traffic load was effectively increased beyond that specified to the CAC scheme. The capacity was 100 bandwidth units, the target BBP was 0.001, and the (unadjusted) workload was that admitted by the “correlated” approach in the previous simulations. This table would be used by the application generator (and CAC scheme) in the simulation. However, the (burst / message) traffic generator would use the adjusted delays shown in Figure 7.9(b), which corresponds to a demand probability table of r: 0 1 2 3 Q(r): 0.87 0.03 0.03 0.07 which is effectively more bursty. It should be noted that other types of variations in the traffic patterns were experimented with, but the above were the only ways in which the resulting BBP was observed to increase. In general, the BBP appears to be tolerant of variations to the delays betWeen messages, message lengths, and cycle durations, as long as the averages of these were maintained in the long run.5 5Needless to say, the independence between applications has to be maintained in all cases. 128 0.0014 I 1 i— l 0.0012 r BBP of trailin bursts +— , BBP of a I bursts --+---- BBP of leading bursts ----a- 0.001 , f“ 0.0008 , ' a. ,I' m m a 0.0006 - 0.0004 ....... a ‘ l """" k ..a ......... 0.0002 .--— ‘* ,,,,,,,,,, a ......... a -------- 8'” . T ........ G ........ +3" 0 1 1 1 1 0 20 40 60 80 100 Percent overlap adjustment Figure 7.8: BBPs resulting from simulations where the intra-application overlap was effectively increased beyond that specified to the CAC scheme. The capacity was 100 bandwidth units, the target BBP was 0.001, and the (unadjusted) workload was that admitted by the “correlated” approach in the previous simulations. 7 .5 Summary In this chapter, a general approach for modeling the bandwidth demands of correlated multiple sources belonging to a distributed application is presented. The model is based on specifying (or estimating) the aggregate bandwidth demand distribution of a set of sources over a shared communication link. The distribution is specified as an empirical probability mass function in tabular form. This table is then combined with other such tables using the tractable algorithm given in Section 7.3. The result is a global (i.e., including all pertinent applications using the link) bandwidth demand distribution from which the burst blocking probability can be computed. Using a simple example of a correlated application, the model and CAC scheme are shown to allow a utilization that falls between a more conservative but less bandwidth- efficient approach and a more overly optimistic approach that does not take any 129 16 (b) After 50% adjustment. Figure 7.9: Diagram of the message overlap profile of an application, with three sources using the same link of concern. Figure (b) shows the effect of decreasing the relative message delays in (a) by 50%. correlation into account. The penalties for using these other approaches are lower bandwidth utilizations for the conservative approach, on the one hand, and higher burst loss rates for the optimistic approach, on the other hand. These penalties are shown to be significant in Section 7.4. That the CAC scheme works in controlling the BBP as intended has been demon- strated using simulations. In addition, additional simulations provide evidence that the CAC scheme is quite “tolerant to variations in the traffic patterns as long as long-term averages in the traffic parameters are maintained. However, the technique appears to be specially sensitive to increases in the traffic load (which is to be ex- pected from all but the most conservative approaches). In addition, it appears to be sensitive (to a lesser extent) to increases in intra-application overlap. Chapter 8 Conclusion 8.1 Summary of Contributions A comprehensive framework for burst-level traffic control in ATM networks has been presented. The framework provides mechanisms for supporting burst-oriented con- nections in which bursts of cells are granted (or denied) bandwidth on-the-fly. This allows such connections to start transmitting bursts without having to fully reserve bandwidth beforehand. Two types of burst-oriented connections are supported: thin logical connections (TLCS) which have zero bandwidth between bursts and fat logical connections (FLCs) which have non-zero reserved bandwidth between bursts (i.e., high-activity periods). The framework includes mechanisms to support the bundling of these connections into virtual paths. Congestion control in this kind of framework is performed at two levels. Burst- level congestion control consists of properly forwarding or dropping bursts based on the availability of bandwidth and optionally on the priority of the bursts. To get any reasonable performance at this level, hardware support will be required from the ATM switching elements. Call-level congestion control is performed at connection establishment by a call admission control (CAC) scheme. The CAC scheme either 130 131 admits or rejects connection establishment requests based on whether the anticipated burst-level congestion will be acceptable to all parties concerned. Three new CAC schemes have been presented and evaluated in this dissertation. The first is a simple scheme based on the Erlang loss system model. Its primary merit is its low computational requirement compared to other schemes using more general traffic models. Despite its simplicity, it appears to be potentially adequate under a wide range of conditions. The second CAC scheme is a burst-level priority scheme based on a more general (non-Poisson and non-renewal) traffic source model. The scheme as presented sup- ports two priority levels for the connections (or bursts). The low and high-priority connections share the same pool of resources but with the low-priority bursts con- strained to use only a subset of the shared bandwidth allocation units. This results in a higher blocking rate for the low-priority connections. The sharing of resources this way results in higher bandwidth utilization (i.e., more admitted connections) than when the two classes of connections are either treated identically (i.e., with the same low BBP) or managed separately (i.e., do not share resources). This two-level scheme can be extended to handle more than two priority levels, but the result would be significantly more complex (computationally) with little expected additional per- formance benefit. The third CAC scheme is an approach that can take into account known traffic correlation among sources. Such correlation is likely to exist among a set of traffic sources associated with a parallel distributed computing application. Most CAC schemes (burst-oriented or otherwise) assume independent traffic sources and this type of correlation will likely degrade the performance of the network unacceptably. The proposed “correlation-aware” scheme is based on a general “set-of-sources” model to capture the correlation and an algorithm similar to that used in the second CAC scheme to compute the aggregate bandwidth demand probabilities. When compared 132 to two alternative approaches, one based on an optimistic “independent sources” assumption and another based on a conservative “maximum overlap” assumption, the proposed scheme appears to strike the proper balance in terms of loss rate and bandwidth utilization. This scheme may be extended to support burst-level priorities like in the second CAC scheme but with the restriction that each set of correlated sources have the same priority.1 The following is a summary of the advantages of the burst-oriented approach over the more common cell-oriented approach in handling bursty traffic.2 0 The loss rate at the burst level may be a more useful performance metric than the cell loss rate as it has direct impact on the performance of end-to—end protocols handling protocol data units larger than a cell. 0 In periods of congestion, cell losses will be shared by all active sources sharing a link (cell loss priority aside) in the cell-oriented approach. In contrast, cell losses will be experienced by the same (blocked) bursts in the burst-oriented approach. In fact, the burst-oriented framework provides a form of “congestion relief” by deliberately dropping the cells of blocked bursts (cf. [5, 65]). e The burst-oriented approach with “on-the—fly” bandwidth reservation requires only enough buffers to support equivalent CBR connections (circuit emulation) in the presence of cell delay jitter. However, if cell queueing beyond that re- quired for absorbing jitter is to be considered in the cell—oriented approach, then the bandwidth demand analysis would be considerably more complicated. This is the case because it would depend on buffer sizes, burst length distributions, and cell interarrival processes. 1Of course, it is mathematically possible to remove this restriction and also to increase the number of priority levels, but the resulting procedure would be computationally impractical. 2Note than in the following, we are primarily concerned with cell losses due to congestion. Cell losses due to physical layer and switch hardware memory bit errors will be experienced equally by both approaches. 133 8.2 Future Work The original motivation and the main emphasis of this dissertation in the use of the burst-oriented approach in the transmission of bursty data traffic. However, with additional work, multimedia traffic, specifically VBR compressed packet video traffic, may also be supported by the framework. The two primary benefits of the burst- oriented approach are its low buffer requirements and its low (queueing) latency. As already pointed out, burst-oriented connections behave much like CBR connections (circuit emulation), except for the head and tail processing. However, a few issues have to be addressed as follows. First, the inherent periodicity of video traffic has to be handled properly. Peri- odicity may be good or bad. In the best case, interleaving of bursts from different sources may occur, resulting in very little and fair burst losses. In the worst case, a set of sources could be caught in a “lock-step” pattern of synchronized arrivals that would result in excessive and unfair burst losses. Note that a connection would typically have to pass through multiple switches and compete with a different set of connections at each switch. Therefore, for a particular connection, a favorable timing (“phase”) at one switch may be disastrous in another switch in the path. Beware of Murphy’s Law. One general solution to the periodicity problem is to introduce randomness in the phase or timing of the frames. In some video coding-decoding schemes (codecs) such as MPEG [26], compressed frames come in different types and sizes. Naturally the bandwidth allocated to the connection would be dictated by the size of the largest frames. Therefore, for most of the smaller frames, the time to transmit them would be smaller than the frame period. This gives the sender some room for randomizing the frame transmission times. This approach, however, does not solve the more subtle problem of periodicity within larger time scales. For example, in MPEG, if the larger 134 I-frames are sent every 12 frames, then a higher-level 12—frame—long period of high activity will result. After the first frame has been transmitted, the source does not have much flexibility in positioning the subsequent I-frames (unless it has control over the encoder which would not be the case in the transmission of precompressed, prerecorded video). Another issue of concern is the debatable (in)appropriateness of burst-level dis- carding in video (as opposed to cell-level discarding during periods of congestion). Depending on the encoding scheme, it is possible that losing a cell in a frame may not necessarily invalidate the whole frame. That is, the remaining cells in the frame may still be useful to the decoder. For example, this would be the case if the data in the cell applied to a region of the picture rather than the picture as a whole. However, tolerance to single cell losses requires the additional overhead of sequence numbers in the cells in order for the decoder to identify the lost cells. In any case, if whole bursts are to be discarded in periods of congestion, then an effort must be made to minimize the perceptability of the losses. One approach is to use burst-level priorities. Multi-layered MPEG coders have been designed that separate the output components of the standard coder into high- priority and low-priority streams (e.g., [58]). The high-priority stream carries most of the “energy” in the picture, and the scheme is able to tolerate higher loss rates in the low-priority stream. These streams may then be carried in different bursts with different priorities. Alternatively or in addition to the above, the more important frames (such as I-frames which have longer temporal significance) can be given higher priority than the less important frames. Moreover, to further minimize the losses of high-priority frames, preemption (of low-priority bursts) may be supported. The above may further be enhanced with limited buffering such as the technique (intended for voice) used in burst-switching [31]. This allows the bulk of a burst to get through even if part of it gets clipped. 135 The above suggestions assume the use of a TLC to carry the video traffic. However, it is also possible to use an F LC instead. One possible approach is to start with a video codec that is designed to produce CBR output such as the p x 64 kbit/s video standard [48]. The constant bit rate output is obtained by the coding scheme by dynamically controlling the degree of quantization (least-significant-bit truncation) in the compressor. The drawback of this approach is that it results in a variable quality output. To produce constant quality output, an F LC may be used to carry the truncated information in a burst-by-burst fashion in addition to the CBR component. This general idea of sending the high-priority stream in a bandwidth-reserved manner while sending the low-priority stream in a less resource-demanding manner has been suggested in [58]. BIBLIOGRAPHY Bibliography [1] Bandula W. Abeysundara and Ahmed E. Kamal. High-speed local area networks and their performance: A survey. ACM Computing Surveys, 23(2):221—264, June 1991. [2] J. M. Aein. A multi-user—class, blocked-calls-cleared, demand access model. IEEE Trans. Commun., COM-26(3):378—385, March 1978. [3] Stanford R. Amstutz. Burst switching—an introduction. IEEE Commun. Mag., pages 36-42, November 1983. [4] Heinrich Armbriister and Klaus Wimmer. Broadband multimedia applications using ATM networks: High-performance computing, high-capacity storage, and high—speed communication. IEEE J. Sel. Areas Commun., 10(9):l382—1396, De- cember 1992. [5] G. Armitage and K. Adams. Packet reassembly during cell loss. IEEE Network, 7(5):26-34, September 1993. [6] E. Arthurs and J. S. Kaufman. Sizing a message store subject to blocking criteria. In M. Arato, A. Butrimenko, and E. Gelenbe, editors, Performance of Computer Systems, pages 547—564. North-Holland, 1979. [7] Jaime Jungok Bae and Tatsuya Suda. Survey of traffic control schemes and protocols in ATM networks. Proc. of the IEEE, 79(2):170—189, February 1991. [8] Krishna Bala, Israel Cidon, and Khosrow Sohraby. Congestion control for high speed packet switched networks. In Proc. INF 000M ’90, pages 520-526. IEEE, 1990. [9] F lavio Bonomi and Kerry W. Fendick. The rate-based flow control framework for the available bit rate ATM service. IEEE Network, pages 25—39, March 1995. [10] Laura J. Bottomley and Arne A. N ilsson. Traffic characterization in a wide area network. In Harry Perros, editor, High-Speed Communication Networks, pages 213—224. Plenum Press, 1992. [11] Jean-Yves Le Boudec. The Asynchronous Transfer Mode: a tutorial. Computer Networks and ISDN Systems, 24:279-309, 1992. 136 137 [12] Andreas D. Bovopoulos. Performance evaluation of a traffic control mechanism for ATM networks. In Proc. INF OCOM ’92, pages 469—478. IEEE, 1992. [13] Pierre E. Boyer and Didier P. Tranchier. A reservation principle with applications to the ATM traffic control. Computer Networks and ISDN Systems, 24:321—334, 1992. [14] Milena Butto’, Elisa Cavallero, and Alberto Tonietti. Effectiveness of the “leaky bucket” policing mechanism in ATM networks. IEEE J. Sel. Areas Commun., 9(3):335-342, April 1991. [15] Ramon Caceres et a1. Characteristics of wide-area TCP/IP conversations. Com— puter Commun. Review, 21(4):101-112, September 1991. [16] Nim K. Cheung. The infrastructure for gigabit computer networks. IEEE Com— mun. Mag., pages 60—68, April 1992. [17] Israel Cidon, Inder Gopal, and Adrian Segall. Fast connection establishment in high speed networks. Computer Commun. Review, 20(4):287-—296, September 1990. [18] Douglas E. Comer. Internetworking with TCP/IP, volume I: Principles, Proto- cols, and Architecture. Prentice-Hall, second edition, 1991. [19] Simon Crosby. In-call renegotiation of traffic parameters. In Proc. INF 0- COM ’93, pages 638—646. IEEE, March 1993. [20] M. Decina et a1. Bandwidth assignment and virtual call blocking in ATM net- works. In Proc. INFOCOM ’90, pages 881—888. IEEE, 1990. [21] W. A. Doeringer et al. Fast connection establishment in large-scale networks. In Proc. INFOCOM ’93, pages 489—496. IEEE, 1993. [22] Bharat Doshi. Performance of in—call buffer/ window allocation schemes for short intermittent file transfers over broadband packet networks. In Proc. INF 0- COM ’92, pages 2463—2471. IEEE, May 1992. [23] Bharat Doshi and Harry Heffes. Performance of an in-call buffer-window reser- vation/allocation scheme for long file transfers. IEEE J. Sel. Areas Commun., 9(7):]013—1023, September 1991. [24] Bharat T. Doshi and Pravin K. Johri. Communication protocols for high speed packet networks. Computer Networks and ISDN Systems, 24:243—273, 1992. [25] Fore Systems, Inc., 1000 Gamma Drive, Suite 504, Pittsburgh, PA 15238. Fore- Runner( TM) ASX-100 ATM Switch Architecture Manual (Release 2.1), 1993. [26] Didier Le Gall. MPEG: A video comnpression standard for multimedia applica- tions. Commun. of the ACM, 34(4):46—58, April 1991. 138 [27] G. Gallassi, G. Rigolio, and L. Fratta. ATM: Bandwidth assignment and band- width enforcement policies. In Proc. GLOBECOM ’89, pages 1788-1793. IEEE, 1989. [28] Roch Guérin, Hamid Ahmadi, and Mahmoud Naghshineh. Equivalent capacity and its application to bandwidth allocation in high-speed networks. IEEE J. Sel. Areas Commun., 9(7):968—981, September 1991. [29] Levent Giin and Roch Guérin. An overview of bandwidth management proce— dures in high-speed networks. In Harry Perros, editor, High-Speed Communica- tion Networks, pages 35—45. Plenum Press, 1992. [30] Ibrahim W. Habib and Tarek N. Saadawi. Controlling flow and avoiding con- gestion in broadband networks. IEEE Commun. Mag., pages 46—53, October 1991. [31] E. Fletcher Haselton. A PCM frame switching concept leading to burst switching network architecture. IEEE Commun. Mag., pages 13-19, September 1983. [32] Harry Heffes and David M. Lucantoni. A Markov modulated characterization of packetized voice and data traffic and related statistical multiplexer performance. IEEE J. Sel. Areas Commun., SAC-4(6):856-867, September 1986. [33] Douglas S. Holtsinger and Harry G. Perros. Performance of the buffered leaky bucket policing mechanism. In Harry Perros, editor, High-Speed Communication Networks, pages 47—69. Plenum Press, 1992. [34] Chengchang Huang and Philip K. McKinley. Communication issues in parallel computing across ATM networks. Parallel 8' Distributed Technology, 2(4):73—86, 1994. [35] Joseph Y. Hui. Resource allocation for broadband networks. IEEE J. Sel. Areas Commun., 6(9):]598—1608, December 1988. [36] A. Iwata et al. ATM connection and traffic management schemes for multimedia internetworking. Commun. of the ACM, 38(2):72-89, February 1995. [37] Raj Jain. FDDI: Current issues and future plans. IEEE Commun. Mag., pages 98—105, September 1993. [38] Raj Jain and Shawn A. Routhier. Packet trains—measurements and a new model for computer network traffic. IEEE. J. Sel. Areas Commun., 4(6):986—995, September 1986. [39] Per Jomer. Connection caching of traffic adaptive dynamic virtual circuits. Com- puter Commun. Review, 19(4):13—24, September 1989. 139 [40] Takashi Kamitake and Tatsuya Suda. Evaluation of an admission control scheme for an ATM network considering fluctuations in cell loss rate. In Proc. GLOBE- COM ’89, pages 1774-1780. IEEE, 1989. [41] Joseph S. Kaufman. Blocking in a shared resource environment. IEEE Trans. Commun., COM—29(10):1474—1481, October 1981. [42] Gary C. Kessler and David A. Train. Metropolitan Area Networks: Concepts, Standards, and Services. McGraw-Hill, 1992. [43] Leonard Kleinrock. The latency/bandwidth tradeoff in gigabit networks. IEEE Commun. Mag., pages 36-40, April 1992. [44] H. T. Kung and Alan Chapman. The FCVC (flow-controlled virtual channels) proposal for ATM networks. In Proc. 1993 Int. Conf. on Network Protocols, pages 116—127, 1993. [45] H. T. Kung and Robert Morris. Credit-based flow control for ATM networks. IEEE Network, pages 40—48, March 1995. [46] Jim Kurose. Open issues and challenges in providing quality of service guarantees in high-speed networks. Computer Commun. Review, 23(1):6-15, January 1993. [47] Will E. Leland et al. On the self-similar nature of Ethernet traffic. Computer Commun. Review, 23(4):183—193, October 1993. [48] Ming Liou. Overview of the px64 kbit/s video coding standard. Commun. of the ACM, 34(4):59—63, April 1991. [49] Basil Maglaris et al. Routing of voice and data in burst-switched networks. IEEE Trans. Commun., 38(6):889—896, June 1990. [50] Peter Newman. ATM local area networks. IEEE Commun. Mag., pages 86—98, March 1994. [51] Ioanis Nikolaidis and Ralf O. Onvural. A bibliography on performance issues in ATM networks. Computer Commun. Review, 22(5):8—23, October 1992. [52] Zhisheng N iu and Haruo Akimaru. Analysis of statistical multiplexer with selec- tive cell discarding control in ATM systems. IEICE Trans., E74(12):4069—4079, December 1991. [53] Katia Obraczka, Peter B. Danzig, and Shih—Hao Li. Internet resource discovery services. IEEE Computer, pages 8-22, September 1993. [54] Hirokazu Ohnishi, Tadanobu Okada, and Kiyohiro Noguchi. Flow control schemes and delay/loss tradeoff in ATM networks. IEEE J. Sel. Areas Com- mun., 6(9), December 1988. 140 [55] Robert T. Olsen and Lawrence H. Landweber. Design and implementation of a fast virtual channel establishment method for ATM networks. In Proc. INF 0- COM ’93, pages 617-627. IEEE, March 1993. [56] Ralf O. Onvural and Ioanis Nikolaidis. Routing in ATM networks. In Harry Perros, editor, High-Speed Communication Networks, pages 139-150. Plenum Press, 1992. [57] Tennis Ott, T. V. Lakshman, and Ali Tabatabai. A scheme for smoothing delay- sensitive traffic offered to ATM networks. In Proc. INFOCOM ’92, pages 776- 785. IEEE, 1992. [58] P. Pancha and M. E1 Zarki. Prioritized transmission of variable bit rate MPEG video. In Proc. GLOBECOM ’92, pages 1135-1139. IEEE, December 1992. [59] Pramod Pancha and Magda El Zarki. Bandwidth requirements of variable bit rate MPEG sources in ATM networks. In Proc. INF OCOM ’93, pages 902-909. IEEE, 1993. [60] M. De Prycker, R. Peschi, and T. Van Landegem. ~B-ISDN and the OSI protocol reference model. IEEE Network, 7(2):10—18, March 1993. [61] K. K. Ramakrishnan and Peter Newman. Integration of rate and credit schemes for ATM flow control. IEEE Network, pages 49-56, March 1995. [62] Erwin P. Rathgeb. Modeling and performance comparison of policing mecha- nisms for ATM networks. IEEE J. Sel. Areas Commun., 9(3):325-334, April 1991. [63] Michael J. Rider. Protocols for ATM access networks. IEEE Network, pages 17-22, January 1989. [64] James W. Roberts. A service system with heterogeneous user requirements— application to multi-services telecommunications systems. In G. Pujolle, editor, Performance of Data Communications Systems, pages 423-431. North-Holland, 1981. [65] Allyn Romanow and Sally Floyd. Dynamics of TCP traffic over ATM networks. Computer Commun. Review, 24(4), October 1994. [66] Reza Rooholamini, Vladimir Cherkassky, and Mark Garver. Finding the right ATM switch for the market. IEEE Computer, pages 16—28, April 1994. [67] Jonathan Rosenberg et al. Multimedia communications for users. IEEE Com- mun. Mag., pages 20-36, May 1992. [68] Sheldon M. Ross. Introduction to Probability Models. Academic Press, fourth edition, 1989. 141 [69] Matthew N. O. Sadiku and A. S. Arvind. Annotated bibliography on Distributed Queue Dual Bus (DQDB). Computer Commun. Review, 24(1):21-35, January 1994. [70] Hiroshi Saito. Call admission control in an ATM network using upper bound of cell loss probability. IEEE Trans. Commun., 40(9):1512-1521, September 1992. [71] Kiyoshi Shimokoshi. Evaluation of policing mechanisms for ATM networks. IE- ICE Trans., E76-B(10):1341-1351, November 1993. [72] John D. Spragins, Joseph L. Hammond, and Krzysztof Pawlikowski. Telecom- munications: Protocols and Design. Addison-Wesley, 1991. [73] William Stallings. Data and Computer Communications. Macmillan, third edi- tion, 1991. [74] William Stallings. Local and Metropolitan Area Networks. Macmillan, fourth edition, 1993. [75] William Stallings. ISDN and Broadband ISDN with Frame Relay and ATM. Prentice-Hall, third edition, 1995. [76] Hiroshi Suzuki, Tutomu Murase, Syohei Sato, and Takao Takeuchi. A burst traffic control strategy for ATM networks. In Proc. CLOBECOM ’90, pages 874—878. IEEE, 1990. [77] Toshikazu Suzuki. ATM adaptation layer protocol. IEEE Commun. Mag., pages 80-83, April 1994. [78] Lajos Takacs. .On Erlang’s formula. The Annals of Mathematical Statistics, 40(1):71-78, 1969. [79] Andrew S. Tanenbaum. Computer Networks. Prentice-Hall, second edition, 1988. [80] Don Tolmie and John Renwick. HIPPI: Simplicity yields success. IEEE Network, 7(1):28—32, January 1993. [81] Jonathan S. Turner. New directions in communications (or which way to the information age?) IEEE Commun. Mag., 24(10):8—15, October 1986. [82] Jonathan S. Turner. Managing bandwidth in ATM networks with bursty traffic. IEEE Network, pages 50—58, September 1992. [83] Faramak Vakil and Hiroshi Saito. On congestion control in ATM networks. IEEE LTS, pages 55-65, August 1991. [84] Jean Walrand. Communication Networks: A First Course. Aksen Associates, 1991. 142 [85] Patricia E. Wirth and Kathleen S. Meier-Hellstern. Traffic models for ISDN and B-ISDN users. In Harry Perros, editor, High-Speed Communication Networks, pages 205-211. Plenum Press, 1992. [86] Nanying Yin, San-Qi Li, and Thomas E. Stern. Congestion control for packet voice by selective packet discarding. IEEE Trans. Commun., 38(5):674-683, May '1990. "1111111111111111“