Publications (27)16.63 Total impact
 [Show abstract] [Hide abstract]
ABSTRACT: We start with the premise, and provide evidence that it is valid, that a Markovmodulated Poisson process (MMPP) is a good model for Internet traffic at the packet/byte level. We present an algorithm to estimate the parameters and size of a discrete MMPP (DMMPP) from a data trace. This algorithm requires only two passes through the data. In tandemnetwork queueing models, the input to a downstream queue is the output from an upstream queue, so the arrival rate is limited by the rate of the upstream queue. We show how to modify the MMPP describing the arrivals to the upstream queue to approximate this effect. To extend this idea to networks that are not tandem, we show how to approximate the superposition of MMPPs without encountering the statespace explosion that occurs in exact computations. Numerical examples that demonstrate the accuracy of these methods are given. We also present a method to convert our estimated DMMPP to a continuoustime MMPP, which is used as the arrival process in a matrixanalytic queueing model.IEEE/ACM Transactions on Networking 01/2004; · 2.01 Impact Factor  [Show abstract] [Hide abstract]
ABSTRACT: With the rapid growth of Internet applications built on TCP/IP such as the World Wide Web and the standardization of traffic management schemes such as Available Bit Rate (ABR) in Asynchronous Transfer Mode (ATM) networks, it is essential to evaluate the performance of feedbackbased protocols using traffic models which are specific to dominant applications. This paper presents a method for analyzing feedbackbased protocols with a Webuserlike input traffic where the source alternates between ‘transfer’ periods followed by ‘think’ periods. Our key results, which are presented for the TCP protocol, are as follows: (1) When the roundtrip time is the same for all users, the goodputs and the fraction of time that the system has some given number of transferring sources are insensitive to the distributions of transfer (file or page) sizes and think times except through the ratio of their means. Thus, apart from network roundtrip times, only the ratio of average transfer sizes and think times of users need be known to size the network for achieving a specific quality of service. (2) The Engset model can be adapted to accurately compute goodputs for TCP and TCP over ATM, with different buffer management schemes. Though only these adaptations are given in the paper, the method based on the Engset model can be applied to analyze other feedback systems, such as ATM ABR, by finding a protocol specific adaptation. Hence, the method we develop is useful not only for analyzing TCP using a source model significantly different from the commonly used persistent sources, but also can be useful for analyzing other feedback schemes. (3) Comparisons of simulated TCP traffic to measured Ethernet traffic shows qualitatively similar second order autocorrelation when think times follow a Pareto distribution with infinite variance. Also, the simulated and measured traffic have long range dependence. In this sense our traffic model, which purports to be Webuserlike, also agrees with measured data traffic.Computer Communications. 01/2003;  [Show abstract] [Hide abstract]
ABSTRACT: Despite the fact that most of todays' Internet traffic is transmitted via the TCP protocol, the performance behavior of networks with TCP traffic is still not well understood. Recent research activities have lead to a number of performance models for TCP traffic, but the degree of accuracy of these models in realistic scenarios is still questionable. This paper provides a comparison of the results (in terms of average throughput per connection) of three different `analytic' TCP models: I. the throughput formula in [Padhye et al. 98], II. the modi ed Engset model of [Heyman et al. 97], and III. the analytic TCP queueing model of [Schwefel 01] that is a packet based extension of (II). Results for all three models are computed for a scenario of N identical TCP sources that transmit data in individual TCP connections of stochastically varying size. The results for the average throughput per connection in the analytic models are compared with simulations of detailed TCP behavior. All of the analytic models are expected to show deficiencies in certain scenarios, since they neglect highly inuential parameters of the actual real simulation model: The approach of Model (I) and (II) only indirectly considers queueing in bottleneck routers, and in certain scenarios those models are not able to adequately describe the impact of bufferspace, neither qualitatively nor quantitatively. Furthermore, (II) is insensitive to the actual distribution of the connection sizes. As a consequence, their prediction would also be insensitive of socalled longrange dependent properties in the traffic that are caused by heavytailed connection size distributions. The simulation results show that such properties cannot be neglected for certain network topologies: LRD properties can even have counterintuitive imp...Proc SPIE 10/2001;  [Show abstract] [Hide abstract]
ABSTRACT: Despite the fact that most of todays' Internet traffic is transmitted via the TCP protocol, the performance behavior of networks with TCP traffic is still not well understood. Recent research activities have lead to a number of performance models for TCP traffic, but the degree of accuracy of these models in realistic scenarios is still questionable. This paper provides a comparison of the results (in terms of average throughput per connection) of three different `analytic' TCP models: I. the throughput formula in [Padhye et al. 98], II. the modified Engset model of [Heyman et al. 97], and III. the analytic TCP queueing model of [Schwefel 01] that is a packet based extension of (II). Results for all three models are computed for a scenario of $N$ identical TCP sources that transmit data in individual TCP connections of stochastically varying size. The results for the average throughput per connection in the analytic models are compared with simulations of detailed TCP behavior. All of the analytic models are expected to show deficiencies in certain scenarios, since they neglect highly influential parameters of the actual real simulation model: The approach of Model (I) and (II) only indirectly considers queueing in bottleneck routers, and in certain scenarios those models are not able to adequately describe the impact of bufferspace, neither qualitatively nor quantitatively. Furthermore, (II) is insensitive to the actual distribution of the connection sizes. As a consequence, their prediction would also be insensitive of socalled longrange dependent properties in the traffic that are caused by heavytailed connection size distributions. The simulation results show that such properties cannot be neglected for certain network topologies: LRD properties can even have counterintuitive impact on the average goodput, namely the goodput can be higher for small buffersizes.© (2001) COPYRIGHT SPIEThe International Society for Optical Engineering. Downloading of the abstract is permitted for personal use only.07/2001;  [Show abstract] [Hide abstract]
ABSTRACT: Heyman (1992) examined three sequences giving the number of cells per frame of a VBR encoding of videoconferences (talking heads); these sequences were produced by hardware encoders using different coding algorithms. Each sequence had a gamma marginal distribution, and the autocorrelation function was geometric up to lags of at least 3 s, which includes all autocorrelation values larger than 0.1. We present an easy to simulate autoregressive process that has these properties. The model is tested by comparing the cellloss rate produced when the data trace was used as the sole source in a simulation of an ATM switch to the cellloss rates produced when traces generated by the model were used as the sourceIEEE/ACM Transactions on Networking 09/1997; 5(4):554560. · 2.01 Impact Factor  [Show abstract] [Hide abstract]
ABSTRACT: The GrassmannTaksarHeyman algorithm is a direct algorithm for computing the steadystate distribution of a finite irreducible Markov chain. We describe our experience in implementing this algorithm on a singleinstruction multipledata parallel processor computer. Our main conclusions are that a lowerlevel language has a performance advantage compared to Fortran, and that data storage is the limiting factor that determines the largest problem that can be solved. As a consequence, we devote considerable attention to storing a block tridiagonal transition matrix.INFORMS Journal on Computing. 01/1997; 9:218223. 
Conference Paper: A New Method for Analysing FeedbackBased Protocols with Applications to Engineering Web Traffic over the Internet.
[Show abstract] [Hide abstract]
ABSTRACT: Most of the studies of feedbackbased flow and congestion control consider only persistent sources which always have data to send. However, with the rapid growth of Internet applications built on TCP/IP such as the World Wide Web and the standardization of traffic management schemes such as Available Bit Rate (ABR) in Asynchronous Transfer Mode (ATM) networks, it is essential to evaluate the performance of feedbackbased protocols using traffic models which are specific to dominant applications. This paper presents a method for analysing feedbackbased protocols with a Webuserlike input traffic where the source alternates between "transfer" periods followed by "think" periods. Our key results, which are presented for the TCP protocol, are:(1) The goodputs and the fraction of time that the system has some given number of transferring sources are insensitive to the distributions of transfer (file or page) sizes and think times except through the ratio of their means. Thus, apart from network roundtrip times, only the ratio of average transfer sizes and think times of users need be known to size the network for achieving a specific quality of service.(2) The Engset model can be adapted to accurately compute goodputs for TCP and TCP over ATM, with different buffer management schemes. Though only these adaptations are given in the paper, the method based on the Engset model can be applied to analyze other feedback systems, such as ATM ABR, by finding a protocol specific adaptation. Hence, the method we develop is useful not only for analysing TCP using a source model significantly different from the commonly used persistent sources, but also can be useful for analysing other feedback schemes.(3) Comparisons of simulated TCP traffic to measured Ethernet traffic shows qualitatively similar autocorrelation when think times follow a Pareto distribution with infinite variance. Also, the simulated and measured traffic have long range dependence. In this sense our traffic model, which purports to be Webuserlike, also agrees with measured traffic.Proceedings of the 1997 ACM SIGMETRICS international conference on Measurement and modeling of computer systems; 01/1997  [Show abstract] [Hide abstract]
ABSTRACT: Intuition suggests that a finite buffer limits the effect of traffic autocorrelations on the queue length. We investigate the extent to which finite buffers can, therefore, be expected to mitigate the effects of longrange dependence (LRD). With traffic sequences generated by the fractional autoregressive integrated moving average (fARIMA) models for LRD, and by AR models for short range dependence (SRD), we investigate the traffic performance for a range of finite buffers, both for single and multiplexed streams. For design, the aim is to `match' a given LRD autocorrelation function with a suitable SRD function whose performance provides, for a wide range of traffic intensities and buffer sizes, a conservative bound on the performance associated with the LRD function. The results suggest that by suitably `dominating' an LRD autocorrelation function by an SRD function, one can obtain conservative performance bounds for a realistic range of traffic intensities and buffer sizes (delays). Also, in several cases, a `crossover' phenomenon is observed for cellloss probabilities as the buffer size increases, i.e., the loss probabilities are smaller for LRD than for SRD for small buffers, with the converse true for large buffers. This suggests that finite buffers, can in some cases, counteract the effects of LRD in traffic arrivals, and thus enable conservative designs to be based on Markovian traffic modelsGlobal Telecommunications Conference, 1996. GLOBECOM '96. 'Communications: The Key to Global Prosperity; 12/1996  [Show abstract] [Hide abstract]
ABSTRACT: The authors explore the influence of longrange dependence in broadband traffic engineering. The classification of stochastic processes {X<sub>t</sub>} into those with short or longrange dependence is based on the asymptotic properties of the variance of the sum S<sub>m </sub>=X<sub>1</sub>+X<sub>2</sub>+···+X<sub>m </sub>. Suppose this process describes the number of packets (or ATM cells) that arrive at a buffer; X<sub>t</sub> is the number that arrive in the tth time slice (e.g., 10 ms). We use a generic buffer model to show how the distribution of S<sub>m</sub> (for all values of m) determines the buffer occupancy. From this model we show that longrange dependence does not affect the buffer occupancy when the busy periods are not large. Numerical experiments show this property is present when data from four video conferences and two entertainment video sequences (which have longrange dependence) are used as the arrival process, even when the transmitting times are long enough to make the probability of buffer overflow 0.07. We generated sample paths from Markov chain models of the video traffic (these have shortrange dependence). Various operating characteristics, computed over a wide range of loadings, closely agree when the data trace and the Markov chain paths are used to drive the model. From this, we conclude that longrange dependence is not a crucial property in determining the buffer behavior of variable bit rate (VBR)video sourcesIEEE/ACM Transactions on Networking 07/1996; · 2.01 Impact Factor  [Show abstract] [Hide abstract]
ABSTRACT: Traffic from video services is expected to be a substantial portion of the traffic carried by emerging broadband integrated networks. For variable bit rate (VBR) coded video, statistical source models are needed to design networks that achieve acceptable picture quality at minimum cost and to design traffic shaping and control mechanisms. For video teleconference traffic Heyman et al. (1992) showed that traffic is sufficiently accurately characterized by a multistate Markov chain model that can be derived from three traffic parameters (mean, correlation, and variance). The present authors describe modeling results for sequences with frequent scene changes (the previously studied video teleconferences have very little scene variation) such as entertainment television, news, and sports broadcasts. The authors analyze 11 long sequences of broadcast video traffic data. Unlike video teleconferences, the different sequences studied have different details regarding distributions of cells per frame. The authors present source models applicable to the different sequences and evaluate their accuracy as predictors of cell losses in asynchronous transfer mode (ATM) networks. The modeling approach is the same for all of the sequences but use of a single model based on a few physically meaningful parameters and applicable to all sequences does not seem to be possibleIEEE/ACM Transactions on Networking 03/1996; · 2.01 Impact Factor  [Show abstract] [Hide abstract]
ABSTRACT: The main contributions of this paper are twofold. First, we prove fundamental, similarly behaving lower and upper bounds, and give an approximation based on the bounds, which is effective for analyzing ATM multiplexers, even when the traffic has many, possibly heterogeneous, sources and their models are of high dimension. Second, we apply our analytic approximation to statistical models of video teleconference traffic, obtain the multiplexing system's capacity as determined by the number of admissible sources for given cellloss probability, buffer size and trunk bandwidth, and, finally, compare with results from simulations, which are driven by actual data from coders. The results are surprisingly close. Our bounds are based on large deviations theory. The main assumption is that the sources are Markovian and timereversible. Our approximation to the steadystate buffer distribution is called Chenoffdominant eigenvalue since one parameter is obtained from Chernoffs theorem and the other is the system's dominant eigenvalue. Fast, effective techniques are given for their computation. In our application we process the output of variable bit rate coders to obtain DAR(1) source models which, while of high dimension, require only knowledge of the mean, variance, and correlation. We require cellloss probability not to exceed 10<sup>6 </sup>, trunk bandwidth ranges from 45 to 150 Mb/s, buffer sizes are such that maximum delays range from 1 to 60 ms, and the number of codersources ranges from 15 to 150. Even for the largest systems, the time for analysis is a fraction of a second, while each simulation takes many hours. Thus, the realtime administration of admission control based on our analytic techniques is feasibleIEEE Journal on Selected Areas in Communications 09/1995; · 3.12 Impact Factor 
Conference Paper: Fundamental Results on the Performance of ATM Multiplexers with Applications to Video Teleconferencing.
[Show abstract] [Hide abstract]
ABSTRACT: The main contributions of this paper are twofold. First, we prove fundamental, similarly behaving lower and upper bounds, and give an approximation based on the bounds, which is effective for analyzing ATM multiplexers, even when the traffic has many, possibly heterogeneous, sources and their models are of high dimension. Second, we apply our analytic approximation to statistical models of video teleconference traffic, obtain the multiplexing system's capacity as determined by the number of admissible sources for given cell loss probability, buffer size and trunk bandwidth, and, finally, compare with results from simulations, which are driven by actual data from coders. The results are surprisingly close. Our bounds are based on Large Deviations theory. Our approximation has two easily calculated parameters, one is from Chernoff's theorem and the other is the system's dominant eigenvalue. A broad range of systems are analyzed and the time for analysis in each case is a fraction of a second.Proceedings of the 1995 ACM SIGMETRICS joint international conference on Measurement and modeling of computer systems; 01/1995 
Conference Paper: Performance Impacts of SelfSimilarity in Traffic (Panel).
01/1995 
Conference Paper: Modeling teleconference traffic from VBR video coders
[Show abstract] [Hide abstract]
ABSTRACT: The authors analyze teleconference traffic, with moderate motion and scene changes, generated by different video codecs. These codecs differ in several aspects of coding, such as in the use of DCT and motion compensation. The results are that, even when traffic is generated using different coding schemes, the number of cells per frame can be described by a gamma (or equivalently negative binomial) distribution and a DAR(1) model determined by three traffic parameters (the mean, variance, and correlation) can be used to accurately model the source. The main contribution of the paper is in showing that the authors' previously published results on source modeling and marginal distributions, which were based on analysis of traffic generated by one type of coder, hold for coders which differ in various ways and particularly differ in the use of motion compensationCommunications, 1994. ICC '94, SUPERCOMM/ICC '94, Conference Record, 'Serving Humanity Through Communications.' IEEE International Conference on; 06/1994 
Conference Paper: Statistical analysis and simulation study of vbr broadcastvideo traffic
[Show abstract] [Hide abstract]
ABSTRACT: Not AvailableComputerAided Modeling, Analysis, and Design of Communication Links and Networks, 1994. (CAMAD '94) Fifth IEEE International Workshop on; 02/1994  [Show abstract] [Hide abstract]
ABSTRACT: Results are presented of a simulation study of the potential multiplexing gains from using variablebitrate (VBR) encoding to multiplex video teleconferencing traffic over an asynchronous transfer mode (ATM) network. Simulated traffic from several video teleconferences is fed into a buffer and multiplexed onto a higherspeed output line. The cell loss resulting from buffer overflow is observed. The major results are as follows: (1) the spacing between cells has a large role in determining the aggregate cell loss rate seen by a collection of calls. The cell loss rate is reduced when either the frame start times are evenly spaced or cells are evenly distributed in the video frame. (2) After the cell loss rate becomes nonnegligible, it grows rapidly as a function of the number of access lines. (3) The cell losses are not distributed uniformly over time but are clustered. A consequence is that the average time between clusters of cell loss and the time to first loss is greater than if the losses were spaced uniformly. (4) A simple model is found to describe when a cluster of cell losses occurs, the duration of the cluster, and how many cells are lost in the clusterIEEE Transactions on Circuits and Systems for Video Technology 01/1994; · 1.82 Impact Factor  IEEE Transactions on Circuits and Systems for Video Technology  TCSV. 01/1994;

Conference Paper: A simulation study of video teleconferencing traffic in ATM networks
[Show abstract] [Hide abstract]
ABSTRACT: Results of a simulation study of the potential multiplexing gains from using variablebitrate encoding to multiplex video teleconferencing traffic over an asynchronous transfer mode (ATM) network are presented. Simulated traffic from several video teleconferences was fed into a buffer and multiplexed onto a higher speed output line. The study observed the cell loss resulting from buffer overflow. The spacings between the start times for the individual video conferences are found to have a large role in determining the aggregate celllossrate seen by the collection of calls. After the celllossrate becomes nonnegligible, it grows rapidly as a function of the number of input lines. The cell losses are not distributed uniformly over time but are clustered, with the consequence that the probability that a call is loss free is much better than indicated by the celllossrateINFOCOM '93. Proceedings.Twelfth Annual Joint Conference of the IEEE Computer and Communications Societies. Networking: Foundation for the Future. IEEE; 02/1993  [Show abstract] [Hide abstract]
ABSTRACT: We consider the simulation of a discrete Markov chain that is so large that numerical solution of the steadystate balance equations cannot be done with available computers, We propose smoothing methods to obtain variance reduction when simulation is used to estimate a function of a subset of the steadystate probabilities. These methods attempt to make each transition provide information about the probabilities of interest. We give an algorithm that converges to the optimal smoothing operator, and some guidelines for picking the parameters of this algorithm. Analytical arguments are used to justify our procedures, and they are buttressed by the results of a numerical example,ACM Trans. Model. Comput. Simul. 01/1993; 3:167189. 
Conference Paper: An almost regenerative simulation of video teleconferencing traffic.
[Show abstract] [Hide abstract]
ABSTRACT: Not AvailableProceedings of the 25th Winter Simulation Conference, Los Angeles, California, USA, December 1215, 1993; 01/1993
Publication Stats
2k  Citations  
16.63  Total Impact Points  
Top Journals
Institutions

1997–2001

AT&T Labs
Austin, Texas, United States


1996

Texas A&M University
College Station, Texas, United States
