US20080043636A1 - Apparatus for size-dependent sampling for managing a data network - Google Patents

Apparatus for size-dependent sampling for managing a data network Download PDF

Info

Publication number
US20080043636A1
US20080043636A1 US11/924,329 US92432907A US2008043636A1 US 20080043636 A1 US20080043636 A1 US 20080043636A1 US 92432907 A US92432907 A US 92432907A US 2008043636 A1 US2008043636 A1 US 2008043636A1
Authority
US
United States
Prior art keywords
sampling
data
usage
network
charging
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/924,329
Inventor
Nicholas Duffield
Carsten Lund
Mikkel Thorup
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US11/924,329 priority Critical patent/US20080043636A1/en
Publication of US20080043636A1 publication Critical patent/US20080043636A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/14Charging, metering or billing arrangements for data wireline or wireless communications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0896Bandwidth or capacity management, i.e. automatically increasing or decreasing capacities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/50Network service management, e.g. ensuring proper service fulfilment according to agreements
    • H04L41/5003Managing SLA; Interaction between SLA and QoS
    • H04L41/5009Determining service level performance parameters or violations of service level contracts, e.g. violations of agreed response time or mean time between failures [MTBF]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/02Capturing of monitoring data
    • H04L43/022Capturing of monitoring data by sampling
    • H04L43/024Capturing of monitoring data by sampling by adaptive sampling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/02Capturing of monitoring data
    • H04L43/026Capturing of monitoring data using flow identification
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/02Capturing of monitoring data
    • H04L43/028Capturing of monitoring data by filtering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/16Threshold monitoring

Definitions

  • the present invention provides apparatus for sampling data flows in a data network.
  • Service providers of data networks are increasingly employing usage measurements as a component in customer charges.
  • One motivation stems from the coarse granularity in the available sizes of access ports into the network. For example, in the sequence of optical carrier transmission facilities OC-3 to OC-12 to OC-48 to OC-192, each port has a factor 4 greater capacity than the next smallest.
  • each port has a factor 4 greater capacity than the next smallest.
  • the service provider In order to manage a date network, the service provider typically determines customer usage at routers and other network elements in order to properly bill the customer.
  • One approach is to maintain byte or packet counters at a customer's access port(s). Such counters are currently very coarsely grained, giving aggregate counts in each direction across an interface over periods of a few minutes.
  • service quality may not be the sole determinant of customer charges. These could also depend, for example, on the remote (i.e. non-customer) IP address involved. This illustrates a broader point that the determinants of a charging scheme may be both numerous and also relatively dynamic. This observation may preclude using counts arising from a set of traffic filters, due to the requirement to have potentially a large number of such filters, and the administrative cost of configuring or reconfiguring such filters.
  • a complementary approach is to measure (or at least summarize) all traffic, and then transmit the measurements to a back-office system for interpretation according to the charging policy. In principle, this could be done by gathering packet headers, or by forming flow statistics.
  • An IP flow is a sequence of IP packets that shares a common property, as source or destination IP address or port number or combinations thereof.
  • a flow may be terminated by a timeout criterion, so that the interpacket time within the flow does not exceed some threshold, or a protocol-based criterion, e.g., by TCP FIN packet.
  • Flow collection schemes have been developed in research environments and have been the subject of standardization efforts. Cisco NetFlow is an operating system feature for the collection and export of flow statistics. These include the identifying property of the flow, its start and end time, the number of packets in the flow, and the total number of bytes of all packets in the flow.
  • the service provider of a data network also typically collects data regarding data usage over the data network as well as parts of the data network.
  • the collection of network usage data is essential for the engineering and management of communications networks.
  • the usage data provided by network elements has been coarse-grained, typically comprising aggregate byte and packet counts in each direction at a given interface, aggregated over time windows of a few minutes.
  • these data are no longer sufficient to engineer and manage networks that are moving beyond the undifferentiated service model of the best-effort Internet. Network operators need more finely differentiated information on the usage of their network.
  • Examples of such information include (i) the relative volumes of traffic using different protocols or applications; (ii) traffic matrices, i.e., the volumes of traffic originating from and/or destined to given ranges of Internet Protocol (IP) addresses or Autonomous Systems (AS); (iii) the time series of packet arrivals together with their IP headers; (iv) the durations of dial-user sessions at modern banks.
  • traffic matrices i.e., the volumes of traffic originating from and/or destined to given ranges of Internet Protocol (IP) addresses or Autonomous Systems (AS); (iii) the time series of packet arrivals together with their IP headers; (iv) the durations of dial-user sessions at modern banks.
  • IP Internet Protocol
  • AS Autonomous Systems
  • Such information can be used to support traffic engineering, network planning, peering policy, customer acquisition, marketing and network security.
  • An important application of traffic matrix estimation is to efficiently redirect traffic from overloaded links. Using this to tune OSPF/IS-IS routing one can typically accommodate
  • a data network With increasing data usage that is driven for the explosive demand for data services, a data network must support greater data traffic. Consequently, the data network must generate more data and associated messaging for managing the data network. A method that ameliorates the generation of management-related messaging and data while preserving the capabilities of managing the data network is therefore of great benefit to the industry.
  • the present invention provides apparatus for sampling data flows in a data network in order to estimate a total data volume in the data network. Sampling the data flows in the data network reduces the network resources that must be expended by the network in order to support the associated activities.
  • the present invention enables the service provider of the data network to control sampling volumes in relation to the desired accuracy.
  • sampling volume is defined as a number of objects selected as the result of sampling, e.g. during a sampling window. It may be a pure number, or may be expressed as a rate, i.e. number of objects per unit time.
  • the control can be either static or can be dynamic for cases in which the data volumes are changing as a function of time.
  • the present invention is not dependent upon the underlying statistical characteristics of the data flows.
  • the disclosure presents an exemplary embodiment with two variations.
  • the exemplary embodiment comprises a data network with a network of routers and dedicated hosts for managing the data network.
  • the first variation enables the service provider to charge a customer for usage of a data network.
  • the method utilizes the sampling of flows that are associated with the customer. The contribution to the usage by a sampled flow is normalized by a number that reflects the probability of sampling. The usage can be adjusted by the service provider in order to compensate for a possibility of overcharging.
  • the method enables the service provider to adjust the sampling rate and the billing period to reduce undercharging in accordance with the goals of the service provider.
  • the second variation enables the service provider to manage a data network in accordance with the measured traffic volume. The service provider can adjust the sampling volume in accordance with the measured sampling volume and with the desired accuracy for both static and dynamic situations.
  • FIG. 1 illustrates a data network utilizing size-dependent sampling, in accordance with the present invention
  • FIG. 2 shows a sampling probability function
  • FIG. 3 shows a complementary cumulative distribution (CCDF) of flow byte sizes
  • FIG. 4 shows a complementary cumulative distribution (CCDF) of bytes per customer-side IP addresses
  • FIG. 5 shows an example of weighted mean relative error vs. an effective sampling period
  • FIG. 6 shows an example of weighted mean relative error vs. an effective sampling period for different flow sizes
  • FIG. 7 is a flow diagram for charging with sampled network usage
  • FIG. 8 shows an example of traffic flow volumes in a data network
  • FIG. 9 shows static and dynamic controlled sampling volumes in relation to FIG. 8 ;
  • FIG. 10 is a flow diagram for controlling the sampling volume in a data network
  • FIG. 11 is a flow diagram for a quasi-random data sampling algorithm
  • FIG. 12 is a flow diagram for root finding algorithm
  • FIG. 13 shows an apparatus for managing a data network in accordance with the present invention.
  • One limitation to comprehensive direct measurement of traffic stems from the immense amounts of measurement data generated.
  • a single optical carrier transmission facility OC-48 at full utilization could generate about 100 GB of packet headers, or several GB of (raw) flow statistics each hour.
  • a common approach to dealing with large data volumes is to sample.
  • a common objection to sampling has been the potential for inaccuracy; customers can be expected to be resistant to being overcharged due to overestimation of the resources that they use.
  • FIG. 1 illustrates data network 100 that utilizes size-dependent sampling in accordance with the present invention.
  • data network 100 supports Internet Protocol (IP) messaging for the users (customers) of host 107 , 109 , 111 , 113 , 115 , 117 , and 119 .
  • IP Internet Protocol
  • a host includes PCs, workstations, mainframes, file servers, and other types of computers.
  • Hosts 107 , 109 , and 111 are configured on data link 121 ; hosts 113 and 115 are on data link 123 ; and hosts 117 and 119 are configured on data link 125 .
  • hosts 117 and 119 are configured on data link 125 .
  • a host e.g.
  • IP messaging is routed through routers 101 , 103 , and 105 and data links 151 , 153 , and 155 .
  • a router forwards packets between hosts and other routers in order to enable hosts not connected to the same link to communicate.
  • Each router has a port (e.g. 127 , 129 , and 131 for router 101 that connects to a data link.
  • ports 133 , 135 , and 137 correspond to router 103
  • ports 137 , 139 , and 141 correspond to router 105 .
  • host 113 supports the billing (charging) of customers and host 115 supports the collection and the utilization of data traffic information regarding data transmission for data network 100 .
  • Hosts 107 , 109 , 111 , 117 , and 119 correspond to users who are the customers for the service provider operating data network 100 .
  • Managing functions that are associated with the billing of customers and the traffic management support the collection of relevant information for the management of data network 100 .
  • “managing a network” denotes the determination of one or more characteristics of the configuration, state, and/or usage of the network and its management subsystems.
  • Host 113 collects information from routers 101 , 103 , and 105 about customer usage.
  • Host 115 collects information about data traffic over the data links. With a typical data network, many flows are transported over the data network. Generating management-related messages to hosts 113 and 115 can be very detrimental to the performance of data network 100 if the number of messages is large. Thus, sampling is supported by data network 100 in order to reduce the number of management-related messages to reduce any associated performance degradations.
  • a router collects information about data traffic over data links through the router. Utilizing the information, the router can adjust its configuration for the current data traffic.
  • the present invention provides a sampling mechanism that specifically addresses concerns of sampling error.
  • Total customer usage is the sum of a number of components, some large, some small. Sampling errors arise predominantly from omission of the larger components, whereas accuracy is less sensitive to omission of the smaller components.
  • accuracy is less sensitive to omission of the smaller components.
  • a simple sampling scheme in which one estimates the total bytes of usage by sampling 1 in every N flows, and then adds together N times the total bytes reported in each sampled flow.
  • the underlying distribution of flow bytes sizes has been found to follow a heavy tailed distribution. In this case, the estimate can be extremely sensitive to the omission or inclusion of the larger flows.
  • such an estimator can have high variance due to the sampling procedure itself. (In the disclosure, the term “flow” is used synonymously with the term “object.”)
  • a flow comprises at least one unit of data (e.g. packet, byte, octet, and ATM cell).
  • the present invention reduces sampling volumes for data network 100 .
  • a heavy-tailed distribution of flow sizes can be turned to an advantage for sampling provided an appropriate sampling algorithm is used.
  • the present invention utilizes size-dependent sampling, in which an object of size x is selected with some size dependent probability p(x).
  • the probability p(x) is 1 for large x. In the case of flows, all sufficiently large flows will always be selected; there is no sampling error for such flows. On the other hand one can have p(x) ⁇ 1 for smaller flows; this reduces the number of samples, but the error involved is small since the underlying flows are small.
  • To estimate the total bytes represented in the original set of flows one sums the quantities x/p(x) over only the sampled flows.
  • routers 101 , 103 , and 105 sample objects (flows). Consequently, routers 101 , 103 , and 105 send messages (“samples”) to host 113 (corresponding to charging the customer for usage) and to host 115 (corresponding to determining traffic volume) comprising attributes of the sampled object (e.g. the customer's identification) and the size of the object.
  • hosts corresponding to charging the customer for usage
  • host 115 corresponding to determining traffic volume
  • Alternative embodiments may store corresponding data at routers 101 , 103 , and 105 for later processing or may utilize intervening processors between routers 101 , 103 , and 105 .
  • Size-dependent sampling has a number of advantages.
  • the sampling probabilities p(x) can be chosen to satisfy a certain optimality criterion for estimator variance as described later.
  • a simple adaptive scheme allows dynamic tuning of p(x) in order to keep the total number of samples within a given bound.
  • the number of flow statistics that are transmitted to the back-end system can be controlled by the service provider.
  • p(x) can vary at different times and across different regions of the network (as needed), but estimation remains unbiased.
  • sampling is composable in the sense that the first three properties above are preserved under successive resampling.
  • size-dependent sampling is applicable to packet sampling as well.
  • performance benefit over 1 in N sampling is smaller in this case, since packet sizes do not have a heavy-tailed distribution.
  • the present invention utilizes an approach to usage-sensitive charging that mirrors the foregoing approach to sampling.
  • the sampling scheme determines the size of the larger flows with no error. Estimation error arises entirely from sampling smaller flows.
  • For billing purposes we wish to measure the total bytes for each billed entity (e.g. for each customer at a given service level) over each billing cycle. Larger totals have a smaller associated sampling error, whereas estimation of total bytes for the smallest customers may be subject to greater error. Therefore, the service provider sets a level L on the total bytes, with a fixed charge for all usage up to L, then a usage sensitive charge for all usage above L. Thus, the service provider only needs to tune the sampling scheme for estimating the usage above L within the desired accuracy.
  • the present invention utilizes sampling as a means to reduce data volume while at the same time obtaining a representative view of the raw data.
  • An elementary way to do this is to sample 1 in N raw data objects, either independently (i.e. each object is selected independently with probability 1/N) or deterministically (objects N, 2N, 3N, . . . are selected and all others are discarded). Only those selected objects are used further for analysis.
  • This sampling strategy clearly reduces the load associated with the subsequent transmission, storage, and analysis of the data by a factor N.
  • sampling parameters need to be bounded to the sampled data in order that extensive properties of the original data stream can be estimated. For example, to estimate the bytes rate in a raw packet stream from samples gathered through 1 in N sampling, one needs to multiply the byte rate of the sampled stream by N. Under a given constraint on resources available for measurement transmission or processing of data, N may vary both temporally and spatially according to traffic volumes. Hence, N is not typically a global variable independent of the raw data.
  • Let x i be the size attribute of interest from the flow i, e.g., the number of packets in the flow, or the total number of bytes in the flow, or any other positive quantity of interest.
  • Each packet in a flow possesses a common attribute, such as IP address (or net), port number, or Type of Service (ToS) field.
  • IP address or net
  • ToS Type of Service
  • Each combination of interest corresponds to attributes as referred by a “color”; c i will be the color of flow i.
  • a color might correspond to a customer address, or this plus a remote network, and possibly a ToS specification.
  • the mapping that associates a particular customer with a set of packet attributes may be relatively complex;. This to be performed by the subsystem that collects and interprets the measurements (e.g. hosts 113 and 115 in the exemplary embodiment). The objective is to estimate the totals for each color c of interest as follows.
  • X ⁇ ( c ) ⁇ c ⁇ x i ( 1 )
  • the present invention supports the sampling of raw packet headers, the set of flow statistics formed from the sampled packets, the stream of flow statistics at some intermediate aggregation point, and the set of aggregate flows at the collection point.
  • the knowledge of the number n of original objects in not required.
  • sampling itself need not make reference to the object color c. This reflects the fact that the colors of interest may not be known at the time of sampling and that it is infeasible to simply accumulate sizes from the original stream for all possible colors.
  • p z (x) min ⁇ 1,x/z ⁇ .
  • a flow with size x is sampled with probability p z (x).
  • the parameter z acts as a threshold: flow of size z or above are always sampled as shown in FIG. 2 .
  • the horizontal axis corresponds to x i (the size of an object 201 ).
  • the parameter z is an example of a “probabilistic parameter.”
  • the statistical variability of the estimate of X(c) provides a measure of confidence of the estimate.
  • the present invention enables the service provider to “tune” the operation of data network 100 in order to achieve the desired accuracy.
  • p z (x i ) is optimal in the sense that Var ⁇ circumflex over (X) ⁇ (c)+z 2 E(N(c)) is minimized with p z (x i ), where E(N(c)) is the expected value of N(c).
  • the disclosure provides a method for controlling the statistical variance based upon operating parameters that the service provider can control.
  • Parameter z is the size threshold above which flows are always sampled.
  • FIG. 3 displays an exemplary complementary cumulative distribution function (CCDF), i.e. the proportion of flows with bytes greater than a given level, of the flow sizes encountered by data network 100 .
  • CCDF complementary cumulative distribution function
  • the approximate linearity on the log-log scale is indicative of a heavy tailed distribution.
  • the distribution of total bytes per customer-side IP address over a given period shares the heavy tailed property as shown in FIG. 4 .
  • WMRE Weighted Mean Relative Error
  • FIG. 5 illustrates an example of substantially better accuracy (smaller WMRE) of optimal sampling as compared with 1 in N sampling, over 4 orders of magnitude of the sampling period.
  • Curve 501 illustrates the relationship for WMRE as a function of the effective sampling period, while curve 503 shows the corresponding relationship with sampling as described in the disclosure.
  • the WMRE for optimal sampling is about only 1%, while for 1 in N sampling it is around 50%.
  • the irregularity of the upper line reflects the sensitivity of the estimates from 1 in N sampling to random inclusion or exclusion of the largest flows during sampling.
  • FIG. 6 displays with WMRE vs. sampling period for a trace of 10 7 flows (corresponding to curve 605 ), as compared with subportions contain 10 6 (corresponding to curve 603 ) and 10 5 (corresponding to curve 601 ) flows.
  • the relative error decrease as the trace length increases, since the byte total for a given IP address is composed of a greater number of contributions. It may be desirable to place lower bounds on z in order to fulfill other objectives, such as limiting the rate at which samples are generated.
  • the behavior from FIG. 6 suggests that is possible to simultaneously fulfill the goal of low relative error provided that the length of the period of observation (e.g. the billing period) is sufficiently long.
  • the exemplary embodiment utilizes the disclosed sampling techniques for charging the customer of data network 100 for usage.
  • Fair charging requires that the deviation between the traffic charged to a customer and the actual traffic be kept to a minimum.
  • the scheme is essentially the best possible, in the sense that variance of X is minimized for a given threshold z.
  • the relative estimation error can be relatively large for colors with small amounts of traffic.
  • ⁇ circumflex over (X) ⁇ (c) that is either 0 (if the flow is not sampled), or z (if it is sampled, wherein the sample is normalized by p z (x)).
  • ⁇ circumflex over (X) ⁇ (c) will be either 0, or at least z.
  • the service provider can charge the traffic of a given color at a fixed fee, plus a usage-sensitive charge only for traffic volumes that exceed a certain level L. (L may depend on the color in question).
  • L may depend on the color in question.
  • the idea is to tune the sampling algorithms so that any usage X(c) that exceeds L can be reliably estimated. Usage X(c) that falls below L does not need to be reliably estimated, since the associated charge is usage-insensitive, i.e., independent of ⁇ circumflex over (X) ⁇ (c) ⁇ L.
  • f c ( ⁇ circumflex over (X) ⁇ (c)) which depends on ⁇ circumflex over (X) ⁇ (c) only through the quantity max ⁇ circumflex over (X) ⁇ (c), L ⁇ , i.e., it is independent of any usage below L.
  • the subscript of f c indicates that the charge may depend on the color c, e.g., through the type of service, or foreign IP address.
  • Equation 4 can also express pricing models in which there is a fixed administrative charge for small customers, whose usage doesn't warrant accurate measurement. Both a c and b c are allowed to depend on the color c in question.
  • Reliable estimation of the volumes X(c) is determined by choosing the sampling threshold z appropriately high for level L in question.
  • the typical estimation error is no more than about ⁇ times X, for some target ⁇ >0. This can be expressed this as the following standard error condition: ⁇ ( ⁇ circumflex over (X) ⁇ ( c )) ⁇ X ( c ) if X ( c )> L (5)
  • Equation 5 the probability of overestimating ⁇ circumflex over (X) ⁇ (c)>L by an amount ⁇ X(c) (i.e., by ⁇ / ⁇ standard deviations) is no more than ⁇ ( ⁇ / ⁇ ), where ⁇ is the standard normal distribution function.
  • the above approach sets limits on the chance that the deviation of the estimated usage above the actual usage exceeds a given amount.
  • a refinement allows the service provider to set a limit on the chance that overcharging occurs. This should be more attractive from the customer's point of view since the chance of the customer being over billed at all can be small. Conversely, the service provider has to accept a small persistent under billing in order to accommodate the potential sampling error.
  • ⁇ ( ⁇ s) is about 0.13%, i.e. about 1 in 740 traffic volumes will be overestimated.
  • the service provider may charge according to ⁇ circumflex over (X) ⁇ ′(c) rather than ⁇ circumflex over (X) ⁇ (c).
  • the customer is billed f c ( ⁇ circumflex over (X) ⁇ ′(c)).
  • the chance that the customer is over billed is approximately equal to ⁇ ( ⁇ s).
  • the service provider determines s 2 z ⁇ 2 L.
  • the condition z* ⁇ z o can be realized by increasing the length of the billing cycle.
  • the thresholds z o and z* control phenomena at different timescales.
  • z* controls the average rate at which samples are taken.
  • z o controls the sampling variance of the estimates ⁇ circumflex over (X) ⁇ (c) of total bytes over the billing timescale, potentially over days, weeks, or even months.
  • the level byte L (under which accurate measurements are not needed) can be chosen to increase with the billing timescale. For example, the service provider may choose L to correspond to a particular quartile of the distribution of byte size, so that only a given proportion of the total bytes transmitted on the network are generated by customers whose total usage does not exceed L during the billing cycle.
  • FIG. 6 shows that the relative error in estimation decreases as the duration of collection of the flow trace increases.
  • FIG. 7 is a flow diagram for charging with sampled network usage in accordance with the exemplary embodiment.
  • threshold z is determined according to the relative error and the unbillable usage.
  • it is determined whether to sample an object that is associated with the customer in accordance with the size of the object and the probabilistic function p z (x). (The discussion with respect to FIG. 11 presents a method for determining whether to sample the objects based upon the size.)
  • the associated usage is determined from the size of the sampled object by dividing the size by the p z (x i ) in step 705 .
  • the usage sensitive pricing for the customer is calculated in step 709 in accordance with Equation 4 and adjustments by the service provider.
  • the usage is reset to 0 so that the usage for the next billing period can be calculated.
  • the present invention also enables the service provider to control the sample volume that is generated by data network 100 .
  • the amount of data is dynamic with time, and consequently the sampling rate needs to adjust accordingly. In other words, dynamic control of the mean sample volume may be needed.
  • An object (flow) may be distinguishable by an attribute.
  • Each object is characterized by a size that may be expressed in a number of packets, bytes (octets), or ATM cells contained in the object. The number is equal to at least one.) In such a case, the object is characterized as being colored.
  • the present invention allows the service provider to estimate the total size of the objects in each color class c.
  • the objective (cost) function C z (p) is minimized locally over each color class.
  • the sampling device does not distinguish colors; however, samples can be later analyzed with respect to any combination of colors.
  • the volume of objects presented for sampling will generally vary with time.
  • a mechanism to control the number of samples must be able to adapt to temporal variations in the rate at which objects are offered for sampling. This is already an issue for the 1 in N sampling algorithm, since it may be necessary to adjust N, both between devices and at different times in a single device, in order to control the sampled volumes.
  • the service provider can control the volume by an appropriate choice of the threshold z.
  • a direct approach to finding z* is to construct an algorithm to find the root, utilizing a set of x i (sizes of the sampled objects).
  • FIG. 12 which is discussed later, illustrates the approach utilized in the exemplary embodiment.
  • the service provider can dynamically adapt (i.e. updating) z knowing only the target and current sample volumes.
  • “sampling window” is defined as being an interval during which objects are presented for sampling. The interval may be measured in time, e.g., in online applications where each object occurs at some time during the window.
  • the objects In offline applications, the objects have already been collected, and are then sampled offline.
  • the interval might be measured in time, i.e. objects collected in a particular time window are presented for sampling, or in number where a certain number of objects are presented for sampling.
  • the endpoint of the window may be determined prior to sampling, or it may depend on the objects, e.g. through the number that are sampled and/or their sizes.
  • FIG. 8 shows an example of traffic flow volumes in data network 100 .
  • data network 100 incurs a sudden increase of the traffic volume.
  • FIG. 9 shows static (curve 901 ) and dynamic controlled (curve 903 ) sampling volumes in relation to FIG. 8 .
  • threshold z By adjusting threshold z, the sampling volume remains substantially constant relative to the sampling volume corresponding to a fixed threshold z.
  • the exemplary embodiment enables the service provider to execute immediate corrective measures.
  • the measured sampling volume N may significantly exceed the target M before the end of the sampling window.
  • the service provider should immediately change the threshold z.
  • the windowing mechanism is a timeout that takes effect if N has not exceeded M by the end of the window.
  • the corresponding emergency control can use timing information. If N already exceeds M at time t from a start of a window of length T, z is immediately replace by zT/t.
  • the a new sampling window can be started at that time. Otherwise, from time t one can reaccumulate the sample count N from zero, and the test and remedy procedure is repeated as needed for the remainder of the sampling window.
  • FIG. 10 is a flow diagram for controlling the sampling volume in data network 100 .
  • the value of z is determined in accordance with a targeted sample volume.
  • FIG. 12 provides a method for determining z.
  • step 1003 it is determined whether to sample the ith object having a size x i .
  • FIG. 11 provides a method for determining whether to sample the ith object.
  • the corresponding traffic volume is estimated by normalizing x i by p z (x i ) and accumulated to the estimated traffic volume in step 1005 .
  • the sampling volume is estimated in step 1009 for data network 100 . If the sampling window is not completed, then step 1003 is repeated.
  • the estimated sampling volume is utilized by the service provider in maintaining data network 100 .
  • the service provider can reconfigure the data network to be better matched to the traffic volume.
  • FIG. 11 is a flow diagram for a quasi-random data sampling algorithm The process shown in FIG. 11 can be utilized by step 703 or by step 1003 in determining whether to sample an object (flow).
  • the variable “count” has a uniformly distributed value between 0 and z-1.
  • count is reset to zero.
  • step 1103 the size of the object x i is compared to z. If x i is greater or equal to z, then the ith object is sampled in sample 1105 . The index i is incremented by 1 in step 1107 so that the next object is considered in the next execution of step 1103 .
  • step 1103 if x i is less than z in step 1103 , then count is incremented by x i in step 1109 . If count is greater than or equal to z in step 1111 , count is decremented by z in step 1113 and the ith object is sampled in step 1115 . However, if count is less than z in step 1111 , index i is incremented by 1 in step 1107 so that the next object is considered for the next execution of step 1103 .
  • FIG. 11 is one embodiment of a quasi-random data sampling algorithm.
  • One skilled in the art appreciates that other quasi-random embodiments can be utilized in order to determine whether to sample an object.
  • FIG. 12 is a flow diagram for root finding algorithm that may be utilized in determining or updating z in step 701 ( FIG. 7 ) or step 1001 ( FIG. 10 ). Discussion of FIG. 12 is facilitated with specific notations as follow.
  • the approach of the process shown in FIG. 12 is to select a candidate z and to determine if the candidate z is satisfactory, too large, or too small.
  • the process utilizes a collection of numbers corresponding to the sizes of previously sampled objects and the target sampling volume M.
  • the process as illustrated in FIG. 12 does not change the value of M. Rather, variables M, B, and C are internal variables that are used for calculations. The process only returns the appropriate value of z in steps 1205 and 1219 .
  • step 1201 M and ⁇ X ⁇ are inputted. Internal variable B is reset to zero.
  • step 1209 z is randomly selected from ⁇ X ⁇ .
  • An efficient implementation may require that z be picked randomly from ⁇ X ⁇ so that the expectation is somewhere in the middle with respect to size. However, assuming that the order of the members from ⁇ X ⁇ is independent of size, one can let z be equal to the first member in ⁇ X ⁇ .
  • set ⁇ Y ⁇ consists of the members of ⁇ X ⁇ whose values are less than z.
  • N (B+C)/z+
  • are equal to the number of elements contained in ⁇ X ⁇ and ⁇ Y ⁇ , respectively.
  • FIG. 13 shows an apparatus 1300 for managing a data network in accordance with the present invention.
  • Apparatus 1300 receives and sends packets that are transported by a data network through packet interface 1303 .
  • Processor 1301 receives packets containing traffic information through link 1302 from packet interface 1303 .
  • apparatus 1300 provides router functionality with routing module 1305 .
  • Routing module 1305 directs packets between packet interface 1303 and packet interface 1307 through links 1304 and 1306 and between packet interface 1303 and packet interface 1309 through links 1304 and 1308 .
  • Processor 1301 configures routing module 1305 through link 1310 in accordance with the traffic information that is received through link 1302 .
  • Processor 1301 executes computer instructions corresponding to the flow diagrams shown in FIGS. 7, 10 , 11 , and 12 .
  • processor 1301 stores traffic information (that is received through link 1302 ) to storage device 1311 through link 1312 for later retrieval (e.g. billing information for charging a customer).
  • the computer system may include at least one computer such as a microprocessor, digital signal processor, and associated peripheral electronic circuitry.

Abstract

The present invention provides apparatus for sampling data flows in a data network in order to estimate a total data volume in the network. Sampling the data flows in the data network reduces the network resources that must be expended by the network to support the associated activity. The present invention enables the service provider of the data network to control sampled volumes in relation to the desired accuracy. The control can be either static or can be dynamic for cases in which the data volumes are changing as a function of time.

Description

  • This application is a continuation of prior application Ser. No. 11/478,960, filed Jun. 27, 2006, which is a continuation of prior application Ser. No. 10/056,683, filed Jan. 24, 2002 now issued as U.S. Pat. No. 7,080,136, which claims priority to provisional U.S. Application No. 60/277,123 (“Control Of Volume And Variance In Network Management”), filed Mar. 19, 2001 and provisional U.S. Application No. 60/300,587 (“Charging from Sampled Network Usage”), filed Jun. 22, 2001. Each of the above cited applications is herein incorporated by reference in its entirety.
  • FIELD OF THE INVENTION
  • The present invention provides apparatus for sampling data flows in a data network.
  • BACKGROUND OF THE INVENTION
  • Service providers of data networks are increasingly employing usage measurements as a component in customer charges. One motivation stems from the coarse granularity in the available sizes of access ports into the network. For example, in the sequence of optical carrier transmission facilities OC-3 to OC-12 to OC-48 to OC-192, each port has a factor 4 greater capacity than the next smallest. Consider a customer charged only according to the access port size. If customer's demand is at the upper end of the capacity of its current port, the customer will experience a sharp increase in charges on moving to the next size up. Moreover, much of the additional resources will not be used, at least initially. Usage based charging can avoid such sharp increases by charging customers for the bandwidth resources that they consume. Another motivation for usage-based charging stems from the fact that in IP networks the bandwidth beyond the access point is typically a shared resource. Customers who are aware of the charges incurred by bandwidth usage have a greater incentive to moderate that usage. Thus, charging can act as a feedback mechanism that discourages customers from attempting to fill the network with their own traffic to the detriment of other customers. Finally, differentiated service quality requires correspondingly differentiated charges. In particular, it is expected that premium services will be charged on a per use basis, even if best effort services remain on a flat (i.e. usage insensitive) fee.
  • In order to manage a date network, the service provider typically determines customer usage at routers and other network elements in order to properly bill the customer. One approach is to maintain byte or packet counters at a customer's access port(s). Such counters are currently very coarsely grained, giving aggregate counts in each direction across an interface over periods of a few minutes. However, even separate counters differentiated by service quality would not suffice for all charging schemes. This is because service quality may not be the sole determinant of customer charges. These could also depend, for example, on the remote (i.e. non-customer) IP address involved. This illustrates a broader point that the determinants of a charging scheme may be both numerous and also relatively dynamic. This observation may preclude using counts arising from a set of traffic filters, due to the requirement to have potentially a large number of such filters, and the administrative cost of configuring or reconfiguring such filters.
  • A complementary approach is to measure (or at least summarize) all traffic, and then transmit the measurements to a back-office system for interpretation according to the charging policy. In principle, this could be done by gathering packet headers, or by forming flow statistics. An IP flow is a sequence of IP packets that shares a common property, as source or destination IP address or port number or combinations thereof. A flow may be terminated by a timeout criterion, so that the interpacket time within the flow does not exceed some threshold, or a protocol-based criterion, e.g., by TCP FIN packet. Flow collection schemes have been developed in research environments and have been the subject of standardization efforts. Cisco NetFlow is an operating system feature for the collection and export of flow statistics. These include the identifying property of the flow, its start and end time, the number of packets in the flow, and the total number of bytes of all packets in the flow.
  • The service provider of a data network also typically collects data regarding data usage over the data network as well as parts of the data network. The collection of network usage data is essential for the engineering and management of communications networks. Until recently, the usage data provided by network elements has been coarse-grained, typically comprising aggregate byte and packet counts in each direction at a given interface, aggregated over time windows of a few minutes. However, these data are no longer sufficient to engineer and manage networks that are moving beyond the undifferentiated service model of the best-effort Internet. Network operators need more finely differentiated information on the usage of their network. Examples of such information include (i) the relative volumes of traffic using different protocols or applications; (ii) traffic matrices, i.e., the volumes of traffic originating from and/or destined to given ranges of Internet Protocol (IP) addresses or Autonomous Systems (AS); (iii) the time series of packet arrivals together with their IP headers; (iv) the durations of dial-user sessions at modern banks. Such information can be used to support traffic engineering, network planning, peering policy, customer acquisition, marketing and network security. An important application of traffic matrix estimation is to efficiently redirect traffic from overloaded links. Using this to tune OSPF/IS-IS routing one can typically accommodate 50% more demand.
  • Concomitant with the increase in detail in the information to be gathered is an increase in its traffic volume. This is most noticeable for traffic data gathered passively, either by packet monitors gathering IP packet header traces or IP flow statistics. As an example, a single OC-48 at full utilization may yield as much as 70 GB of IP packet headers or 3 GB of flow statistics per hour. The volume of data exported for further analysis may be potentially decreased at the measurement point through either filtering or aggregation. Neither of these approaches may be appropriate for all purposes. Filtering allows us to restrict attention to a particular subset of data, e.g., all traffic to or from a pre-determined range of IP addresses of interest. However, not all questions can be answered in such a manner. For example, in determining the most popular destination web site for traffic on a given link, one generally does not know in advance which address or address ranges to look for. On the other hand, aggregation and other forms of analysis at the measurement site have two disadvantages. First, the time-scale to implement and modify such features in network elements are very long, typically a small number of years. Second, the absence of raw measured data would limit exploratory studies of network traffic.
  • With increasing data usage that is driven for the explosive demand for data services, a data network must support greater data traffic. Consequently, the data network must generate more data and associated messaging for managing the data network. A method that ameliorates the generation of management-related messaging and data while preserving the capabilities of managing the data network is therefore of great benefit to the industry.
  • BRIEF SUMMARY OF THE INVENTION
  • The present invention provides apparatus for sampling data flows in a data network in order to estimate a total data volume in the data network. Sampling the data flows in the data network reduces the network resources that must be expended by the network in order to support the associated activities. The present invention enables the service provider of the data network to control sampling volumes in relation to the desired accuracy. (In the disclosure “sampling volume” is defined as a number of objects selected as the result of sampling, e.g. during a sampling window. It may be a pure number, or may be expressed as a rate, i.e. number of objects per unit time.) The control can be either static or can be dynamic for cases in which the data volumes are changing as a function of time. Moreover, the present invention is not dependent upon the underlying statistical characteristics of the data flows.
  • The disclosure presents an exemplary embodiment with two variations. The exemplary embodiment comprises a data network with a network of routers and dedicated hosts for managing the data network. The first variation enables the service provider to charge a customer for usage of a data network. The method utilizes the sampling of flows that are associated with the customer. The contribution to the usage by a sampled flow is normalized by a number that reflects the probability of sampling. The usage can be adjusted by the service provider in order to compensate for a possibility of overcharging. In addition, the method enables the service provider to adjust the sampling rate and the billing period to reduce undercharging in accordance with the goals of the service provider. The second variation enables the service provider to manage a data network in accordance with the measured traffic volume. The service provider can adjust the sampling volume in accordance with the measured sampling volume and with the desired accuracy for both static and dynamic situations.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a data network utilizing size-dependent sampling, in accordance with the present invention;
  • FIG. 2 shows a sampling probability function;
  • FIG. 3 shows a complementary cumulative distribution (CCDF) of flow byte sizes;
  • FIG. 4 shows a complementary cumulative distribution (CCDF) of bytes per customer-side IP addresses;
  • FIG. 5 shows an example of weighted mean relative error vs. an effective sampling period;
  • FIG. 6 shows an example of weighted mean relative error vs. an effective sampling period for different flow sizes;
  • FIG. 7 is a flow diagram for charging with sampled network usage;
  • FIG. 8 shows an example of traffic flow volumes in a data network;
  • FIG. 9 shows static and dynamic controlled sampling volumes in relation to FIG. 8;
  • FIG. 10 is a flow diagram for controlling the sampling volume in a data network;
  • FIG. 11 is a flow diagram for a quasi-random data sampling algorithm;
  • FIG. 12 is a flow diagram for root finding algorithm; and
  • FIG. 13 shows an apparatus for managing a data network in accordance with the present invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • One limitation to comprehensive direct measurement of traffic stems from the immense amounts of measurement data generated. For example, a single optical carrier transmission facility OC-48 at full utilization could generate about 100 GB of packet headers, or several GB of (raw) flow statistics each hour. The demands on computational resources at the measurement point, transmission bandwidth for measured data, and back-end systems for storage and analysis of data, all increase costs for the service provider.
  • A common approach to dealing with large data volumes is to sample. A common objection to sampling has been the potential for inaccuracy; customers can be expected to be resistant to being overcharged due to overestimation of the resources that they use.
  • FIG. 1 illustrates data network 100 that utilizes size-dependent sampling in accordance with the present invention. In the exemplary embodiment of the invention, data network 100 supports Internet Protocol (IP) messaging for the users (customers) of host 107, 109, 111, 113, 115, 117, and 119. (However, variations of the embodiment can support other data protocols.) A host includes PCs, workstations, mainframes, file servers, and other types of computers. Hosts 107, 109, and 111 are configured on data link 121; hosts 113 and 115 are on data link 123; and hosts 117 and 119 are configured on data link 125. In order for a host (e.g. host 111) to communicate to another host (e.g. host 119) on a different data link, IP messaging is routed through routers 101, 103, and 105 and data links 151, 153, and 155. A router forwards packets between hosts and other routers in order to enable hosts not connected to the same link to communicate. Each router has a port (e.g. 127, 129, and 131 for router 101 that connects to a data link. Similarly, ports 133, 135, and 137 correspond to router 103, and ports 137, 139, and 141 correspond to router 105.
  • In the exemplary embodiment, host 113 supports the billing (charging) of customers and host 115 supports the collection and the utilization of data traffic information regarding data transmission for data network 100. Hosts 107, 109, 111, 117, and 119 correspond to users who are the customers for the service provider operating data network 100. Managing functions that are associated with the billing of customers and the traffic management support the collection of relevant information for the management of data network 100. (In the disclosure, “managing a network” denotes the determination of one or more characteristics of the configuration, state, and/or usage of the network and its management subsystems. The characteristics are then reported for subsequent activities such as billing or marketing, and/or using them to assist reconfigure and/or reengineer the network and its management subsystems.) Host 113 collects information from routers 101, 103, and 105 about customer usage. Host 115 collects information about data traffic over the data links. With a typical data network, many flows are transported over the data network. Generating management-related messages to hosts 113 and 115 can be very detrimental to the performance of data network 100 if the number of messages is large. Thus, sampling is supported by data network 100 in order to reduce the number of management-related messages to reduce any associated performance degradations.
  • With alternative embodiments, a router (e.g. 101, 103, and 105) collects information about data traffic over data links through the router. Utilizing the information, the router can adjust its configuration for the current data traffic.
  • The present invention provides a sampling mechanism that specifically addresses concerns of sampling error. Total customer usage is the sum of a number of components, some large, some small. Sampling errors arise predominantly from omission of the larger components, whereas accuracy is less sensitive to omission of the smaller components. For example, consider a simple sampling scheme in which one estimates the total bytes of usage by sampling 1 in every N flows, and then adds together N times the total bytes reported in each sampled flow. The underlying distribution of flow bytes sizes has been found to follow a heavy tailed distribution. In this case, the estimate can be extremely sensitive to the omission or inclusion of the larger flows. Generally, such an estimator can have high variance due to the sampling procedure itself. (In the disclosure, the term “flow” is used synonymously with the term “object.”)
  • The present invention does not require any knowledge of the underlying statistical information of the data traffic for data network 100. For example, the associated probability relating to the size of a flow can assume any form, including a heavy-tailed probability distribution. A flow (object) comprises at least one unit of data (e.g. packet, byte, octet, and ATM cell).
  • Additionally, the present invention reduces sampling volumes for data network 100. A heavy-tailed distribution of flow sizes can be turned to an advantage for sampling provided an appropriate sampling algorithm is used. The present invention utilizes size-dependent sampling, in which an object of size x is selected with some size dependent probability p(x). The probability p(x) is 1 for large x. In the case of flows, all sufficiently large flows will always be selected; there is no sampling error for such flows. On the other hand one can have p(x)<1 for smaller flows; this reduces the number of samples, but the error involved is small since the underlying flows are small. To estimate the total bytes represented in the original set of flows, one sums the quantities x/p(x) over only the sampled flows. Applying the renormalization factor 1/p(x) to the small flows compensates for the fact that that might have been omitted. In fact, it can be shown that this sum is an unbiased estimator of the actual total bytes (i.e. its average value over all possible random samplings is equal to the actual total bytes. Moreover, uniform sampling is a special case of this scheme with p(x) constant and equal to 1/N.)
  • With the exemplary embodiment of the invention (as shown as data network 100 in FIG. 1), routers 101, 103, and 105 sample objects (flows). Consequently, routers 101, 103, and 105 send messages (“samples”) to host 113 (corresponding to charging the customer for usage) and to host 115 (corresponding to determining traffic volume) comprising attributes of the sampled object (e.g. the customer's identification) and the size of the object. Alternative embodiments may store corresponding data at routers 101, 103, and 105 for later processing or may utilize intervening processors between routers 101, 103, and 105.
  • Size-dependent sampling has a number of advantages. First, the sampling probabilities p(x) can be chosen to satisfy a certain optimality criterion for estimator variance as described later. Second, a simple adaptive scheme allows dynamic tuning of p(x) in order to keep the total number of samples within a given bound. Thus, in the context of flow measurement, the number of flow statistics that are transmitted to the back-end system (host 113 and host 115) can be controlled by the service provider. Third, on binding the sampling parameters (i.e. p(x)) to the data x in constructing the rescaled size x/p(x), the need to keep independent track of p(x) (or even the original flow sizes x) is obviated. Thus, p(x) can vary at different times and across different regions of the network (as needed), but estimation remains unbiased. Fourth, sampling is composable in the sense that the first three properties above are preserved under successive resampling. Thus, one could progressively resample at different points in the measurement system in order to limit sample volumes. Also, size-dependent sampling is applicable to packet sampling as well. However, one expects the performance benefit over 1 in N sampling to be smaller in this case, since packet sizes do not have a heavy-tailed distribution.
  • The present invention utilizes an approach to usage-sensitive charging that mirrors the foregoing approach to sampling. The sampling scheme determines the size of the larger flows with no error. Estimation error arises entirely from sampling smaller flows. For billing purposes we wish to measure the total bytes for each billed entity (e.g. for each customer at a given service level) over each billing cycle. Larger totals have a smaller associated sampling error, whereas estimation of total bytes for the smallest customers may be subject to greater error. Therefore, the service provider sets a level L on the total bytes, with a fixed charge for all usage up to L, then a usage sensitive charge for all usage above L. Thus, the service provider only needs to tune the sampling scheme for estimating the usage above L within the desired accuracy.
  • Moreover, the potentially massive volumes of data to be gathered have important consequences for resource usage at each stage in the chain leading from data collection to data analysis. First, computational resources on network elements are scarce, and hence measurement functions may need to be de-prioritized in favor of basic packet forwarding and routing operations, particularly under heavy loads. Second, the transmission of raw measurement data to collection points can consume significant amounts of network bandwidth. Third, sophisticated and costly computing platforms are required for the storage and analysis of large volume of raw measurement data.
  • The present invention utilizes sampling as a means to reduce data volume while at the same time obtaining a representative view of the raw data. An elementary way to do this is to sample 1 in N raw data objects, either independently (i.e. each object is selected independently with probability 1/N) or deterministically (objects N, 2N, 3N, . . . are selected and all others are discarded). Only those selected objects are used further for analysis. This sampling strategy clearly reduces the load associated with the subsequent transmission, storage, and analysis of the data by a factor N.
  • However, besides the ability to reduce data volumes, the statistical properties of any proposed sampling scheme must be evaluated. The sampling parameters (N in the above example) need to be bounded to the sampled data in order that extensive properties of the original data stream can be estimated. For example, to estimate the bytes rate in a raw packet stream from samples gathered through 1 in N sampling, one needs to multiply the byte rate of the sampled stream by N. Under a given constraint on resources available for measurement transmission or processing of data, N may vary both temporally and spatially according to traffic volumes. Hence, N is not typically a global variable independent of the raw data.
  • Although one expects random sampling to yield unbiased estimates of properties of the typical raw data objects, there may be a significant impact of the variance of such estimates. A striking feature of flow statistics is that the distributions of the number of packet and bytes in flows are heavy-tailed. Consider the problem of reducing reported flow export volumes by sampling 1 in every N flow statistics. Sampling from heavy tailed distributions is particularly problematic, since the inclusion or exclusion of a small number of data points can lead to large changes in estimates of the mean. This has the consequence that estimates of the total byte rates on a link using a subset of flows selected by 1 in N sampling can be subject to high variance due to the sampling procedure itself. A sampling strategy that samples all big flows and a sufficient fraction of the smaller flows may reduce the estimator variance.
  • The basis of the sampling scheme is that sufficiently large objects (that may comprise packets or asynchronous transfer mode cells) are always sampled, while smaller objects are sampled with progressively smaller probability. A set of objects (flows) are labeled by i=1, 2, . . . , n corresponding to summaries generated by measurements in the network during some time period. Let xi be the size attribute of interest from the flow i, e.g., the number of packets in the flow, or the total number of bytes in the flow, or any other positive quantity of interest. Each packet in a flow possesses a common attribute, such as IP address (or net), port number, or Type of Service (ToS) field. Each combination of interest corresponds to attributes as referred by a “color”; ci will be the color of flow i. In the context of billing, a color might correspond to a customer address, or this plus a remote network, and possibly a ToS specification. The mapping that associates a particular customer with a set of packet attributes may be relatively complex;. This to be performed by the subsystem that collects and interprets the measurements (e.g. hosts 113 and 115 in the exemplary embodiment). The objective is to estimate the totals for each color c of interest as follows. X ( c ) = c x i ( 1 )
  • The present invention supports the sampling of raw packet headers, the set of flow statistics formed from the sampled packets, the stream of flow statistics at some intermediate aggregation point, and the set of aggregate flows at the collection point. The knowledge of the number n of original objects in not required. Furthermore, sampling itself need not make reference to the object color c. This reflects the fact that the colors of interest may not be known at the time of sampling and that it is infeasible to simply accumulate sizes from the original stream for all possible colors.
  • For each positive number z, one defines the sampling probability function pz(x)=min{1,x/z}. In the sampling scheme, a flow with size x is sampled with probability pz(x). The parameter z acts as a threshold: flow of size z or above are always sampled as shown in FIG. 2. The horizontal axis corresponds to xi (the size of an object 201). (In the disclosure, the parameter z is an example of a “probabilistic parameter.”) Each independent random variable wi has the values 1 with probability pz(xi) and 0 otherwise. Thus wi indicates whether flow i is to be sampled (wi=1) or not (wi=0). Each sampled value xi is renormalized by division by pz(xi). Thus, the estimate of the X (c) is given by: X ^ ( c ) = c w i x i / p z ( x i ) ( 2 )
  • In order to manage data network 100, the statistical variability of the estimate of X(c) provides a measure of confidence of the estimate. Moreover, the present invention enables the service provider to “tune” the operation of data network 100 in order to achieve the desired accuracy. In fact, pz(xi) is optimal in the sense that Var {circumflex over (X)}(c)+z2E(N(c)) is minimized with pz(xi), where E(N(c)) is the expected value of N(c). As will be explained later, the disclosure provides a method for controlling the statistical variance based upon operating parameters that the service provider can control. Parameter z is the size threshold above which flows are always sampled. The larger the value of z, the less likely that a given flow will be sampled and consequently the greater the variance associated with sampling it. If z is small, then Var {circumflex over (X)}(c)+z2E(N(c)) is more easily minimized by making Var {circumflex over (X)}(c) small, which occurs if one samples more of the flows. Conversely, if z is large, then Var {circumflex over (X)}(c)+z2E(N(c)) is more easily minimized by making E(N(c)) small, which occurs if one samples less of the flows.
  • Data networks supporting IP (as in data network 100) typically encounter heavy-tailed distributions of byte and packet sizes of IP. FIG. 3 displays an exemplary complementary cumulative distribution function (CCDF), i.e. the proportion of flows with bytes greater than a given level, of the flow sizes encountered by data network 100. The approximate linearity on the log-log scale is indicative of a heavy tailed distribution. The distribution of total bytes per customer-side IP address over a given period shares the heavy tailed property as shown in FIG. 4.
  • A statistic for comparing estimated usage with its actual usage is quantified by the Weighted Mean Relative Error (WMRE). WMRE = c X ^ ( c ) - X ( c ) c X ( c ) ( 3 )
    The WMRE averages the per-color absolute relative errors. WMRE gives greater weight to relative errors for large volume colors than for those with small volumes.
  • FIG. 5 illustrates an example of substantially better accuracy (smaller WMRE) of optimal sampling as compared with 1 in N sampling, over 4 orders of magnitude of the sampling period. Curve 501 illustrates the relationship for WMRE as a function of the effective sampling period, while curve 503 shows the corresponding relationship with sampling as described in the disclosure. With an effective sampling period of 100, the WMRE for optimal sampling is about only 1%, while for 1 in N sampling it is around 50%. The irregularity of the upper line reflects the sensitivity of the estimates from 1 in N sampling to random inclusion or exclusion of the largest flows during sampling. These features demonstrate the potential for inaccuracy arising from naive sampling from heavy-tailed distributions.
  • FIG. 6 displays with WMRE vs. sampling period for a trace of 107 flows (corresponding to curve 605), as compared with subportions contain 106 (corresponding to curve 603) and 105 (corresponding to curve 601) flows. The relative error decrease as the trace length increases, since the byte total for a given IP address is composed of a greater number of contributions. It may be desirable to place lower bounds on z in order to fulfill other objectives, such as limiting the rate at which samples are generated. The behavior from FIG. 6 suggests that is possible to simultaneously fulfill the goal of low relative error provided that the length of the period of observation (e.g. the billing period) is sufficiently long.
  • The exemplary embodiment utilizes the disclosed sampling techniques for charging the customer of data network 100 for usage. Fair charging requires that the deviation between the traffic charged to a customer and the actual traffic be kept to a minimum. The scheme is essentially the best possible, in the sense that variance of X is minimized for a given threshold z. However, the relative estimation error can be relatively large for colors with small amounts of traffic. As an extreme example, suppose the traffic associated with color c has total size X(c)<z. Each flow in that traffic thus has size less than z and will hence have a contribution to the estimate {circumflex over (X)}(c) that is either 0 (if the flow is not sampled), or z (if it is sampled, wherein the sample is normalized by pz(x)). Hence, {circumflex over (X)}(c) will be either 0, or at least z.
  • As a simple solution to the problem of estimating the small traffic volumes, the service provider can charge the traffic of a given color at a fixed fee, plus a usage-sensitive charge only for traffic volumes that exceed a certain level L. (L may depend on the color in question). The idea is to tune the sampling algorithms so that any usage X(c) that exceeds L can be reliably estimated. Usage X(c) that falls below L does not need to be reliably estimated, since the associated charge is usage-insensitive, i.e., independent of {circumflex over (X)}(c)<L.
  • Generally, one can consider traffic to be charged according to some function fc({circumflex over (X)}(c)) which depends on {circumflex over (X)}(c) only through the quantity max{{circumflex over (X)}(c), L}, i.e., it is independent of any usage below L. The subscript of fc indicates that the charge may depend on the color c, e.g., through the type of service, or foreign IP address. In the exemplary embodiment, the service charges the customer according to:
    f c({circumflex over (X)}(c))=a c +b c max{{circumflex over (X)}(c), L}  (4)
  • where “ac” is a fixed charge, than can encompass, e.g., port charges and administrative charges, “bc” is a per byte charge on traffic transmitted during the billing cycle, and “L” is the minimum usage. Equation 4 can also express pricing models in which there is a fixed administrative charge for small customers, whose usage doesn't warrant accurate measurement. Both ac and bc are allowed to depend on the color c in question.
  • Reliable estimation of the volumes X(c) is determined by choosing the sampling threshold z appropriately high for level L in question. The larger the level L and the larger the deviation of {circumflex over (X)}(c) from X(c) that can be tolerated, the higher a sampling level z one can allow.
  • The variance of all estimates for {circumflex over (X)}(c) greater than the level L can be controlled. This corresponds as a condition on the standard error, i.e., the ratio of standard deviation σ({circumflex over (X)}(c))=sqrt(Var {circumflex over (X)}(c)) to the mean X(c). In the exemplary embodiment, the typical estimation error is no more than about ε times X, for some target ε>0. This can be expressed this as the following standard error condition:
    σ({circumflex over (X)}(c))<εX(c) if X(c)>L   (5)
  • For example, with ε=0.05 the standard deviation cannot be more than 5% of the mean.
  • If {circumflex over (X)}(c) is derived from a large number of flows of independent sizes then {circumflex over (X)}(c) is roughly normally distributed. From Equation 5, the probability of overestimating {circumflex over (X)}(c)>L by an amount δX(c) (i.e., by δ/ε standard deviations) is no more than φ(−δ/ε), where φ is the standard normal distribution function. Thus, with ε=0.05, the probability of overestimating {circumflex over (X)}(c) by more than 10% (corresponding to δ) is approximately equal to (φ(−2)=2.23% (since 10%=2×5%).
  • The above approach sets limits on the chance that the deviation of the estimated usage above the actual usage exceeds a given amount. A refinement allows the service provider to set a limit on the chance that overcharging occurs. This should be more attractive from the customer's point of view since the chance of the customer being over billed at all can be small. Conversely, the service provider has to accept a small persistent under billing in order to accommodate the potential sampling error.
  • The distribution of {circumflex over (X)}(c) can be well approximated by a normal distribution when it is derived from a large number of constituent samples. If the probability of {circumflex over (X)}(c) being at least s standard deviations above the expected value X(c) is sufficiently small, then the calculated usage can be adjusted as follows:
    {circumflex over (X)}′(c)={circumflex over (X)}(c)−s√{square root over (z {circumflex over (X)}(c))}  (6)
    “s” is the number of standard deviations away from X(c) above which over-estimation is sufficiently rare. As an example, with s=3, φ(−s) is about 0.13%, i.e. about 1 in 740 traffic volumes will be overestimated. The service provider may charge according to {circumflex over (X)}′(c) rather than {circumflex over (X)}(c). In such a case, the customer is billed fc({circumflex over (X)}′(c)). Thus, the chance that the customer is over billed is approximately equal to φ(−s).
  • For the service provider, the difference {circumflex over (X)}(c)−{circumflex over (X)}″ (c)=s√{square root over (z {circumflex over (X)}(c))} represents unbillable revenue. In the charging scheme (as in Equation 4), this leads to under billing by a fraction roughly s√{square root over (2/X(c))}. Given the minimum billed volume L, the fraction of underbilling is no more than s√{square root over (z/L)}. (In variations of the exemplary embodiment, underbilling can be systematically compensated for in the charging rate bc). Thus, in order to limit potential undercharging to a fraction of no more than about η, the service provider determines s2z<η2L. In the example of s=3, underbilling by a fraction of no more than η=10% then requires selecting z and L such that z is less than about L/1000.
  • Table 1 shows the tradeoff of overcharging and unbillable usage.
    TABLE 1
    TRADE-OFF BETWEEN OVERCHARGING
    AND UNBILLABLE TRAFFIC
    overcharged
    Unbillable usage customers
    s = 0 −.1% 50% 
    s = 1 3.1% 3%
    s = 2 6.2% 0
  • Consider flows that present themselves for sampling at a rate ρ, in which the flow sizes have a distribution function F, i.e., F(x) is the proportion of flows that have size less than or equal to x. With a sampling threshold z, samples are produced at an average rate r=ρ∫F(dx)pz(x). Suppose there is a target maximum rate of samples r*<ρ. Then the service provider determines the sampling threshold z such that ρ∫F(dx)ρz(x)<r*. Using the fact that pz(x) is a decreasing function in z, it can be shown that this observation requires z≧z*, where z* is the unique solution z to the equation ρ∫F(dx) pz(x)=r*.
  • Let zo denote the maximum sampling threshold allowed in order to control sampling variance, e.g., z≦zo2L. The goals of controlling sample volume and variance are compatible provided that z* <=zo, for then any sampling threshold z in the interval [z*, zo] has the property of being sufficiently small to yield small sampling variance, and sufficiently large to restrict the average sampling rate no greater than the desired rate r*.
  • The condition z*≦zo can be realized by increasing the length of the billing cycle. The thresholds zo and z* control phenomena at different timescales. z* controls the average rate at which samples are taken. On the other hand, zo controls the sampling variance of the estimates {circumflex over (X)}(c) of total bytes over the billing timescale, potentially over days, weeks, or even months. The level byte L (under which accurate measurements are not needed) can be chosen to increase with the billing timescale. For example, the service provider may choose L to correspond to a particular quartile of the distribution of byte size, so that only a given proportion of the total bytes transmitted on the network are generated by customers whose total usage does not exceed L during the billing cycle. Increasing the length of the billing cycle will increase the corresponding quartile L, and hence also zo since is proportional to L. Support for this approach is provided by FIG. 6, which shows that the relative error in estimation decreases as the duration of collection of the flow trace increases.
  • FIG. 7 is a flow diagram for charging with sampled network usage in accordance with the exemplary embodiment. In step 701, threshold z is determined according to the relative error and the unbillable usage. In step 703, it is determined whether to sample an object that is associated with the customer in accordance with the size of the object and the probabilistic function pz(x). (The discussion with respect to FIG. 11 presents a method for determining whether to sample the objects based upon the size.) The associated usage is determined from the size of the sampled object by dividing the size by the pz(xi) in step 705. At the end of the billing in step 707, the usage sensitive pricing for the customer is calculated in step 709 in accordance with Equation 4 and adjustments by the service provider. In step 711, the usage is reset to 0 so that the usage for the next billing period can be calculated.
  • The present invention, as disclosed by the exemplary embodiment, also enables the service provider to control the sample volume that is generated by data network 100. Moreover, in data network 100, the amount of data is dynamic with time, and consequently the sampling rate needs to adjust accordingly. In other words, dynamic control of the mean sample volume may be needed.
  • An object (flow) may be distinguishable by an attribute. (Each object is characterized by a size that may be expressed in a number of packets, bytes (octets), or ATM cells contained in the object. The number is equal to at least one.) In such a case, the object is characterized as being colored. The present invention allows the service provider to estimate the total size of the objects in each color class c. If ci is the color of packet i, then X ( c ) = c x i
    is the total size of the objects with color c, and the unbiased estimator is then {circumflex over (X)}=Σwiwr(xi), in which {circumflex over (X)}(c) is obtained simply by summing the sampled normalized sizes of color c. Let X ^ ( c ) = c w i
    be the number of sampled objects with color c. By the linearity of expectation, E(N)=ΣE({circumflex over (N)}(c)). Also, since each xi is picked independently, {circumflex over (X)}(c) is independent for each c, and hence Var Var X = c Var X ^ ( c ) .
    Thus,
    C z(ρ)=Var {circumflex over (X)}+z 2 E({circumflex over (N)})   (7)
    where p is a probability function that is utilized for determining if an object is to be sampled.
  • The objective (cost) function Cz(p) is minimized locally over each color class. With variations of the exemplary embodiment, there may be scenarios in which there are different objectives for different colors. However, in the exemplary embodiment, the sampling device does not distinguish colors; however, samples can be later analyzed with respect to any combination of colors.
  • Finer control of sampling by color, within a given volume constraint, can only increase estimator variance. By applying a different threshold zc to the sampling of packets for each color, the service provider can control the sampling volume for each color. However, this approach increases the aggregate variance of {circumflex over (X)}(c).
  • In a dynamic context the volume of objects presented for sampling will generally vary with time. Thus, in order to be useful, a mechanism to control the number of samples must be able to adapt to temporal variations in the rate at which objects are offered for sampling. This is already an issue for the 1 in N sampling algorithm, since it may be necessary to adjust N, both between devices and at different times in a single device, in order to control the sampled volumes. For the optimal algorithm, the service provider can control the volume by an appropriate choice of the threshold z. Moreover, one can dynamically adapt (i.e. updating) z knowing only the target and current sample volumes.
  • Consider the case in which the targeted sampling volume M is less than n, which is the total number of objects from which to sample. N ^ z = i w i
  • is the total number of samples obtained using the sampling function pz. The expected number of samples N z = E ( N ^ z ) = i p z ( x i )
    is a non-increasing function of z. A direct approach to finding z* is to construct an algorithm to find the root, utilizing a set of xi (sizes of the sampled objects). FIG. 12, which is discussed later, illustrates the approach utilized in the exemplary embodiment.
  • Alternatively, the service provider can dynamically adapt (i.e. updating) z knowing only the target and current sample volumes. One approach is update z by:
    Z k+1 =Z k {circumflex over (N)}/M   (8)
    where M is the target sampling volume and {circumflex over (N)} is the measured sampling volume and where both correspond to the kth sampling window. As another alternative for dynamically updating z, the service provider can utilize the following:
    Z k+1 =Z k({circumflex over (N)}−{circumflex over (R)}/(M−{circumflex over (R)})   (9)
    where M is the target sampling volume, {circumflex over (N)} is the measured sampling volume, and {circumflex over (R)} is the measured sampling volume for objects having a size greater than zk, and where all correspond to the kth sampling window. (In the disclosure, “sampling window” is defined as being an interval during which objects are presented for sampling. The interval may be measured in time, e.g., in online applications where each object occurs at some time during the window. In offline applications, the objects have already been collected, and are then sampled offline. In this case, the interval might be measured in time, i.e. objects collected in a particular time window are presented for sampling, or in number where a certain number of objects are presented for sampling. The endpoint of the window may be determined prior to sampling, or it may depend on the objects, e.g. through the number that are sampled and/or their sizes.)
  • FIG. 8 shows an example of traffic flow volumes in data network 100. At approximately 100 seconds, data network 100 incurs a sudden increase of the traffic volume. FIG. 9 shows static (curve 901) and dynamic controlled (curve 903) sampling volumes in relation to FIG. 8. By adjusting threshold z, the sampling volume remains substantially constant relative to the sampling volume corresponding to a fixed threshold z.
  • If the arrival rate of objects to be sampled grows noticeably over a time scale shorter than the time duration (window width) of a sampling window, the exemplary embodiment enables the service provider to execute immediate corrective measures. The measured sampling volume N may significantly exceed the target M before the end of the sampling window. In the exemplary embodiment, if a target sample volume is already exceeded before the end of a window, the service provider should immediately change the threshold z. In this context, the windowing mechanism is a timeout that takes effect if N has not exceeded M by the end of the window. There are several variations of the exemplary embodiment. The corresponding emergency control can use timing information. If N already exceeds M at time t from a start of a window of length T, z is immediately replace by zT/t. Furthermore, if data network 100 provides control over the window boundaries, the a new sampling window can be started at that time. Otherwise, from time t one can reaccumulate the sample count N from zero, and the test and remedy procedure is repeated as needed for the remainder of the sampling window.
  • The target sampling volume M can be reduced to compensate for sampling variability. With a target sampling volume M, one can expect a relative error on N of about 1/√M. In order to guard against statistical fluctuations of up to s standard deviations from a target sampling volume M, the target sampling volume can be adjusted by:
    M s =M−s√M   (10)
    w
    here M, is the compensated target sampling volume.
  • FIG. 10 is a flow diagram for controlling the sampling volume in data network 100. The value of z is determined in accordance with a targeted sample volume. (FIG. 12 provides a method for determining z.) In step 1003, it is determined whether to sample the ith object having a size xi. (FIG. 11 provides a method for determining whether to sample the ith object.) If the object is sampled, then the corresponding traffic volume is estimated by normalizing xi by pz(xi) and accumulated to the estimated traffic volume in step 1005. At the end of the sampling window as determined by step 1007, the sampling volume is estimated in step 1009 for data network 100. If the sampling window is not completed, then step 1003 is repeated. In step 1011, the estimated sampling volume is utilized by the service provider in maintaining data network 100. There are a spectrum of associated activities, including traffic engineering studies, network planning, peering policy, customer acquisition, marketing, and network security. As part of the network planning activity, the service provider can reconfigure the data network to be better matched to the traffic volume.
  • FIG. 11 is a flow diagram for a quasi-random data sampling algorithm The process shown in FIG. 11 can be utilized by step 703 or by step 1003 in determining whether to sample an object (flow). In the exemplary embodiment as shown in FIG. 11, it is assumed that the variable “count” has a uniformly distributed value between 0 and z-1. In step 1101, count is reset to zero. In step 1103, the size of the object xi is compared to z. If xi is greater or equal to z, then the ith object is sampled in sample 1105. The index i is incremented by 1 in step 1107 so that the next object is considered in the next execution of step 1103. However, if xi is less than z in step 1103, then count is incremented by xi in step 1109. If count is greater than or equal to z in step 1111, count is decremented by z in step 1113 and the ith object is sampled in step 1115. However, if count is less than z in step 1111, index i is incremented by 1 in step 1107 so that the next object is considered for the next execution of step 1103.
  • FIG. 11 is one embodiment of a quasi-random data sampling algorithm. One skilled in the art appreciates that other quasi-random embodiments can be utilized in order to determine whether to sample an object.
  • FIG. 12 is a flow diagram for root finding algorithm that may be utilized in determining or updating z in step 701 (FIG. 7) or step 1001 (FIG. 10). Discussion of FIG. 12 is facilitated with specific notations as follow.
      • {X} is a set {xi}, where each xi is the size of the ith object
      • {X|condition} is a subset of {X}, where each member satisfies the given condition
      • |X| is a number that is equal to the number of members in the set {X}
      • sum{Y} is a number equal to the sum of the members of {Y}
  • The approach of the process shown in FIG. 12 is to select a candidate z and to determine if the candidate z is satisfactory, too large, or too small. The process utilizes a collection of numbers corresponding to the sizes of previously sampled objects and the target sampling volume M. However, the process as illustrated in FIG. 12 does not change the value of M. Rather, variables M, B, and C are internal variables that are used for calculations. The process only returns the appropriate value of z in steps 1205 and 1219.
  • In step 1201, M and {X} are inputted. Internal variable B is reset to zero. In step 1203, the number of members in {X} is compared to zero. If so, z=B/M is returned in step 1205 and the routine is exited. In step 1209, z is randomly selected from {X}. An efficient implementation may require that z be picked randomly from {X} so that the expectation is somewhere in the middle with respect to size. However, assuming that the order of the members from {X} is independent of size, one can let z be equal to the first member in {X}. In step 1211, set {Y} consists of the members of {X} whose values are less than z. In step 1213, C=sum{Y}, where C is an internal variable that is used for calculations. In step 1215, N=(B+C)/z+|X|−|Y|. |X| and |Y| are equal to the number of elements contained in {X} and {Y}, respectively. In step 1217, N is compared to M. If so, z is equal to xi that was selected in step 1209. If N is not equal to M, then step 1221 determines if N is greater than M. If so, {X}={X|x>z} in step 1223. In other words, members of set {X} are removed from the set whose values that are smaller or equal to z. Also, B=B+sum {X|x<=z}. In other words, B is incremented by the sum of the members that are removed from the set {X}. Step 1203 is then repeated. If step 1221 determines that N is not greater than M, then N is less than M. In that case, step 1225 is executed. In step 1225, set {X} is equal to set {Y}, where {Y} consists of the members of the previous set {X} that are less than z (as determined by step 1211). Also, M=M−(|X|−|Y|). In other words, M is reduced by |X|−|Y|. Step 1203 is then repeated.
  • FIG. 13 shows an apparatus 1300 for managing a data network in accordance with the present invention. Apparatus 1300 receives and sends packets that are transported by a data network through packet interface 1303. Processor 1301 receives packets containing traffic information through link 1302 from packet interface 1303. In a variation of the embodiment, apparatus 1300 provides router functionality with routing module 1305. Routing module 1305 directs packets between packet interface 1303 and packet interface 1307 through links 1304 and 1306 and between packet interface 1303 and packet interface 1309 through links 1304 and 1308. Processor 1301 configures routing module 1305 through link 1310 in accordance with the traffic information that is received through link 1302. Processor 1301 executes computer instructions corresponding to the flow diagrams shown in FIGS. 7, 10, 11, and 12.
  • In another variation of the embodiment, processor 1301 stores traffic information (that is received through link 1302) to storage device 1311 through link 1312 for later retrieval (e.g. billing information for charging a customer).
  • As can be appreciated by one skilled in the art, a computer system with an associated computer-readable medium containing instructions for controlling the computer system can be utilized to implement the exemplary embodiments that are disclosed herein. The computer system may include at least one computer such as a microprocessor, digital signal processor, and associated peripheral electronic circuitry.
  • It is to be understood that the above-described embodiment is merely an illustrative principle of the invention and that many variations may be devised by those skilled in the art without departing from the scope of the invention. It is, therefore, intended that such variations be included with the scope of the claims.

Claims (22)

1. An apparatus for managing a data network, comprising:
means for receiving an object, wherein the object is characterized by at least one attribute and wherein the object comprises at least one data element;
means for determining whether to sample the object in accordance with a probabilistic parameter; and
means for sampling the object if it is determined that sampling is required.
2. The apparatus of claim 1, wherein the probabilistic parameter is determined from a cost function.
3. The apparatus of claim 2, wherein the cost function relates a network resource to a quality of measurements.
4. The apparatus of claim 3, wherein the network resource corresponds to a sampling volume and the quality of measurements corresponds to a sampling accuracy.
5. The apparatus of claim 1, further comprising:
means for aggregating a plurality of samples in accordance with the at least one attribute.
6. The apparatus of claim 1, wherein said means for determining utilizes one of the at least one attribute to determine whether to sample the object.
7. The apparatus of claim 6, wherein said one of the at least one attribute comprises a size of the object, wherein the size includes a contribution of the at least one data element.
8. The apparatus of claim 6, wherein the object comprises at least one data element, wherein the data element is selected from a group comprising at least one of: an octet, an Internet Protocol (IP) packet, a frame relay packet, or an Asynchronous Transfer Mode (ATM) cell.
9. The apparatus of claim 1, further comprising:
means for determining a measured usage of the data network in accordance with the at least one attribute; and
means for charging a customer for the measured usage in accordance with a charging function, wherein the customer is associated with the at least one attribute and wherein the customer is presented a bill for a billing period and wherein a charging accuracy is related to the charging function and an accuracy of the measured usage.
10. The apparatus of claim 9, wherein said means for charging utilizes a minimum usage and a usage charge, or wherein said means for charging utilizes a fixed charge.
11. The apparatus of claim 1, further comprising:
means for calculating an estimated sampling volume.
12. A method for managing a data network, comprising:
receiving an object, wherein the object is characterized by at least one attribute and wherein the object comprises at least one data element;
determining whether to sample the object in accordance with a probabilistic parameter; and
sampling the object if it is determined that sampling is required.
13. The method of claim 12, wherein the probabilistic parameter is determined from a cost function.
14. The method of claim 13, wherein the cost function relates a network resource to a quality of measurements.
15. The method of claim 14, wherein the network resource corresponds to a sampling volume and the quality of measurements corresponds to a sampling accuracy.
16. The method of claim 12, further comprising:
aggregating a plurality of samples in accordance with the at least one attribute.
17. The method of claim 12, wherein said determining utilizes one of the at least one attribute to determine whether to sample the object.
18. The method of claim 17, wherein said one of the at least one attribute comprises a size of the object, wherein the size includes a contribution of the at least one data element.
19. The method of claim 17, wherein the object comprises at least one data element, wherein the data element is selected from a group comprising at least one of: an octet, an Internet Protocol (IP) packet, a frame relay packet, or an Asynchronous Transfer Mode (ATM) cell.
20. The method of claim 12, further comprising:
determining a measured usage of the data network in accordance with the at least one attribute; and
charging a customer for the measured usage in accordance with a charging function, wherein the customer is associated with the at least one attribute and wherein the customer is presented a bill for a billing period and wherein a charging accuracy is related to the charging function and an accuracy of the measured usage.
21. The method of claim 20, wherein said charging utilizes a minimum usage and a usage charge, or wherein said charging utilizes a fixed charge.
22. The method of claim 12, further comprising:
calculating an estimated sampling volume.
US11/924,329 2001-03-18 2007-10-25 Apparatus for size-dependent sampling for managing a data network Abandoned US20080043636A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/924,329 US20080043636A1 (en) 2001-03-18 2007-10-25 Apparatus for size-dependent sampling for managing a data network

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US27712301P 2001-03-18 2001-03-18
US30058701P 2001-06-22 2001-06-22
US10/056,683 US7080136B2 (en) 2001-03-18 2002-01-24 Method and apparatus for size-dependent sampling for managing a data network
US11/478,960 US7299283B1 (en) 2001-03-18 2006-06-27 Apparatus for size-dependent sampling for managing a data network
US11/924,329 US20080043636A1 (en) 2001-03-18 2007-10-25 Apparatus for size-dependent sampling for managing a data network

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/478,960 Continuation US7299283B1 (en) 2001-03-18 2006-06-27 Apparatus for size-dependent sampling for managing a data network

Publications (1)

Publication Number Publication Date
US20080043636A1 true US20080043636A1 (en) 2008-02-21

Family

ID=27369081

Family Applications (3)

Application Number Title Priority Date Filing Date
US10/056,683 Expired - Lifetime US7080136B2 (en) 2001-03-18 2002-01-24 Method and apparatus for size-dependent sampling for managing a data network
US11/478,960 Expired - Fee Related US7299283B1 (en) 2001-03-18 2006-06-27 Apparatus for size-dependent sampling for managing a data network
US11/924,329 Abandoned US20080043636A1 (en) 2001-03-18 2007-10-25 Apparatus for size-dependent sampling for managing a data network

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US10/056,683 Expired - Lifetime US7080136B2 (en) 2001-03-18 2002-01-24 Method and apparatus for size-dependent sampling for managing a data network
US11/478,960 Expired - Fee Related US7299283B1 (en) 2001-03-18 2006-06-27 Apparatus for size-dependent sampling for managing a data network

Country Status (1)

Country Link
US (3) US7080136B2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070011329A1 (en) * 2005-07-06 2007-01-11 Cisco Technology, Inc. Techniques for accounting for multiple transactions in a transport control protocol (TCP) payload
US20090144304A1 (en) * 2007-11-30 2009-06-04 Josh Stephens Method for summarizing flow information of network devices
US20090232012A1 (en) * 2005-06-03 2009-09-17 Fraunhofer-Gesellschaft Zur Forderung Der Angewandten Forschung E.V. Method and Monitoring System for Sample-Analysis of Data Comprising a Multitute of Data Packets
US20110149794A1 (en) * 2009-12-21 2011-06-23 Electronics And Telecommunications Research Institute Apparatus and method for dynamically sampling of flow

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020188710A1 (en) * 2001-03-18 2002-12-12 At&T Corp. Size-dependent sampling for managing a data network
US7080136B2 (en) 2001-03-18 2006-07-18 At & T Corp. Method and apparatus for size-dependent sampling for managing a data network
US7536455B2 (en) * 2001-03-18 2009-05-19 At&T Corp. Optimal combination of sampled measurements
US7027396B1 (en) 2002-02-13 2006-04-11 At&T Corp. Traffic matrix computation for a backbone network supporting virtual private networks
US7487121B2 (en) 2002-07-08 2009-02-03 Convergys Cmg Utah Flexible event correlation aggregation tool
US20040006608A1 (en) * 2002-07-08 2004-01-08 Convergys Cmg Utah Flexible network element interface
US8285825B1 (en) * 2002-11-13 2012-10-09 Novell, Inc. Method and system for managing network resources based on a dynamic quorum
GB2399708B (en) * 2003-03-20 2006-03-22 Parc Technologies Ltd Assisted determination of data flows in communication/data networks
US6999895B2 (en) * 2003-03-21 2006-02-14 International Business Machines Corporation Method and structure for dynamic sampling method in on-line process monitoring
CA2933668C (en) 2004-04-23 2019-01-08 The Nielsen Company (Us), Llc Methods and apparatus to maintain audience privacy while determining viewing of video-on-demand programs
US7653007B2 (en) * 2004-06-04 2010-01-26 Alcatel-Lucent Usa Inc. Per-flow traffic estimation
US7460487B2 (en) * 2004-06-04 2008-12-02 Lucent Technologies Inc. Accelerated per-flow traffic estimation
US7782793B2 (en) * 2005-09-15 2010-08-24 Alcatel Lucent Statistical trace-based methods for real-time traffic classification
US8081581B2 (en) * 2008-01-30 2011-12-20 Microsoft Corporation Sampling rules for information dissemination
US7990982B2 (en) * 2008-12-15 2011-08-02 At&T Intellectual Property I, L.P. Methods and apparatus to bound network traffic estimation error for multistage measurement sampling and aggregation
US8064359B2 (en) * 2008-12-23 2011-11-22 At&T Intellectual Property I, L.P. System and method for spatially consistent sampling of flow records at constrained, content-dependent rates
US7957315B2 (en) * 2008-12-23 2011-06-07 At&T Intellectual Property Ii, L.P. System and method for sampling network traffic
US8443080B2 (en) * 2010-05-06 2013-05-14 Nec Laboratories America, Inc. System and method for determining application dependency paths in a data center
US8307006B2 (en) 2010-06-30 2012-11-06 The Nielsen Company (Us), Llc Methods and apparatus to obtain anonymous audience measurement data from network server data for particular demographic and usage profiles
US10652318B2 (en) * 2012-08-13 2020-05-12 Verisign, Inc. Systems and methods for load balancing using predictive routing
US10691082B2 (en) * 2017-12-05 2020-06-23 Cisco Technology, Inc. Dynamically adjusting sample rates based on performance of a machine-learning based model for performing a network assurance function in a network assurance system
CN115357813B (en) * 2022-10-20 2023-04-07 荣耀终端有限公司 Sampling method and device and electronic equipment

Citations (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5197002A (en) * 1989-12-22 1993-03-23 Bell Communications Research, Inc. Methods and apparatus for dynamic hashing
US5235644A (en) * 1990-06-29 1993-08-10 Digital Equipment Corporation Probabilistic cryptographic processing method
US6061331A (en) * 1998-07-28 2000-05-09 Gte Laboratories Incorporated Method and apparatus for estimating source-destination traffic in a packet-switched communications network
US6119109A (en) * 1996-09-30 2000-09-12 Digital Vision Laboratories Corporation Information distribution system and billing system used for the information distribution system
US6230204B1 (en) * 1997-12-19 2001-05-08 Micron Electronics, Inc. Method and system for estimating usage of computer resources
US6330008B1 (en) * 1997-02-24 2001-12-11 Torrent Systems, Inc. Apparatuses and methods for monitoring performance of parallel computing
US6347224B1 (en) * 1996-03-29 2002-02-12 British Telecommunications Public Limited Company Charging systems for services in communications
US6360261B1 (en) * 1997-02-14 2002-03-19 Webtrends Corporation System and method for analyzing remote traffic data in distributed computing environment
US6421435B1 (en) * 1998-11-30 2002-07-16 Qwest Communications International Inc. SS7 network planning tool
US20020165958A1 (en) * 2001-03-18 2002-11-07 At&T Corp. Apparatus for size-dependent sampling for managing a data network
US20020188710A1 (en) * 2001-03-18 2002-12-12 At&T Corp. Size-dependent sampling for managing a data network
US6725263B1 (en) * 2000-03-21 2004-04-20 Level 3 Communications, Inc. Systems and methods for analyzing network traffic
US6735553B1 (en) * 2000-07-13 2004-05-11 Netpredict, Inc. Use of model calibration to achieve high accuracy in analysis of computer networks
US6738349B1 (en) * 2000-03-01 2004-05-18 Tektronix, Inc. Non-intrusive measurement of end-to-end network properties
US6775267B1 (en) * 1999-12-30 2004-08-10 At&T Corp Method for billing IP broadband subscribers
US6823225B1 (en) * 1997-02-12 2004-11-23 Im Networks, Inc. Apparatus for distributing and playing audio information
US6853619B1 (en) * 1999-02-26 2005-02-08 Ipanema Technologies System and method for measuring the transfer durations and loss rates in high volume telecommunication networks
US6920112B1 (en) * 1998-06-29 2005-07-19 Cisco Technology, Inc. Sampling packets for network monitoring
US6947723B1 (en) * 2002-01-14 2005-09-20 Cellco Partnership Postpay spending limit using a cellular network usage governor
US6990073B1 (en) * 2001-03-28 2006-01-24 Lsi Logic Corporation Data packet congestion management technique
US7058590B2 (en) * 2001-05-04 2006-06-06 Hewlett-Packard Development Company, L.P. System and method for generating conversion-related estimates utilizing adaptive sample size
US7130306B1 (en) * 2000-10-25 2006-10-31 Pasi Into Loukas Pulse group soliton transmission
US7130612B1 (en) * 2000-05-30 2006-10-31 At&T Corp. System and method for providing wireless services within a wireless local area network
US7215637B1 (en) * 2000-04-17 2007-05-08 Juniper Networks, Inc. Systems and methods for processing packets

Patent Citations (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5197002A (en) * 1989-12-22 1993-03-23 Bell Communications Research, Inc. Methods and apparatus for dynamic hashing
US5235644A (en) * 1990-06-29 1993-08-10 Digital Equipment Corporation Probabilistic cryptographic processing method
US6347224B1 (en) * 1996-03-29 2002-02-12 British Telecommunications Public Limited Company Charging systems for services in communications
US6119109A (en) * 1996-09-30 2000-09-12 Digital Vision Laboratories Corporation Information distribution system and billing system used for the information distribution system
US6823225B1 (en) * 1997-02-12 2004-11-23 Im Networks, Inc. Apparatus for distributing and playing audio information
US6360261B1 (en) * 1997-02-14 2002-03-19 Webtrends Corporation System and method for analyzing remote traffic data in distributed computing environment
US6330008B1 (en) * 1997-02-24 2001-12-11 Torrent Systems, Inc. Apparatuses and methods for monitoring performance of parallel computing
US6230204B1 (en) * 1997-12-19 2001-05-08 Micron Electronics, Inc. Method and system for estimating usage of computer resources
US6920112B1 (en) * 1998-06-29 2005-07-19 Cisco Technology, Inc. Sampling packets for network monitoring
US6061331A (en) * 1998-07-28 2000-05-09 Gte Laboratories Incorporated Method and apparatus for estimating source-destination traffic in a packet-switched communications network
US6421435B1 (en) * 1998-11-30 2002-07-16 Qwest Communications International Inc. SS7 network planning tool
US6853619B1 (en) * 1999-02-26 2005-02-08 Ipanema Technologies System and method for measuring the transfer durations and loss rates in high volume telecommunication networks
US6775267B1 (en) * 1999-12-30 2004-08-10 At&T Corp Method for billing IP broadband subscribers
US6738349B1 (en) * 2000-03-01 2004-05-18 Tektronix, Inc. Non-intrusive measurement of end-to-end network properties
US6725263B1 (en) * 2000-03-21 2004-04-20 Level 3 Communications, Inc. Systems and methods for analyzing network traffic
US7215637B1 (en) * 2000-04-17 2007-05-08 Juniper Networks, Inc. Systems and methods for processing packets
US7130612B1 (en) * 2000-05-30 2006-10-31 At&T Corp. System and method for providing wireless services within a wireless local area network
US6735553B1 (en) * 2000-07-13 2004-05-11 Netpredict, Inc. Use of model calibration to achieve high accuracy in analysis of computer networks
US7130306B1 (en) * 2000-10-25 2006-10-31 Pasi Into Loukas Pulse group soliton transmission
US20020188710A1 (en) * 2001-03-18 2002-12-12 At&T Corp. Size-dependent sampling for managing a data network
US20020165958A1 (en) * 2001-03-18 2002-11-07 At&T Corp. Apparatus for size-dependent sampling for managing a data network
US7299283B1 (en) * 2001-03-18 2007-11-20 At&T Corp. Apparatus for size-dependent sampling for managing a data network
US6990073B1 (en) * 2001-03-28 2006-01-24 Lsi Logic Corporation Data packet congestion management technique
US7058590B2 (en) * 2001-05-04 2006-06-06 Hewlett-Packard Development Company, L.P. System and method for generating conversion-related estimates utilizing adaptive sample size
US6947723B1 (en) * 2002-01-14 2005-09-20 Cellco Partnership Postpay spending limit using a cellular network usage governor

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090232012A1 (en) * 2005-06-03 2009-09-17 Fraunhofer-Gesellschaft Zur Forderung Der Angewandten Forschung E.V. Method and Monitoring System for Sample-Analysis of Data Comprising a Multitute of Data Packets
US8045470B2 (en) * 2005-06-03 2011-10-25 Fraunhofer-Gesellschaft Zur Forderung Der Angewandten Forschung E.V. Method and monitoring system for sample-analysis of data comprising a multitute of data packets
US20070011329A1 (en) * 2005-07-06 2007-01-11 Cisco Technology, Inc. Techniques for accounting for multiple transactions in a transport control protocol (TCP) payload
US8438281B2 (en) * 2005-07-06 2013-05-07 Cisco Technology, Inc. Techniques for accounting for multiple transactions in a transport control protocol (TCP) payload
US9716636B2 (en) 2005-07-06 2017-07-25 Cisco Technology, Inc. Techniques for accounting for multiple transactions in a transport control protocol (TCP) payload
US20090144304A1 (en) * 2007-11-30 2009-06-04 Josh Stephens Method for summarizing flow information of network devices
US9331919B2 (en) * 2007-11-30 2016-05-03 Solarwinds Worldwide, Llc Method for summarizing flow information of network devices
US20110149794A1 (en) * 2009-12-21 2011-06-23 Electronics And Telecommunications Research Institute Apparatus and method for dynamically sampling of flow

Also Published As

Publication number Publication date
US20020165958A1 (en) 2002-11-07
US7080136B2 (en) 2006-07-18
US7299283B1 (en) 2007-11-20

Similar Documents

Publication Publication Date Title
US7299283B1 (en) Apparatus for size-dependent sampling for managing a data network
US8028055B2 (en) Optimal combination of sampled measurements
US11316790B2 (en) System and method for managing bandwidth usage rates in a packet-switched network
US20020188710A1 (en) Size-dependent sampling for managing a data network
US8601155B2 (en) Telemetry stream performance analysis and optimization
Duffield et al. Predicting resource usage and estimation accuracy in an IP flow measurement collection infrastructure
EP2834949B1 (en) Congestion control and resource allocation in split architecture networks
Ott et al. Sred: stabilized red
US7668966B2 (en) Data network controller
CN102150394B (en) Systems and methods for determining top spreaders
US7924739B2 (en) Method and apparatus for one-way passive loss measurements using sampled flow statistics
US7124180B1 (en) Internet usage data recording system and method employing a configurable rule engine for the processing and correlation of network data
US20200014549A1 (en) Computer Network Service Providing System Including Self Adjusting Volume Enforcement Functionality
US20040044753A1 (en) Method and system for dynamic business management of a network
US7366790B1 (en) System and method of active latency detection for network applications
Varghese et al. The measurement manifesto
Cottrell et al. Experiences and results from a new high performance network and application monitoring toolkit
EP1911203A1 (en) Optimal combination of sampled measurements
US20230231783A1 (en) Traffic matrix estimation with the QSP-MLE model and recursive improvement
Thorup Charging from sampled network usage
Ivanovich et al. Modelling GPRS data traffic
Jarrett Congestion detection within multi-service TCP/IP networks using wavelets
Thorup Learn More, Sample Less: Control of Volume and Variance in Network Measurement
Osman Internet backbone network traffic in Sudan
Müller Traffic Profiles and Performance Modelling of Heterogeneous Networks

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION